text
stringlengths
234
589k
id
stringlengths
47
47
dump
stringclasses
62 values
url
stringlengths
16
734
date
stringlengths
20
20
file_path
stringlengths
109
155
language
stringclasses
1 value
language_score
float64
0.65
1
token_count
int64
57
124k
score
float64
2.52
4.91
int_score
int64
3
5
AI is shaping the world in more ways than we can imagine, and its advancement has brought a wave of unprecedented disruption across every sector. We need to train our future workforce to be adept with the necessary skills and also make them aware of the ethical responsibilities that come with the power of Artificial Intelligence (AI). This has sparked conversations on delivering hands-on artificial intelligence (AI) education for K12 students. To enable AI for K-12, educators and curriculum developers need the right courseware that supports hands-on learning. Cloud-based technologies offer safe and scalable solutions for AI learning. The surge in AI enabled education is evident from a number of state schools taking up initiatives of integrating classrooms with AI tools for education. Florida became the pioneer in adopting AI integrated curriculum for K-12. Similarly, a school district in Gwinnet County,Georgia, has introduced AI integrated curriculum, bringing technology into classrooms and facilitating students with AI literacy. Achieving this phenomenal task of introducing AI education to school students has its own set of challenges. In this blog, we’ll take a deep dive into what it takes for curriculum developers to achieve this monumental task and how to equip students with hands-on AI education through a tech-infused learning platform. Key Elements to Enable AI Education for K-12 To introduce students to Artificial Intelligence (AI) core concepts and technologies, textbook education cannot suffice. Educators need to go beyond the conventional pedagogy. They need tech-infused courseware that not only makes abstracts from textbooks come to life but equips students with hands-on practice. With the right courseware and resources students can learn by doing, building and breaking, trying and testing. Whether spinning up a virtual machine (VM) or developing a chatbot, students can go all in with AI technologies. School faculty will need to address the following requirements to be able to deliver AI education among school students. - Enable hands-on learning: As we have already discussed why hands-on learning is imperative when it comes to technology-driven education, let’s delve deeper into its significance. Artificial Intelligence (AI) and Robotics are complex topics, it requires analytical thinking and problem-solving skills among students. These skills can only be developed with hands-on practical experiences. How hands-on virtual labs can serve as the best tools for AI education in K-12 - These are cloud-based virtual environments, that can simulate real-world challenges to equip students with practical skills. - Non-production environments for safe learning. - Students can experience guided hands-on learning and explore various aspects of AI including Machine Learning, Natural Language Processing and Computer Vision. - Develop necessary courseware: Developing the right courseware includes selecting the kind of labs that serve your training objectives, helping students with practice materials and providing practice tests for evaluating students’ knowledge and skills. Also instructors who have been teaching in the conventional system of education might not find themselves well-versed with managing cloud-based labs. To make the integration easier, school faculty needs pre-built labs like the AI-900 labs provided by CloudLabs. These are pre-built and completely ready-to-use labs, managed end-to-end by CloudLabs team. The school faculty would not need any IT support or any coding experience to launch and distribute these labs. - Instructor Upskilling: Instructor upskilling is a crucial part of any educational program. To make sure adoption of AI education at K-12 is successful, professional development programs must be put in action as a part of AI infused training. To help teachers get familiar with what their classroom sessions will look like, allowing them to get hands-on with AI-900 labs can be the first step. However, instructors must be given additional support with lesson plans and instructional resources Additional steps that can help instructors in integrating AI education for K-12 curriculum: - Access to age-appropriate teaching materials, resources, and curriculum modules designed to engage students at different developmental levels. - Continuous learning opportunities to stay abreast of evolving educational approaches and emerging AI technologies relevant to different age groups. - In collaboartion with curriculum developers, instructors need guidance in integrating present curricula with AI concepts and applications, with tailored lesson plans and content aligning with teaching standards set by the education department of the respective state. Challenges Faced by K-12 schools in delivering AI training: District schools while implementing AI education at K-12 level face several challenges. The lack of required courseware, limited experience with using software technology for both instructors and students accompanied with data privacy issues impedes the large-scale adoption of AI at K-12. - Lack of resources : The most significant challenge faced by schools is the lack of courseware and resources. To facilitate AI-enabled learning in classrooms, pre-built AI-900 labs can be an essential courseware. However, to use cloud-based labs, students need cloud resources to complete various tasks. For example: students will require Azure Bot Services to develop chatbots, Azure Computer Vision to perform image recognition and so on during their lab tasks. The availability of these cloud-based resources is a major challenge. While Microsoft’s Azure for Students program is a good option for beginners, it comes with its own setbacks. The free $100 cloud credit provided by Microsoft cannot be replenished before the year end. Also, certain resources are restricted to access under this program. Consequently, school faculty would need to invest extensive time developing their own resources. Since this program is tied to students’ Azure accounts, instructors do not get any visibility into users’ lab activities, the resources they are provisioning or the credits they have consumed. Hence tracking student progress and assessment would be challenging. School faculty needs a better alternative one that supports all the required resources, for the required duration along with centralized control of all students’ environments. - Individual student assessment is exhausting and time consuming for instructors: There is nothing conventional about AI infused classroom training, and hence student assessment also needs a better alternative, one that is not time consuming and exhausting. The additional perks of this alternative must include Skill Validation to identify if the learners were fully able to grasp the concepts and apply their training into practice. This model of assessment which is completely automated and can demonstrate learners’ skills is made possible with CloudLabs’ Lab Validation feature. - Determining Certification Success Among Students: School faculty needs detailed analysis into student performance and rate of lab completion to determine certification success. By tracking student activity, lab status, and lab scores, they can identify areas of improvement and develop data-informed strategies for a successful certification program. How CloudLabs Managed to Enable AI Education at Scale The challenges that restrict the adoption of AI for K-12 at scale can be effectively mitigated with CloudLabs’ capabilities. CloudLabs is a virtual labs platform that aims to make IT training immersive with hands-on experiences. However, it’s the multiple features of the platform and the keen emphasis on meeting every unique requirement of our partners that makes CloudLabs the most preferred solution to enable artificial intelligence in K-12 education. - Instantly available pre-built labs: CloudLabs has simplified the deployment of virtual labs with readily available pre-configured labs that are fully managed end-to-end. These labs require no complex setup or any additional configuration. The easy accessibility to virtual labs makes it the ideal solution with no IT overhead for the school faculty and minimum infrastructure costs. - Easy Lab Distribution: Lab deployment is a seamless experience that students can achieve in a few simple steps. Once the instructor shares the URL, students can access their labs on the CloudLabs portal. Since the labs are pre-deployed, the virtual environments (VMs) easily spin up in a few minutes. The students need nothing more than a powerful internet connection. Instructors can easily scale the number of labs as per the student count with no infrastructure dependency. The entire process involves no trouble shooting making it a highly viable option for the school faculty. - Optimized lab infrastructure: CloudLabs provide end-to-end lab provisioning. The lab infrastructure is highly optimized to meet the user requirements. All the necessary resources are pre-configured, and licenses are extended to all users depending upon the training requirement. The instructors can focus solely on training without spending time on provisioning resources or fine-tuning the VMs. - A comprehensive courseware for hands-on learning: Instructors require comprehensive courseware for impactful classroom training, offering essential student support. Those engaging with AI technologies in cloud-based lab environments for the first time may encounter challenges navigating lab tasks, leading to errors, overlooked steps, and unintended outcomes. To overcome this challenge, CloudLabs provides detailed lab guides with step-by-step instructions to follow for a given lab assignment. Students also get a set of practice tests to evaluate their skills and knowledge gained during the hands-on sessions. This alleviates a heavy task for the instructors of curating courseware right from scratch. - Lab Validation to determine certification success : To determine the success of the certification program, it is necessary to practice Skill Validation for students. Students can turn on the Lab Validation feature and demonstrate their skills gained with hands-on practice. After completing the tasks, students can validate their labs by clicking the lab validation button in the virtual machine. A successful lab validation status entails successful task completion, whereas if any step is missed or done incorrectly, students will receive a review message to reattempt and correct it. While students can practice self-assessment as they get a second chance to rectify their mistakes and learn from it, instructors can automate the student assessment process saving them more time. Why CloudLabs is the Ideal Solution for AI in K-12 Education: - Improved learning outcomes with end-to-end student support: To ensure successful learning outcomes, end-to-end student support was crucial. Instructors can closely monitor student activities and jump in whenever students need their support through Shadow Labs. Shadow labs provide an over-the-shoulder view of student labs, to provide real time support. - Streamline student assessment with lab validation: With Lab Validation feature, auto grading of student labs made instructor workload significantly lighter. Instructors can get a complete overview of class performance for a given lab module with individual lab score. - Auto-generate certification reports: Determine certification success by analyzing the rate of successful lab completion. Offers Power BI reports for users’ lab activities and lab scores. - A transparent billing system: CloudLabs follow a fixed-billing system for pre-built labs. No extra charge on resource consumption or exceeded budget. Additional Highlights of CloudLabs Platform For a Successful AI Training Program: - Admin Console: Instructors can have full visibility and control into students’ lab environments with CloudLabs’ Admin Console.The Admin Console is an easy-to-use self-service portal that grant admin controls to instructors based on which instructors can: - Enable user access for registered students and distribute labs. - Monitor students’ lab status, whether Approved, Deploying, Failed or Deleted. - Approve necessary licenses for students. - Control user labs with Start/Suspend feature. - Access Validation report for each attendee to track their progress. - Override validation status, with this feature instructors can change validation status for users, as per the need: For example. ‘Failed’ validation status can be changed to ‘Succeeded’ - CloudLabs Portal for Learners : The CloudLabs portal is designed for smooth learning experience, with quick onboarding, simple navigation, and an immersive user interface. Learners can optimize their learning journey with a series of features embedded into the portal. - Learners can easily launch their labs in a few simple steps with the signup link and activation code/voucher code. - The virtual machine comes with a split screen dedicated to all necessary details such as lab duration and navigational instructions. - Navigational instructions are divided into a series of tabs for lab guides, environment details, and resource information that comes in handy while working with the labs. - The lab guides offer step-by-step instructions to accomplish the lab task successfully. - Environment details offer Azure login credentials and virtual machine credentials. - Learners also can keep track of the resources deployed in the given environment and their maximum uptime limit. By starting and stopping the virtual machine, learners can efficiently manage their resources. - Students who have completed their lab task can click on the lab validation button and undergo a self- assessment to identify mistakes and correct them with the review error provided by the platform. - Compliance with students’ privacy protection policies: CloudLabs stands out as the ideal training platform for district schools due to its strict adherence to US data privacy and protection policies for school students. By complying with the Children’s Online Privacy Protection Act (“COPPA”) and the Family Educational Rights and Privacy Act (“FERPA”), CloudLabs ensures the complete protection of students’ data and privacy rights, making it a trusted and secure choice for educational institutions. - CloudLabs’ Automation Framework : Instructors get fully managed labs with a strong automation framework reducing workload and enabling a scalable certification program - A single lab template can be reused multiple times to replicate similar lab environments effortlessly. - Schedule lab duration after which all labs get deleted automatically. - CloudLabs’ auto-grading capabilities helps student assessment at scale. CloudLabs is a virtual training platform that strives to make IT training immersive with its hands-on application. However, navigating through cloud-based training can get tricky. To help instructors and school admins setup a certification program such as the AI for K12, CloudLabs pre-built AI-900 labs can help mitigate the various challenges with its customer-centric features allowing for a seamless integration with the school curriculum. Learn more about CloudLabs virtual labs and see how leading educational institutions are transforming their campus with Cloudlabs’ capabilities. Manesh Raveendran is the CEO and Founder of Spektra Systems. He is known for his thought leadership and for crafting end-to-end technology solutions in cloud marketplaces. His passion lies in harnessing the potential of public cloud partnerships to drive business growth. With over two decades of experience, Manesh collaborates with global CXOs, designing innovative cloud solutions and services to drive customer success and address critical business challenges.
<urn:uuid:4c046b0a-cdf2-4fd8-9a7b-144414d1153f>
CC-MAIN-2024-38
https://cloudlabs.ai/blog/complete-guide-for-schools-to-deliver-hands-on-ai-for-k-12-education/
2024-09-16T02:15:56Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651668.29/warc/CC-MAIN-20240916012328-20240916042328-00161.warc.gz
en
0.933435
3,061
3.5625
4
Banking bribery schemes, often hidden in the shadows of the global financial system, are illicit arrangements where individuals or entities offer, give, receive, or solicit something of value to influence the actions of bank officials or related parties. These schemes can range from small-scale frauds to massive operations involving billions of dollars, and they have far-reaching implications not only for the institutions involved but also for the global economy at large. The Anatomy of a Banking Bribery Scheme At its core, a banking bribery scheme typically involves two main parties: the ‘giver’ and the ‘receiver’. The ‘giver’ offers something of value, which could be money, gifts, or other favors, to the ‘receiver’ in exchange for some form of preferential treatment. This could be in the form of approving a loan, manipulating financial data, or overlooking certain regulatory checks. Common Elements in a Banking Bribery Scheme Element | Description | Bribe Giver | Often a corporate entity or individual seeking undue advantage. | Bribe Receiver | Typically a bank official or someone in a position of power within the financial institution. | Bribe | Can be money, gifts, holidays, or any other form of undue advantage. | Objective | The desired outcome, e.g., loan approval, regulatory leniency, etc. | The Global Impact of Banking Bribery The repercussions of these schemes are vast and varied: - Economic Distortions: They can lead to misallocation of resources, where decisions are made based on bribes rather than economic merit. This can result in bad loans, unwise investments, and ultimately financial crises. - Loss of Trust: Public trust in financial institutions diminishes, leading to reduced investments and economic downturns. - Regulatory Backlash: As these schemes come to light, regulatory bodies often respond with stricter regulations, which can increase the cost of compliance for all banks, even those not involved in bribery. The Significance of Addressing Banking Bribery Addressing and curbing banking bribery is of paramount importance for several reasons: - Economic Stability: A transparent and fair banking system is foundational for economic growth and stability. - Investor Confidence: Investors are more likely to invest in countries and institutions they perceive as clean and free from corruption. - Social Justice: Bribery often results in the rich getting richer at the expense of the general populace. Addressing it is a step towards a more equitable society. Comparison Table: Impact of Addressing vs. Ignoring Banking Bribery Aspect | Addressing Bribery | Ignoring Bribery | Economic Growth | Positive, as resources are allocated based on merit. | Stunted, due to misallocation of resources. | Public Trust | High, leading to more investments and economic activities. | Low, leading to economic downturns and reduced investments. | Regulatory Environment | Proactive, with regulations that are fair and not overly punitive. | Reactive, with potentially harsh regulations introduced in response to scandals. | Historical Context of Banking Bribery Schemes The history of banking is, unfortunately, intertwined with instances of bribery and corruption. From ancient civilizations to modern financial hubs, bribery schemes have evolved in complexity, scale, and impact. Understanding this historical context provides a clearer picture of the challenges faced today and the lessons learned from past misdeeds. Notorious Banking Bribery Cases Through the Ages The Medici Bank (15th Century) One of the earliest and most influential banks in the world, the Medici Bank of Renaissance Italy, was not immune to allegations of bribery. The Medici family, while known for their patronage of the arts, also used their financial power to influence political decisions, often through questionable means. The South Sea Bubble (1720) The South Sea Company, in an attempt to monopolize trade routes, was involved in one of the first major stock market crashes. Key to this disaster was the bribery of members of the British aristocracy and even the royal family to endorse and inflate the company’s value. BCCI Scandal (1991) The Bank of Credit and Commerce International (BCCI) was embroiled in a massive scandal involving money laundering, bribery, and support of terrorism. With operations in 78 countries and assets over $20 billion, its closure was a significant event in banking history. The 1MDB Scandal (2015) As mentioned earlier, the 1Malaysia Development Berhad (1MDB) scandal involved the misappropriation of billions of dollars, with funds being laundered through the global financial system. High-ranking officials in Malaysia and the United Arab Emirates were implicated, and the scandal had repercussions for banks and financial institutions worldwide. The Evolution of Bribery Schemes Ancient and Medieval Times. In ancient civilizations, bribery was often straightforward, involving direct payments to officials to overlook regulations or grant favors. Gold, land, and other tangible assets were the primary currencies of bribery. The Age of Exploration and Colonialism. As global trade routes expanded, so did the opportunities for bribery. Companies vied for monopolistic control over lucrative trade routes, often resorting to bribing local chieftains or colonial administrators. Industrial Age to Early 20th Century. The rise of industrial magnates and the expansion of global banking saw more sophisticated bribery schemes. Offshore accounts, shell companies, and complex financial instruments became tools for masking illicit transactions. Late 20th Century to Present. Modern bribery schemes often involve digital transactions, making them harder to trace. However, advancements in forensic accounting, international cooperation, and whistleblower protections have also improved the detection and prosecution of such activities. Evolution of Bribery Tools Over the Ages Era | Primary Tools of Bribery | Ancient Times | Gold, land, tangible assets | Age of Exploration | Trade monopolies, direct payments | Industrial Age | Offshore accounts, shell companies | Modern Era | Digital transactions, complex financial instruments | Mechanisms of Banking Bribery Banking bribery, a clandestine operation often buried deep within layers of financial transactions, is a sophisticated endeavor that requires intricate planning and execution. The mechanisms behind these schemes have evolved over time, adapting to the ever-changing landscape of global finance. Central to these operations are high-ranking officials and corporate insiders, whose positions of power and influence are exploited to facilitate these illicit activities. Orchestrating a Banking Bribery Scheme Identification of Targets The first step often involves identifying individuals or entities that can be influenced. This could be a bank official responsible for approving loans, a regulator overseeing banking operations, or even politicians who can sway financial policies. Establishing Communication Channels Discreet communication channels are established, ensuring that conversations and transactions remain hidden. This could involve encrypted messaging platforms, face-to-face meetings in undisclosed locations, or even using intermediaries to relay messages. The actual transfer of bribes has become increasingly sophisticated. Methods include: - Shell Companies: Entities created to disguise the true source and destination of funds. - Offshore Accounts: Bank accounts in jurisdictions with strict bank secrecy laws, making it difficult to trace the origins of funds. - Over-invoicing or Under-invoicing: Manipulating the price of goods or services to move money illicitly. - Hawala System: An informal and ancient system of transferring money based on trust, prevalent in some parts of the world. Layering and Integration To further obscure the origins of illicit funds, multiple transactions are conducted across various banks and countries, a process known as layering. Once sufficiently “distanced” from their source, these funds are then integrated into the legitimate financial system, appearing as legal assets. The Role of High-Ranking Officials and Corporate Insiders - Access to Critical Information: These individuals often have access to confidential information, giving them insights into regulatory checks, internal audits, and upcoming policy changes. This knowledge can be invaluable in planning and executing bribery schemes. - Influence Over Decisions: High-ranking officials can sway decisions in favor of those offering bribes. This could involve approving questionable loans, overlooking regulatory violations, or even influencing the drafting of financial policies. - Facilitation of Transactions: Corporate insiders, especially those in pivotal roles like treasury or compliance, can facilitate illicit transactions, ensuring they go unnoticed or are masked as legitimate. - Damage Control: In the event of suspicions or investigations, these individuals can manipulate internal reports, influence internal audits, or even stall external investigations, providing time to cover tracks or adjust strategies. Role of Officials in Bribery Schemes Role | Function | Impact | Bank Official | Approve loans, manipulate interest rates | Direct financial gains for bribe-givers | Regulator | Overlook violations, influence policies | Regulatory leniency, policy advantages | Corporate Insider | Facilitate illicit transactions, manipulate reports | Concealment of bribery, evasion of detection | Global Repercussions of Banking Bribery Schemes Banking bribery schemes, while often initiated by individuals or specific entities, have ripple effects that extend far beyond their immediate sphere of operation. These illicit activities can destabilize economies, strain international relations, and erode the public’s trust in financial institutions. The global implications of such schemes underscore the importance of understanding and addressing them.. Impact on Global Economies - Economic Distortions: Banking bribery can lead to a misallocation of resources. Decisions influenced by bribes, rather than economic merit, can result in bad loans, unwise investments, and even financial crises. Countries may see skewed economic growth, favoring sectors or industries that are involved in bribery over those that are not. - Reduced Foreign Investments: Countries perceived to have rampant banking bribery may become less attractive to foreign investors. The unpredictability and perceived risk associated with such environments can deter investments, slowing economic growth. - Currency Fluctuations: Large-scale bribery schemes can lead to significant capital flows in and out of countries, causing currency value fluctuations. This can impact trade balances, inflation rates, and overall economic stability. Strain on International Relations Diplomatic Tensions: Revelations of bribery involving officials from multiple countries can strain diplomatic ties. Countries may be accused of harboring or protecting individuals involved in bribery, leading to tensions or even sanctions. Trade Implications: Countries embroiled in bribery scandals may face trade restrictions or boycotts, impacting their export-import dynamics and potentially leading to trade deficits. International Legal Battles: Bribery schemes that span multiple countries can result in complex international legal battles. Extradition requests, asset seizures, and international arbitration can strain relations and resources. Undermining of Public Trust - Loss of Faith in Financial Institutions: When major banks or financial institutions are implicated in bribery schemes, the public’s trust in these entities diminishes. This can lead to reduced deposits, stock value drops, and even bank runs in extreme cases. - Skepticism Towards Regulatory Bodies: If regulatory bodies are perceived as being complicit or ineffective in preventing bribery, the public may lose faith in their ability to safeguard the financial system. - General Cynicism: Repeated instances of banking bribery can foster a sense of cynicism among the public, leading to decreased engagement in financial and civic activities. People may feel that corruption is inevitable, reducing their motivation to seek change or hold entities accountable. Global Repercussions of Banking Bribery Repercussion | Description | Potential Impact | Economic Distortions | Misallocation of resources due to bribery-influenced decisions | Financial crises, skewed economic growth | Diplomatic Tensions | Strained relations due to cross-border bribery incidents | Sanctions, reduced diplomatic cooperation | Loss of Public Trust | Eroded faith in banks and regulators | Bank runs, reduced civic engagement | Efforts to Combat Banking Bribery The pervasive nature of banking bribery schemes, coupled with their detrimental global repercussions, has necessitated concerted efforts to combat these illicit activities. From international collaborations to internal checks within financial institutions, various strategies have been employed to detect, prevent, and penalize bribery. Central to these efforts are international law enforcement agencies, robust internal controls, and the brave whistleblowers who bring such schemes to light. International Cooperation and Law Enforcement Initiatives Treaties and Conventions Several international treaties aim to combat corruption and bribery. Notable among these is the OECD Anti-Bribery Convention, which mandates member countries to criminalize bribery of foreign public officials in international business transactions. Joint Task Forces Countries often form joint task forces to investigate and combat cross-border bribery schemes. These task forces facilitate the sharing of intelligence, resources, and expertise. Asset Freezing and Recovery International cooperation has enabled the freezing and recovery of assets obtained through bribery. Organizations like the Stolen Asset Recovery Initiative (StAR), a partnership between the World Bank Group and the United Nations, assist countries in recovering stolen assets. Countries collaborate to extradite individuals implicated in bribery schemes, ensuring they face justice even if they flee their home country. The Role of Internal Controls - Compliance Programs: Financial institutions are increasingly investing in robust compliance programs that monitor transactions, identify red flags, and ensure adherence to anti-bribery laws. - Regular Audits: Regular internal and external audits help in detecting discrepancies that might indicate bribery or other illicit activities. These audits provide an additional layer of oversight and scrutiny. - Employee Training: Banks and financial institutions often conduct training programs to educate employees about the risks of bribery, the importance of ethical conduct, and the mechanisms to report suspicious activities. - Technology and AI: Advanced technologies, including artificial intelligence, are being employed to monitor and analyze vast amounts of transaction data, helping in the early detection of potential bribery schemes. Whistleblowers: The Unsung Heroes Reporting Mechanisms. Many institutions have established secure and anonymous reporting mechanisms, allowing whistleblowers to report bribery without fear of retaliation. Legal Protections. Countries are increasingly enacting laws to protect whistleblowers from harassment, job loss, and other forms of retaliation. Financial Incentives. In some jurisdictions, whistleblowers are offered financial rewards for providing information that leads to the successful prosecution of bribery cases. Key Strategies to Combat Banking Bribery Strategy | Description | Impact | International Treaties | Agreements between countries to criminalize and combat bribery | Standardized global approach to bribery | Compliance Programs | Internal systems to monitor and prevent bribery | Early detection and prevention of illicit activities | Whistleblower Protections | Mechanisms to protect and incentivize whistleblowers | Encourages reporting of bribery, leading to increased detection | The Future of Banking Integrity As the global financial landscape continues to evolve, so does the commitment to banking integrity. The scars of past bribery schemes serve as stark reminders of the need for vigilance, innovation, and proactive measures to safeguard the financial sector. Looking ahead, banks are adopting a multifaceted approach, leveraging technology, especially artificial intelligence (AI), to ensure a future where financial transactions are transparent, ethical, and devoid of corrupt practices. Proactive Measures by Banks - Enhanced Due Diligence: Banks are intensifying their due diligence processes, especially for high-risk clients and transactions. This involves a deeper scrutiny of clients’ backgrounds, sources of funds, and the nature of their transactions. - Strengthened Compliance Frameworks: Financial institutions are bolstering their compliance departments, ensuring they are equipped with the latest tools, training, and resources to monitor and enforce anti-bribery regulations. - Employee Ethics Training: Recognizing that the fight against bribery begins from within, banks are investing in regular ethics training for employees, emphasizing the importance of integrity and the consequences of bribery. - Collaborative Initiatives: Banks are increasingly collaborating, sharing information, best practices, and intelligence to collectively combat bribery schemes. Industry forums and associations play a pivotal role in facilitating such collaborations. In conclusion, in the intricate tapestry of the global financial landscape, two pillars stand out as paramount: transparency and integrity. The banking sector, the lifeblood of economies worldwide, thrives on the trust placed in it by millions. This trust, however, is not a given; it is earned through consistent demonstrations of transparency in operations and unwavering commitment to ethical practices. Reflecting on the myriad challenges the sector has faced, from high-profile bribery scandals to subtle internal compromises, it becomes evident that the path to integrity is fraught with temptations and pitfalls. Yet, it is this very journey, with its trials and tribulations, that underscores the importance of steadfast dedication to ethical principles. Every instance of corruption unveiled, every whistleblower’s revelation, and every regulatory intervention serves as a reminder of the delicate balance that needs to be maintained. The battle against corruption in the financial world is ongoing. It is a battle fought not just in courtrooms or boardrooms but in the daily decisions of bank employees, in the policies set by institutions, and in the vigilance of the public. As technology advances and the world becomes more interconnected, new challenges will undoubtedly arise. However, armed with lessons from the past and a collective commitment to a transparent future, the banking sector is well-equipped to face these challenges head-on.
<urn:uuid:c7dfd93d-23e6-4370-8946-277ea2fd0649>
CC-MAIN-2024-38
https://hextechsecurity.com/banking-bribery-schemes/
2024-09-17T09:27:05Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651750.27/warc/CC-MAIN-20240917072424-20240917102424-00061.warc.gz
en
0.925429
3,563
2.875
3
When it comes to cyber security statistics, most people would be horrified if they knew just how many cyber security attacks per day are attempted. Of course, not all are successful in breaching your information security defenses. Still, the cyber security statistics for 2020 demonstrate just how likely is a cyber attack against you and your organization. Table of Contents ToggleThe cyber security statistics by year show that cybercrime has been increasing ever since the stats were first recorded. The cyber security facts are that the number of cyber attacks per day continues to increase. Cyber attack statistics don’t give a true picture of how many hack attempts a day are successful because most organizations don’t like to publicize the weakness of their information security. So we don’t really know what percentage of successful cyber attacks happen. But we do have an indication of the number of attacks from the data collected by firewalls, anti-virus tools, IT security stats, and the providers of internet and network services. However, you can be sure that the cybersecurity facts will be that the rate of success is increasing in line with the number of attacks. This article will give you an insight into the state of cybercrime in the world today, looking at some of the statistics of hacking and other pertinent information security facts and figures for 2020. Here are some of the headlines for the global cyber war statistics in 2020. Most analysts expect 2021 to be worse. - Every 40 seconds, a new cyber attack starts. - Ransomware attacks are increasing at a rate of 400% year on year. - Over 25,000 different malicious applications are detected and blocked every day. - Each day hackers attack over 30,000 websites. - Over 65% of organizations worldwide have had at least one cyber attack against them. - Email is responsible for propagating 95% of all malware. - 43% of all cyber attacks are made on small businesses. The most illuminating of the cybercrime facts is that there is now a whole sector in IT that specializes in cyber attacks. Cyber crime is no longer just the preserve of curious and clever individuals sitting in their bedrooms. It is now an organized industry with big money behind it, some of it coming from nation-states and large corporations. To counter these cybercrime facts, information security and cyber security teams in many organizations have some of the largest budgets in IT, well over 50% of the total. Most governments are now heavily investing in developing and operating defenses against cyber crime, recognizing the economic risks of ignoring the threats. Some cyber security statistics Here are just a few of the cybersecurity stats for 2020 that demonstrate the scale of the challenge to information security: - 300,000 thousand new pieces of malware are created every day. These are designed to steal data and include spyware, adware, Trojans, and viruses. - Ransomware cost businesses a total of $20 billion in 2020. There were nearly 550,000 cyber attacks per day involving ransomware. The average amount demanded was nearly a quarter of a million dollars. - Between 9th March 2020 and 6th April 2020, hackers registered over 300,000 websites that used coronavirus-related keywords to lure victims. Many of these used web-spamming techniques to place them earlier in search engine results than non-malicious sites. - 70% of all data breaches globally are financially motivated. This is increasing year on year. Nearly all of today’s cyber criminals want to make money, and some make quite a lot of it. Each year the total amount generated by cybercrime is over $1.5 trillion. - 25% of all data breaches are motivated by espionage or stealing commercial information. - 2020 saw an increase in U.S. healthcare data breaches of 25% over 2019. Healthcare organizations reported 642 large data breaches. This is three times higher than the number reported in 2010. - By 2022 worldwide spending on just the cyber security part of information security is expected to reach $133 billion. - In 2020 over 20% of organizations worldwide have experienced an attack on Internet of Things (IoT) devices. There are now more than 8 billion connected IoT devices. As well as use in domestic appliances, they are also widely used in healthcare to provide real-time services to patients. - 63% of all data breaches in organizations are caused by compromised usernames and passwords. Surveys continue to highlight that in organizations without strong password control, users will use weak and predictable passwords such as “Letme1n” and “QWERTYUIOP.” Additionally, surveys have shown that over 70% of people use the same password on multiple platforms and websites, making it very easy for hackers to compromise information security across a number of systems after a successful attack on one of them. - The average time taken to detect a breach in information security is 7 months. Hackers today are skilled in developing and deploying malicious code that is difficult to detect. When coupled with the IT security facts that patching of systems and applications isn’t always promptly done and low-cost anti-virus solutions are often preferred against effective ones, this means that hackers can steal data for a long time. - In 2020 there were 23,000 denial of service (DoS or DDoS) attacks every 24 hours, disrupting the business of the target organization and attempting to extort money to stop the attack. Frequently asked questions: Hers are some questions that are often asked about the statistics of hacking attacks: How many victims of cyber crime are there each day? There are about 4000 cyber crime attacks every day in the U.S. alone. Globally 30,000 websites are hacked daily. What is the biggest hack in history? The largest data breach ever is believed to have been in 2013 when cybercriminals stole the credentials of over 3 billion Yahoo account holders. The company revealed in 2017 that the information security for every one of their customers had been breached, including users of the Tumblr and Flickr photo storage services. What is the most dangerous cyber attack? The most dangerous cyber attacks tend to be on governments and healthcare organizations. In 1999 when Jonathan James hacked into the US Department of Defense and NASA, which compromised national security. He stole software worth $1.7 million and made the agency shut down for weeks. In 2017 a cyber attack using the Wannacry ransomware virus breached the information security of the National Health Service in England for days, infected over 30,000 computers, and resulting in canceled patient appointments and a huge financial bill for restoring the hacked systems. Which organizations are attacked the most by ransomware? According to cyber attack stats, healthcare, financial, government, and retail organizations are most targeted by ransomware attacks. How many people are affected by cyber crimes? In the U.S. approximately 90,000 people per month are affected by cyber crimes. What is the most dangerous cyber crime? Ransomware is probably the most dangerous of the cyber attacks. Because ransomware can be easily delivered using seemingly innocent emails, it represents a very high risk to every organization. If the hacker manages to encrypt the victim’s servers, data storage, and backup devices, it can be almost impossible to restore them. This would stop the organization from operating until the ransom is paid and the hacker releases the locked devices. Paying the ransom could easily wipe out the company’s finances, and it is increasingly difficult to get any business insurance that covers this type of loss. There is also no guarantee that the hacker will unlock the systems after payment has been made. The hacked organization could easily go out of business after an attack. If the target is a government, the impact could be massive, preventing key services from operating for some time. What country gets hacked the most? The U.S currently has the highest rate of cyber crime, over 23.6% more than any other country. The country that is most vulnerable to cyber attack is Indonesia, followed by China and Thailand. It is thought that more than 100 of the Thai government websites have been hacked and that 85% of these are used by hackers to distribute malware around the world. Which country has the best cyber security? The cyber security statistics 2020 show that the country believed to have the best cybersecurity is Denmark. This is based on a survey that analyzed the percentage of devices infected with malware, the number of successful attacks, and the robustness of cybersecurity and information security legislation. Other top-performing countries include Sweden, Germany, Ireland, and Japan. Which country has the worst cyber security? The country with the worst level of cyber security according to cyber security stats 2020, is Algeria. It is the country with the poorest cyber security legislation and is not well prepared for cyber attacks. Other countries with low levels of cybersecurity include Tajikstan, Syria, and Iran. Cyber security is often seen as something that only big organizations need to do something about. Governments, financial institutions, and major technology companies have been managing cyber security for some time. Today, every organization or individual who uses IT in any form is at risk from cyber criminals and needs good cyber security. The hackers want your data. The hackers want your money. They know ways to get past poor information security and cyber security defenses. The threats are real; the hackers are out there probing systems every minute of every day. Defenses that you put in place a few years ago and haven’t updated since are now useless. The cyber security statistics 2020 show the scale of the problem last year. This year it will only get worse, so if you want to maintain the security of your data and information, do something now about your cyber security , before it’s too late.
<urn:uuid:43f0ab30-0f39-4a06-b32f-fae11c863b66>
CC-MAIN-2024-38
https://itchronicles.com/information-security/cyber-security-statistics-2020/
2024-09-18T10:55:58Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651895.12/warc/CC-MAIN-20240918100941-20240918130941-00861.warc.gz
en
0.956702
1,990
2.53125
3
What are patches? Patches are not actually threats; they are add-ons designed by software creators to fix security vulnerabilities and other known bugs. The idea is that these software creators catch vulnerabilities on their own and patch them before they are exploited, but often an attack by a hacker is the signal that alerts software-makers of security flaws. How to Recognize This Threat: The threat in this case is not keeping your software updated with the latest patches. Hackers often look to exploit known vulnerabilities, banking on the fact that many do not update their software in a timely manner (or at all). How to Prevent This Threat: Major software developers have regular updates scheduled and will often alert users directly on their devices. Admins should stay abreast of these updates by subscribing to vulnerability mailing lists as well, since some patches are needed sooner than scheduled updates, and should also set policies to ensure all networked computers have the latest patches. A Managed Security Service subscription from Firewalls.com will help you with your network configuration and ensuring the latest updates are coming through, leaving those nitty-gritty details to us!
<urn:uuid:c4e0e1fd-3d97-4d19-b85d-734d38160e82>
CC-MAIN-2024-38
https://www.firewalls.com/blog/security-terms/patches/
2024-09-20T23:26:46Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725701425385.95/warc/CC-MAIN-20240920222945-20240921012945-00661.warc.gz
en
0.959423
230
2.84375
3
A cyber crime inevitably engages a computing device and a network where the computer can be used as the weapon of the crime or it can be the target as well. As a definition of crime, “Offences that are committed against individuals or groups of individuals with a criminal motive to intentionally harm the reputation of the victim or cause physical or mental harm, or loss, to the victim directly or indirectly, using modern telecommunication networks such as Internet (networks including but not limited to Chat rooms, emails, notice boards and groups) and mobile phones (Bluetooth/ SMS/ MMS)." It also endangers the national security along with the personal. Issues like hacking, unwarranted mass-surveillance, copyright infringement, child pornography have become the raving problems in the sphere. Also a violation of woman’s dignity through cyber space has stroked a higher rate than ever. White Paper: The Importance of Gamification in Cybersecurity A cyber criminal is an individual entity who takes advantage of a computer and network system or makes it a target or sometimes both in order to commit cyber crime. There are generally three ways that a cyber criminal works- The criminal targets a computer system for malicious activities like data theft, identity theft, virus spread and so on. The criminal owns a personal computer to commit crimes like fraudulent activities, spamming and many more. The criminal saves the crime documents like illegal data and in the computer and uses it as an accessory. Criminals often work in groups as much like any other crime scenes. For cyber based crimes it sometimes needs a group of people who can work as the followings- Programmers who can write coded programs for the organization s/he is working. Distributors who can sell and distribute stolen data, information or goods that has been collected from associated cybercriminals. Hackers who can track and massacre networks or systems. Fraudsters to make plots for phishing and spamming. System providers who hosts malicious and illegal sites. Cashiers who provide names and account details to criminals and manage the drop accounts. Money mules that manage the bank accounts and keep eye on the wire transfers. Tellers that launders with illegal money via digital and foreign exchange methods. And lastly leaders who are in connection with the criminal organizations. They generally lack in technical knowledge and send teams to do a task. According to IPC section 43 of ‘Information technology Act, 2000’, whoever deletes or destroys or alters or causes disruption of any computer with the intention of damaging the whole data of the computer system without permission from the owner shall be liable to pay fine upto 1 crore to the person affected. Also according to section 43a of ‘Information Technology (Amendment) Act, 2008’, if a body corporate is liable for maintaining the protection for a person’s data and is accused of negligence of the same or failure in protection, the body corporate is accountable to pay compensation to the affected. According to section 66 that deals with hacking enables a criminal with 3 years of imprisonment or fine. There are other penalties as well for different kind of cyber crimes. White Paper: Building Cyber Security and Data Management Awareness White Paper By: Qualys A complete Vulnerability Management (VM) solution can monitor your environment, enabling you to discover devices running in your network, and determine whether they are vulnerable to attack. The VM solution helps you to find fixes to the underlying problems, and protect yourself while those fixes are being implemented. This whitepaper provides insights on the best practices that will save... White Paper By: Circadence Top-notch cybersecurity is an essential part of our world. Data breaches can bring a business to its knees in the blink of an eye. Gamification has a tremendous opportunity to revolutionize the speed, efficacy and relevancy of training in the quickly evolving landscape of the Cybersecurity sector. Cybersecurity awareness trainings are usually a boring affair, by training... White Paper By: Circadence Cyber‐attacks and threats against the financial services sector are ongoing – common targets include banks, payment processing companies, investment firms, and other organizations that manage financial transactions. A 2016 study reported that 83% of financial services companies cite defending against cyber threats and protecting personal data as one of their biggest challenges in... White Paper By: Lastline Today’s sophisticated malware is a major culprit in many of the rampant cybersecurity incidents. Unfortunately for organizations, advanced malware is getting harder to detect. Malware assaults are so common that many IT managers admit that their enterprise networks are likely to experience a cyberattack at some point because their conventional security systems cannot effectively... White Paper By: SPECOPS If you have identified the need for a self-service password reset solution, you are likely familiar with the cost-savings, usability, and security benefits. The next step is identifying the criteria to use when evaluating the desired outcome of your investment. A self-service password reset and management solution not only reduces the number of help desk password tickets but also... White Paper By: SPECOPS Passwords are the thin layer protecting our personal information from the “unknown.” A few key proactive measures can make that layer impenetrable. Shifting some burden from individuals to password policies that promote stronger passwords is the next logical step. A proactive password security approach can go a long way for both organizations and users. Password security is a...
<urn:uuid:df0ed5e3-79ba-45ff-9f66-3e8b6d872b1f>
CC-MAIN-2024-38
https://whatis.ciowhitepapersreview.com/definition/cyber-criminal/
2024-09-07T17:22:44Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650898.24/warc/CC-MAIN-20240907162417-20240907192417-00061.warc.gz
en
0.935362
1,121
3.296875
3
DAX, or Data Analysis Expressions, is a vital language used in Power BI, Excel, and other Microsoft BI tools for data modeling and reporting. One of the more advanced features of DAX are window functions. These functions allow users to perform calculations across sets of rows that are related to the current row. Apply semantics, introduced in the enhanced versions of DAX, enable more nuanced and efficient computations in these contexts. Apply semantics effectively modify how table expressions are computed by allowing for complex manipulations that encompass multiple rows within DAX window functions. This modification enhances the capability of DAX to handle intricate scenarios in business intelligence applications. For those new to DAX or seeking to deepen their expertise, numerous resources are available. These include comprehensive guides and definitive books that lay out foundational and advanced concepts. Learning DAX through these resources can significantly boost one's analytical skills, enabling better data handling and reporting in various business environments. Understanding the use of semantics in DAX window functions is essential for anyone working with DAX in data modeling. The technique involves a novel method of computing table expressions when selecting multiple rows. This allows for more precise and efficient data analysis. DAX, or Data Analysis Expressions, is a formula language used extensively in different Microsoft products such as Microsoft 365, SQL Server Analysis Services, and Power Pivot in Excel. Mastery of DAX can dramatically improve one's ability to handle complex data operations in these platforms. Therefore, learning DAX through structured guides can significantly enhance your proficiency. For those eager to delve deeper into DAX, resources such as books and online tutorials can be invaluable. They serve as definitive guides, offering detailed insights and practical examples to help sharpen your skills in data analysis using DAX. Investing time in these resources can greatly advance your understanding and application of this powerful tool. Within DAX, WINDOW functions initially identify all columns that lack corresponding outer columns. For each unique combination of existing values in these columns within WINDOW's parent context, the function evaluates and returns the corresponding rows. The final output from WINDOW is the union of these rows. In DAX, the AND function restricts its input to two arguments. For scenarios requiring an AND operation on multiple expressions, you could chain several calculations together, or more efficiently, utilize the AND operator (&&), which enables a more streamlined expression by combining multiple conditions. Data Analysis Expressions (DAX) encompass a suite of formulas, functions, operators, and constants that empower users to create measures, dimensions, and custom tables. These expressions return one or more values, addressing complex data analysis challenges through new relational dynamics among different data variables. The DAX function, CALCULATE, is pivotal as it evaluates an expression within a specifically modified filter context. Furthermore, CALCULATETABLE serves a similar role but tailored to scenarios involving tables. Understanding Semantics, Window Functions DAX, Apply Semantics DAX, DAX Window Functions, Semantic Analysis DAX, Implementing Window Functions, DAX Query Optimization, Advanced DAX Techniques
<urn:uuid:08ff034b-3a26-4121-893a-0dce553221c2>
CC-MAIN-2024-38
https://www.hubsite365.com/en-ww/crm-pages/understanding-apply-semantics-for-window-functions-in-dax.htm
2024-09-08T20:03:27Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651017.43/warc/CC-MAIN-20240908181815-20240908211815-00861.warc.gz
en
0.856356
627
3.1875
3
Kefreen Batista, Vice President of Technologies at Globant, discusses how robotics can transform operations for the better. Robotics is transforming operational efficiency in a variety of industries, paving the way for a more automated and efficient future. When we talk about this subject, it is very common for previous generations to remember Rosie, from The Jetsons, a classic animation of the twentieth century. The character is characterized as a humanoid who did all the household chores and related to the futuristic family in the cartoon, as if he were a real human being. Although technology is not as autonomous as it is, full of mannerisms, ideas and attitudes, some of these aspects are similar to what we live today, only in a slightly different way. To contextualize this observation, two characteristics are highlighted to define the functions of robotics: that of reducing risks and increasing the scalability of processes. These systems help develop processes with the objective of mitigating risks of process criticality due to human error, in addition to developing activities that go beyond our capacity, with more precision and at scale. Robotics can be applied in various sectors. In the health area, for example, where automation helps a person with a chronic disease to remember to take medicine every day through a simple application or technology operating to support the doctor in decision-making during a consultation. In manufacturing, robots automate production lines, increasing efficiency and reducing errors. While in the logistics sector, robotic systems optimize the movement of goods, ensuring faster and more accurate deliveries. Finally, in customer service, chatbots and virtual assistants provide immediate support, improving the user experience. Within what we conceptualize as robotics, the goal is always to generate an impact on results in the face of the complexity and feasibility of the processes. With the support of AI, systems can be operated through large-scale language models, the LLM, as well as smaller and leaner models, the SLM, which are focused on deeper knowledge of each sector of a company, thus enabling the naturalness of the interaction between human and machine in problem solving. The success of these services, however, is evaluated from the point where the user does not remember that he is using technology in operations, ensuring maximum naturalness in these processes. In addition, it is important that ethics and transparency are guaranteed, which can be defined according to agreement between the participants, and that strict policies are applied to ensure that all technological solutions respect the privacy and rights of users are adopted. As we dive into the technological revolution, it is clear that the junction between AI and robotics is paving a path towards efficiency, safety and ethics converging in harmony. Even if this current model is not a ‘Rosie’, which solves problems with such a personality, we are on a positive technological path that brings effective solutions to our daily lives. This rise still promises significant transformations, with profound impacts that optimize several sectors of the market. However, these systems must be operated with data transparency, ensuring that this progress is underpinned by trust in and respect for core human values. Click below to share this article
<urn:uuid:6fb41b2b-7347-4fdd-ad3c-2f70d4d482df>
CC-MAIN-2024-38
https://www.intelligenttechchannels.com/latam/2024/08/28/what-should-the-role-of-robotics-be-to-enhance-operational-efficiency/
2024-09-08T19:25:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651017.43/warc/CC-MAIN-20240908181815-20240908211815-00861.warc.gz
en
0.954628
632
2.796875
3
If you started implementing business continuity management, probably the biggest challenge you are facing is writing the business continuity plans. Why is it so difficult? Well, you have to think of various scenarios under which a disaster (or other kind of disruption of business activities) can occur, and you have to think of a way how to handle such exceptionally rare but potentially catastrophic incidents. The problems that people who write such plans usually have include what the plan should contain (what are the main elements), how long (how detailed) it should be, what steps to include etc. One of the best solutions to all these dilemmas is using the BS 25999-2 standard, which together with BS 25999-1 defines a framework as to how the plans should be written. According to those standards, the business continuity plans should consist of (1) incident response plan, and (2) recovery plans. An incident response plan is usually a single plan written for the whole organization, and describes what has to be done immediately after a disaster occurs – reducing the effects of the incident, communicating to emergency services, evacuating the building, gathering at assembly points, organizing transport to alternative locations etc. Recovery plans are usually written separately for each critical activity, and the steps to be included in the recovery plans are usually the following: when and how to communicate with various stakeholders (employees and their families, shareholders, customers, partners, government bodies, public media etc.), how to assemble the team, how to recover the infrastructure, how to check whether the applications are functioning and whether the access rights are appropriate, how to check which data is missing or has been corrupted by the disaster, how to recover the data, and how to decide when the recovery is completed so that normal operations can begin. Disaster recovery plans (the recovery plans of ICT infrastructure) are the ones to be written with great care because they should describe how to set each system running within the recovery time objective of a particular critical activity. This is usually done by writing a detailed recovery plan for each system to be recovered. The rule of the thumb says that the level of details in all these plans should be such that other employees (or external staff) should be able to execute the plan if the people working with that critical activity are not available. Therefore, use common sense when writing the plans – they should be understandable to anyone, not just you. In my experience, the biggest challenge when writing these plans is that employees have to face something completely different, something they never had to think about. To overcome such a problem it is best to organize a workshop where, with or without a moderator, they could share their views about what would happen if… , how to react when…, etc. The truth is, the mere fact that your employees have started thinking about business continuity is 50% of the job done – with such an approach, the results of business continuity planning will be much better. This free webinar will also help you: Writing a business continuity plan according to ISO 22301.
<urn:uuid:d7dc6368-5593-40a7-a463-27414beb6b73>
CC-MAIN-2024-38
https://advisera.com/27001academy/blog/2010/04/08/how-to-write-business-continuity-plans/
2024-09-13T17:50:51Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651535.66/warc/CC-MAIN-20240913165920-20240913195920-00461.warc.gz
en
0.964863
621
2.859375
3
What is a Monitor? A Monitor is an IT infrastructure component discovered and subsequently provisioned in Motadata for the purpose of monitoring. The monitor facilitates judging the real-time state of the system and generates insights into its performance. At the application level, a monitor is any entity that is provisioned by a Motadata user after running a discovery. Motadata keeps track of the various performance metrics for an IT infrastructure at the monitor level. The data polling for the metrics is done by Motadata at the monitor level. The analysis and the insights provided by Motadata via any of its multiple features, be it, Topology, Metric explorer, Log explorer, or Alerts and Policies is done by using the values of the metrics collected at the monitor level. Let us understand how a monitor is created in Motadata. - A user creates a credential profile and a discovery profile and maps the discovery profile to the credential profile. - The user then starts a discovery run to discover the devices from a network by executing the discovery profile they created. - The user then goes on to provision the discovered devices so that Motadata can keep track of these devices to ‘monitor’ their performance. - These provisioned devices that are ‘monitored’ by Motadata are known as Monitors.
<urn:uuid:8be8eae7-f471-4268-bd4c-69830fa17804>
CC-MAIN-2024-38
https://www.motadata.com/documentations/what-is-a-monitor/
2024-09-13T17:26:13Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651535.66/warc/CC-MAIN-20240913165920-20240913195920-00461.warc.gz
en
0.936547
268
2.515625
3
By Maila Zahra, Air University Islamabad and Zia Muhammad, North Dakota State University Post-quantum cryptography aims to develop secure cryptographic algorithms to protect against most quantum attacks. The threats of quantum computers to the current cryptographic system will lead to the development of more secure algorithms schemes to ensure that the communication between channels remains secure and the ability to perform different efficient tasks. The key to sustainable cryptography is to adopt new algorithms that are powerful, resilient, and secure. Security and Future of Post-Quantum Cryptography There are a few noticeable questions that arise in our mind, particularly when we think of post-quantum future perspectives. What impact would post-quantum have on departments such as government, finance, and healthcare? How does it impact future security protocols, data protection, and privacy? How will it implement in real-world systems? What happens if a large number of post-quantum computers will use? Post-quantum cryptography has a significant impact on these departmental elements like the government, as it does depend upon secure communication during the exchange of sensitive data over the channel. Implementing post-quantum cryptography will protect the potential attacks on critical infrastructure. Now a day’s, in finance, crypto-currencies such as bitcoins are widely in use, and post-quantum assures the security of these digital currencies. Post-quantum will protect and secure patients’ sensitive data, medical devices, and new research in medical fields. How Post-Quantum Cryptography Works and Why It Matters Quantum computing has been proven effective to breaks the classical cryptographic components because it can solve factoring problems and break complex algorithms. To defend us against such attacks, post-quantum cryptography came into existence. It utilizes complex mathematical frameworks to resist these quantum attacks and ensure the security of underlined algorithms. It ensures that sensitive information remains protected. Moreover, it strengthens the public and private keys, symmetric encryption, and all models that can be used to mitigate the risk of attacks. post-quantum cryptography has paramount importance, and its significance can not be denied. In the current interconnected digital world, there is an uncontrolled reliance on technology and the need for online communication and data is key to security and privacy. Researchers are dedicating their efforts to building highly complex mathematical algorithms and key-sharing schemes that can sustain quantum attacks effectively. Build algorithms are tested against various attacks and scenarios to ensure the reliance on both classical and quantum cryptography, ultimately paving the way forward toward development and secure cryptographic solution. Challenges and Limitations of Post-Quantum Cryptography Despite the benefits and importance of post-quantum cryptography, there are also some drawbacks and difficulties that require further research and development. Some of the challenges include the complexities of keys, the overhead of key processing, and the underlined legacy infrastructure. These factors can affect the efficiency and performance of post-quantum cryptography, as well as its adoption and implementation in real-world systems. One of the important aspects of post-quantum cryptography is the deployment in underlined legacy systems and making it compatible with existing protocols. These newly developed post-quantum cryptographic schemes and protocols need compatibility solutions with real-world systems and existing applications. There would be a need to focus on the choice of algorithms, the design of protocols, the evaluation of performance, and the verification of strength. As with the development of every new technology, there are unintended consequences. For example, in terms of quantum computing, it is a new threat to underlined crypto infrastructure. But this is how we evolve, innovate and come up with more advanced novel solutions. This is how post-quantum cryptography was born to promote security and privacy in the presence of quantum computers. The post-quantum cryptography holds immense potential to safeguard the security of our digital future against any quantum threats. Therefore, there is a need to dedicate efforts to strengthen the current infrastructure with post-quantum cryptography. About the Author Maila Zahra is a student of cybersecurity at Air University, Islamabad, Pakistan. She also works for the National Science and Technology Park and actively participates in national-level hackathons. Her research interests include quantum computing, post-quantum cryptography, and various cutting-edge topics in the field of cybersecurity. Zia Muhammad is a Ph.D. scholar at the Department of Computer Science, North Dakota State University (NDSU). He is a cybersecurity professional, academician, and researcher who has taken professional training and certifications. He has authored several publications in peer-reviewed conferences and journals in the field of cybersecurity. Zia can be reached online at ([email protected], https://www.linkedin.com/in/zianoedar/).
<urn:uuid:da10be02-f8fd-473b-9d05-df432d759a97>
CC-MAIN-2024-38
https://www.cyberdefensemagazine.com/post-quantum-cryptography-safeguarding-the-digital-future-and-bolstering-security-in-critical-sectors/
2024-09-15T00:08:39Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651601.85/warc/CC-MAIN-20240914225323-20240915015323-00361.warc.gz
en
0.937435
960
3.265625
3
Let’s get right to it. If you are unaware of Bitcoin, AceHost can educate you on what it is. Bitcoin is a decentralized digital currency, without a central bank or single administrator, that can be sent from user to user on the peer-to-peer bitcoin network without the need for intermediaries. What is Bitcoin Mining? Bitcoin Mining is performed by high-powered computers that solve complex computational math problems. The problems are so complex they cannot be solved by hand and are so complicated to tax even incredibly powerful computers. In simpler form, Bitcoin mining is the process of creating new bitcoins by solving a computational puzzle. To take it a step further, if you have been mining Bitcoin and turning out a profit, you probably don’t keep the miners in your living space. However, did you know that you can store miners in a colocation facility? AceHost is going to dive deeper into a colocation facility and give you the information you need. Bitcoin Miner Hosting and Colocation When hosting your servers, the colocation center is used to host ASIC miners. These centers guarantee that everything is properly set up and remove any worries from the server owner. Colocation is when individuals bring their own server to a data center. These data centers are ideal for ASIC mining Bitcoins. Colocations are convenient and offer more space. The facilities have racks and cabinets in which servers can be placed and each cabinet has a connection to the internet power. Additionally, you can always feel at ease because colocation centers have strong security. AceHost is a colocation center. ASICs and GPUs What are ASICs? Is the acronym for “application-specific integrated circuit.” It is a type of circuit that has been designed for a single specific purpose. An ASIC miner refers to a device that uses microprocessors for the sole purpose of mining digital currency. A key takeaway is ASIC is specialized for bitcoin or cryptocurrencies. Bitcoin miners review and verify previous bitcoin transactions and create new blocks so that the data can be added to the blockchain. There are other devices you can use to mine bitcoin like GPU. GPU stands for “Graphical Processing Unit.” They have a higher processing speed than Central processing units (CPUs). GPUs are responsible for digital rendering in a computer system. GPUs are more useful in blockchain mining due to their high speed and efficiency. In fact, GPUs mine 800 times faster than CPUs. Is Colocation Right for You? Bitcoin’s popularity is beginning to increase, which means the number of companies offering colocation is increasing too. When looking for a colocation you may want to consider: - If you want specialist services - If you are willing to pay other people to eliminate the hassle from your life - If you seek redundancy - If you want safety and security These are the four factors to consider if you are considering using a colocation center. If you decide to contact a data center provider there are also a few things to understand before going to a colocation center. - What are your power requirements? - How many kilowatts do you need? - What is the amount of space you need in terms of racks? - What budget do you have designated for colocation? - How much money can you spend and remain profitable? Colocation facilities are great if you are invested in Bitcoin mining and bitcoin miner hosting. AceHost is available to be the safe center for you and help you with your bitcoin needs. AceHost will be upfront with costs and help answer all of your questions.
<urn:uuid:8edde733-100b-4089-ad3d-d8f0fbf64d43>
CC-MAIN-2024-38
https://blog.acehost.com/uncategorized/what-does-bitcoin-miner-hosting-mean/
2024-09-16T05:16:00Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651676.3/warc/CC-MAIN-20240916044225-20240916074225-00261.warc.gz
en
0.948082
739
2.765625
3
Table of Contents As the pace of technological advancement accelerates, staying updated on the emerging trends in software development are crucial for developers and tech freaks. From artificial intelligence to Quantum Computing, these technological developments have just started realizing new avenues for innovation. In this article, we will discuss the top five emerging technologies shaping software development’s future. What’s New in the World of Software Development? Emerging trends and technologies are the most upcoming wave of progress in software development. Technologies hold the keys to understanding these for any firm to remain relevant and ahead of the evolving tech landscape. With new methodologies springing up every day, keeping up can become really difficult, and using a custom software development company can produce fast, reliable results. Top 5 Trends Transforming Software Development 1. Artificial Intelligence & Machine Learning Artificial Intelligence and Machine Learning have become an intrinsic part of many software applications. AI systems may imitate human intelligence in performing tasks, while ML algorithms allow software to learn from its data and improve over some time. Some of the use cases of AI & ML are: - Predictive Coding: AI algorithms can help in predicting or preventing potential coding errors or improving existing ones. - Automated Testing: With the help of ML-driven tools, test cases can be created and executed automatically, thereby improving software quality. 2. Quantum Computing Quantum computing is the paradigm shift from classical computing. It utilizes quantum bits or qubits to process complex computations efficiently. Quantum computers shall perform intricate problems—mainly cryptographic challenges and optimization problems—by orders more efficiently than classical computers. New programming languages and frameworks are being developed that shall harness the quantum algorithms. IBM and Google are leading in quantum research. 3. Blockchain Technology Blockchain technology ensures secure and transparent transactions through decentralized ledgers. Some areas that blockchain technology is projected to affect include supply chain management, digital identity verification, and DeFi. Some of the blockchain technology use cases are: - Smart Contracts: It involves automated self-executable contracts that enforce an agreement without the need for intermediaries. - Decentralized Apps (dApps): dApps create applications which run on peer-to-peer networks thereby making them more secure as opposed to relying on central servers 4. 5G Technology 5G is the next generation of mobile network technology. It is much faster with less latency than its predecessors. Many companies are exploring 5G for a diverse array of applications from smart cities to sophisticated telemedicine. Some advanced implications for software development that make this technology important for 5G are: - Improved Data Transfer: It provides real-time data processing and flawless connectivity to IoT devices. - New Opportunities: It allows breakthroughs in AR and VR apps. 5. Edge Computing Edge computing refers to a computational model that allows data processing and storage to be relocated right at the source or closer to it. It thereby reduces latency and bandwidth usage by handling data locally rather than in centralized cloud data centers. Thus, the key advantages of Edge computing are: - Low Latency: Critical for applications requiring immediate data processing, such as autonomous vehicles and industrial automation. - Real-Time Data Processing: This will improve performance for IoT devices and real-time analytics. Edge computing is being used in diverse sectors from manufacturing and healthcare to smart infrastructure. How These Technologies Affect Software Development Understanding these technologies and their combination can make a huge difference in one’s mindset about how to solve problems and drive innovation in products. Careers in artificial intelligence, quantum computing, blockchain, 5G, and edge computing will be some of the most demanded sets of skills and hence specialization and career growth will be promising. Each technology augments the possibilities of learning as following: - AI & ML: Familiarity with data science concepts and tools - Quantum Computing: New programming paradigms, quantum algorithms. - Blockchain technology: This helps us learn about smart contract development, dApps, and blockchain. - 5G technology: This will help us learn about its impact on networked applications and IoT. - Edge Computing: The technology helps to learn about distributed computing and real-time data processing. Real-World Use Cases These are just some of the real-world use cases of emerging technologies across multiple industries that showcase a roadmap to future innovations. - AI-powered insights are changing medical diagnostics with IBM’s Watson. - Google’s Sycamore processor created quantum supremacy, powering complex problems at speeds unthinkable to classical computers. - The smart contracts of Ethereum are making a difference in DeFi applications. Staying Ahead in the Fast-Paced World of Tech Rapidly changing technology requires continuous learning and adaptation. Grasp these emerging technologies, and this will push you to the forefront among software development companies—not only improving your skills but also moving you to the leading edge of progress. ABOUT THE AUTHOR IPwithease is aimed at sharing knowledge across varied domains like Network, Security, Virtualization, Software, Wireless, etc.
<urn:uuid:7777b0dd-16de-4c25-ad65-93bd270e2bbc>
CC-MAIN-2024-38
https://ipwithease.com/top-5-emerging-trends-in-software-development/
2024-09-16T05:55:56Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651676.3/warc/CC-MAIN-20240916044225-20240916074225-00261.warc.gz
en
0.897718
1,041
2.71875
3
What is penetration testing, why is pen testing important and what are the benefits of penetration testing for your business? Discover the critical penetration testing benefits that safeguard your business. Understanding these benefits is key to reinforcing your cybersecurity posture. What is penetration testing and why is penetration testing important? Penetration testing, also known in the industry as pen testing, is a form of ethical hacking. Penetration testing uses simulated cyberattack methods against your networks, applications and computer systems to identify potential vulnerabilities or weaknesses. Penetration tests look to uncover these vulnerabilities to understand the company’s risk of a cyberattack or security breach, including: - Where a cybercriminal may attempt to access the organisation’s networks - How they would gain access to systems - What defences are already in place and how they cope under attack - The potential impact of a breach on the business Read our complete guide to penetration testing. Not all penetration testing services are equal. Penetration testing shouldn’t be confused with vulnerability scanning or vulnerability assessments. A penetration test provides a much more comprehensive report of the company’s security posture, including how networks, applications and systems could be hacked. Many organisations choose CREST accredited penetration testing providers to perform pen tests on their infrastructure or applications. Find out about the benefit of penetration testing under the CREST accreditation and why CREST penetration testing is the gold standard in the cybersecurity industry. Penetration testing is important to understand any company’s digital infrastructure and to identify any potential threats. The insights provided by penetration testing can be used to improve company security policies and address any vulnerabilities with patching and other remediation. It allows the business to take action before a breach occurs. A penetration test should be thought of in the same way as a financial audit for your business. But instead of looking at invoices and accounts, penetration testing (or pen testing) is a security exercise where a cyber security expert attempts to find and exploit vulnerabilities. Businesses are recommended to perform a penetration test whenever the following happens: - You discover a potential new security threat - You create or update an application - You relocate offices or migrate networks or move to adopt remote or hybrid working - You create a new database or data storage site - You have recently been attacked What are the categories of penetration testing? There are some common penetration testing services available to assess security risks across networks, applications and computer systems, whether internal or external. These are designed to diagnose any security risks or weaknesses in your infrastructure before cybercriminals or hackers can exploit them. Businesses are building more of their frameworks online, especially in the acceleration of digital transformation seen in the past few years. Unfortunately, while this can bring many business continuity and flexibility benefits for organisations, it also expands the attack surface, making them more susceptible to attack. More than 60% of internet-based cyberattacks are aimed at web applications. So, what is application penetration testing? Web application penetration testing will find any weaknesses in the target system. The benefit of penetration testing your applications is that the pen tester can also check the functionality of websites to pinpoint any failings. A web application penetration testing service will supply the protection needed to safeguard sensitive data. Regular web application penetration testing will defend against every conceivable online threat. With the increase in mobile use and mobile devices now a major part of our everyday lives, organisations must take necessary action to secure their mobile applications and protect the business, its reputation and most importantly, its customers. A mobile application penetration test will look for a range of exploitable vulnerabilities that cybercriminals may take advantage of. Disgruntled employees, or negligent staff members, can fall prey to phishing attacks and pose a security risk. An internal security breach could prove disastrous for any business. The benefits of penetration testing of internal environments, systems and procedures mean that businesses can ensure they have all the right countermeasures in place to prevent unauthorised access to privileged information. Internal network penetration testing is designed to simulate a cyberattack from within the organisation itself, highlighting potential issues and safeguarding against threats from malicious insiders. By mimicking real-world cyberattacks, the benefit of penetration testing on external networks means businesses can identify any gaps in external network infrastructure to allow the necessary remediations. Using the same techniques that a hacker would, pen testers – or ethical hackers – conduct external network penetration testing to simulate a real-world attack and understand if data is secure. On completion of the external penetration testing, pen testers issue a comprehensive report. Using this information, any security flaws can be addressed, eliminating potential threats before they can cause damage. What are the penetration testing methods? There are three different types of penetration testing; Black box penetration testing, white box penetration testing and grey box penetration testing. The benefits of penetration using each of these types are the same – to attempt to gain access to an organisation’s networks, computer systems, software or applications using the same methods as an attacker might use in order to exploit any weaknesses or vulnerabilities. Penetration testing can also follow different methodologies. Standards such as The OWASP (Open Web Application Security Project®) Top 10 outline the most critical security risks to web applications. Following this standard, the penetration tester can identify common risks and vulnerabilities. As mentioned, the CREST penetration testing method is universal and highly respected. CREST set a strict code of conduct around preparation and scoping best practices, penetration testing execution, post-testing reporting delivery and data protection. Only CREST penetration testing service providers can promise to conduct pen testing services to this gold standard. Black Box Testing This method examines functionality with no prior knowledge of the system, application or infrastructure being tested. White Box Testing This cybersecurity testing method looks at the internal source coding structure aided by full information disclosure on the target. Grey Box Testing This method is similar to white box testing but with only limited knowledge of the system, application or environment being targeted. The benefits of penetration testing regularly, such as quarterly, or at a minimum annually, will ensure the business is continuously safeguarded. You will be able to: - Identify any security issues or vulnerabilities and remediate them with the right controls - Benchmark your existing processes and security controls - Understand where software or applications have developed bugs or not been patched sufficiently - Ensure business continuity by preventing disruptions caused by attacks - Support any regulatory compliance requirements such as GDPR (General Data Protection Regulations) or PCI-DSS (Payment Card Industry Data Security Standard) - Provide assurance to senior management, stakeholders, partners and most importantly maintain trust with customers that their data is protected There are clear benefits of penetration testing for organisations. Pen tests will certainly uncover your security weaknesses and how vulnerable your company is to cyberattack. They can also identify potential threats to your cybersecurity. By conducting penetration testing, you can safeguard your security posture before a cybercriminal has a chance to exploit your vulnerabilities. If any weaknesses are identified during pen testing, they must be addressed as soon as possible. Any vulnerabilities that are left unpatched are likely to be exploited by bad actors and will compromise the business. This helps to reduce information security risk and reports can be shared with senior management to improve cybersecurity awareness. What are the advantages and benefits of penetration testing? The tactics, techniques and procedures (TTPs) that cybercriminals use to attack networks, software systems and applications are growing in volume and sophistication. The benefits of penetration testing mean that not only can businesses safeguard their cyber security before a cybercriminal has a chance to exploit vulnerabilities – they can also improve internal security management processes. Another benefit of penetration testing is that organisations can test the effectiveness of their Intrusion Detection systems and teams to see if the attempted attack is identified. This will remove time constraints around annual testing, allow for a deeper and wider variety of penetration testing to be done, and protect the organisation more effectively against cyber attacks. One of the key benefits of penetration testing is that it can provide organisations with a clear picture of their attack surface and risk profile. If any gaps are left unpatched or unaddressed, bad actors are likely to exploit and compromise the business. Penetration testing ensures security controls and processes are in place, so gaps are remediated in a timely manner. A benefit of penetration testing is the understanding of where budget or investment is needed in order to remediate issues. It can also shed light on cyber awareness training needs within the organisation. Benefits of penetration testing more regularly also include regular checks on systems, networks and applications so the time to remediate doesn’t leave the business vulnerable to attack. Any weaknesses left unpatched are a huge risk to business operations and will be exploited by threat actors. The benefit of penetration testing to monitor weaknesses ensures that remediation actions are completed promptly and helps reduce information security risk. The average cost of a data breach in the UK is now $4.35 million, up by 12% on the previous year. Ignoring vulnerabilities can lead to millions in damages to business operations, company reputation and fines. The benefits of penetration testing help to avoid these costs by preventing cyberattacks before they occur. As mentioned, organisations can also make a more informed investment in cyber security where it’s most needed to utilise the budget more efficiently. Firstly, it’s not always possible for organisations to hire security professionals in-house, especially when considering the cyber security skills gap. There are specialised skills and qualifications needed to conduct penetration testing. The cyber security industry is short of 2.7 million workers. By outsourcing to a managed cyber penetration testing provider, businesses free internal staff to work on in-house projects. The key benefit of penetration testing conducted with a cyber security specialist partner, is the access to industry-wide insight and extensive knowledge of the entire threat landscape. One of the key benefits of penetration testing for business continuity is the timely mitigation of any issues that may be identified in the pentest. Networks, systems and applications with vulnerabilities are at a much higher risk of exploitation. Threat actors use the same security tools that are utilised by pen testers to find those companies that have vulnerabilities. The benefits of penetration testing in terms of compliance mean that the organisation can support information security and compliance requirements such as GDPR (General Data Protection Regulations), PCI DSS and ISO 27001 by supplying more up to date information and reporting. Another benefit of penetration testing regularly is that the business can demonstrate audit trails and evidence their commitment to regulatory compliance. of penetration tests identify critical vulnerabilities of pen tests found sensitive data begin transferred is all it takes for a hacker to breach a network and gain access to data and systems What is continuous penetration testing? Conducting penetration testing just once every 12 months is no longer sufficient to protect businesses. The speed at which digital transformation is moving and tools and technology are updated poses a constant security threat. While annual pen tests, or half year or quarterly tests, provide a moment-in-time snapshot of a company’s potential vulnerabilities, in isolation they can’t paint an accurate picture of long-term security risks. It’s imperative that continuous penetration testing is conducted to protect networks, applications and systems. The cycle of continuous penetration testing should start from the baseline penetration test. Alongside this baseline penetration testing, continuous penetration testing identifies new weaknesses that could be exploited. The process should then include steps which define the scope and assets to be continuously tested, a schedule of regular security testing, remediation of any issues identified, retesting and ongoing tracking of upgrades, misconfigurations and newly reported threats and vulnerabilities. Continuous penetration testing will enable organisations to protect their security posture on an ongoing basis – before cybercriminals attempt to exploit their vulnerabilities. DigitalXRAID’s Penetration Testing Services DigitalXRAID’s penetration testing services will identify any weaknesses and vulnerabilities in your systems, networks and applications. We give you the chance to remedy issues before threat actors can exploit them, protecting you from attacks. DigitalXRAID is one of the first managed cyber security service providers to gain CREST certification for our penetration testing services. This makes us one of the top penetration testing providers in the world. If there’s a vulnerability, DigitalXRAID’s penetration testing experts will find it. For more information on our penetration testing services and how we can support you in staying a step ahead of cyber criminals, speak to an expert. For an in-depth view of the benefits of penetration testing by DigitalXRAID experts and to get tailored quote: scope your project. With cyber security services operating on both the offensive and defensive sides, DigitalXRAID have a much deeper understanding of what techniques are being used for both attack and defense. Therefore, our CREST certified pen testing team dive deeper, uncovering vulnerabilities that others tend to miss. DigitalXRAID’s security testers can offer penetration testing services, including: - Internal Penetration Testing Services - External Penetration Testing Services - PCI DSS Penetration Testing Services - Red Teaming - Social Engineering - Mobile app Penetration Testing Services - Web application Penetration Testing Services Protect your business with the Benefits of Penetration Testing A security partner you can trust Make sure you’re truly protected by putting your networks, systems and applications to the test. As with all cyber security, the benefits of penetration testing forms a more robust security posture. We’ll work with you to identify and remedy weaknesses in your security before a malicious party exploits them. Protect Your Business & Your Reputation. With a continued focus on security, you can rest assured that breaches and exploits won't be holding you back.
<urn:uuid:08d2db38-7459-4238-a393-b361937679bd>
CC-MAIN-2024-38
https://www.digitalxraid.com/benefits-of-penetration-testing/
2024-09-16T05:28:55Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651676.3/warc/CC-MAIN-20240916044225-20240916074225-00261.warc.gz
en
0.929717
2,851
2.875
3
Several concepts are introduced in the API. - Engine Classes and Algorithms An engine class defines a cryptographic service in an abstract fashion (without a concrete implementation). A cryptographic service is always associated with a particular algorithm or type, and it either provides cryptographic operations (such as those for digital signatures or message digests), generates or supplies the cryptographic material (keys or parameters) required for cryptographic operations, or generates data objects (keystores or certificates) that encapsulate cryptographic keys (that can be used in a cryptographic operation) in a secure fashion. For example, two of the engine classes are the class provides access to the functionality of a digital signature algorithm. A DSAKeyFactory supplies a DSA private or public key (from its encoding or transparent specification) in a format usable by theinitSign methods, respectively, of a DSASignature object.The Java™ Cryptography Architecture encompasses the classes of the Java 2 SDK Security package related to cryptography, including the engine classes. Users of the API request and use instances of the engine classes to carry out corresponding operations. The following engine classes are defined in the Java 2 SDK:In the 1.4 release of the Java 2 SDK, the following new engines were added and the services are provided by theMessageDigest : used to calculate the message digest (hash) of specified data.Signature : used to sign data and verify digital signatures.KeyPairGenerator : used to generate a pair of public and private keys suitable for a specified algorithm.KeyFactory : used to convert opaque cryptographic keys of typeKey into key specifications (transparent representations of the underlying key material), and vice versa.CertificateFactory : used to create public key certificates and Certificate Revocation Lists (CRLs).KeyStore : used to create and manage a keystore. A keystore is a database of keys. Private keys in a keystore have a certificate chain associated with them, that authenticates the corresponding public key. A keystore also contains certificates from trusted entities.AlgorithmParameters : used to manage the parameters for a particular algorithm, including parameter encoding and decoding.AlgorithmParameterGenerator : used to generate a set of parameters suitable for a specified algorithm.SecureRandom : used to generate random or pseudo-random numbers. : used to build certificate chains (also known as certification paths).CertPathValidator : used to validate certificate chains.CertStore : used to retrieveCertificate objects from a repository. Note: A generator creates objects with brand-new contents, whereas a factory creates objects from existing material (for example, an encoding).An engine class provides the interface to the functionality of a specific type of cryptographic service (independent of a particular cryptographic algorithm). It defines Application Programming Interface (API) methods that allow applications to access the specific type of cryptographic service it provides. The actual implementations (from one or more providers) are those for specific algorithms. The engine class, for example, provides access to the functionality of a digital signature algorithm. The actual implementation supplied in aSignatureSpi subclass would be that for a specific kind of signature algorithm, such as SHA-1 with DSA, SHA-1 with RSA, or MD5 with RSA.The application interfaces supplied by an engine class are implemented in terms of a Service Provider Interface (SPI). That is, for each engine class, there is a corresponding abstract SPI class, which defines the SPI methods that cryptographic service providers must implement. An instance of an engine class, the API object, encapsulates (as a private field) an instance of the corresponding SPI class, the SPI object. All API methods of an API object are declared final and their implementations invoke the corresponding SPI methods of the encapsulated SPI object. An instance of an engine class (and of its corresponding SPI class) is created by a call to the factory method of the engine class.The name of each SPI class is the same as that of the corresponding engine class, followed by . For example, the SPI class corresponding to theSignature engine class is theSignatureSpi class.Each SPI class is abstract. To supply the implementation of a particular type of service, for a specific algorithm, a provider must subclass the corresponding SPI class and provide implementations for all the abstract methods. Another example of an engine class is the class, which provides access to a message digest algorithm. Its implementations, inMessageDigestSpi subclasses, can be those of various message digest algorithms such as SHA-1, MD5, or MD2.As a final example, the engine class supports the conversion from opaque keys to transparent key specifications, and vice versa. (See the Key Specification Interfaces and Classes section.) TheKeyFactorySpi subclass supplies an actual implementation for a specific type of key, for example, DSA public and private keys. - Implementations and Providers Implementations for various cryptographic services are provided by JCA Cryptographic Service Providers. Cryptographic service providers are essentially packages that supply one or more cryptographic service implementations. The Engine Classes and Algorithms section includes a list of some of the implementations supplied by IBMJCE. Other providers might define their own implementations of these services or of other services. - Using Factory Methods to Obtain Implementation Instances For each engine class in the API, a particular implementation is requested and instantiated by calling a factory method on the engine class. A factory method is a static method that returns an instance of a class. The basic mechanism for obtaining an appropriate object, for example, is as follows: A user requests such an object by calling thegetInstance method in theSignature class, specifying the name of a signature algorithm (such as SHA1withDSA), and, optionally, the name of the provider or theProvider method finds an implementation that satisfies the supplied algorithm and provider parameters. If no provider is specified,getInstance searches the registered providers, in preference order, for one with an implementation of the specified algorithm. See TheProvider Class for more information about registering providers.
<urn:uuid:191ede11-00e3-467d-b032-75c25154b2d4>
CC-MAIN-2024-38
https://www.ibm.com/docs/en/cloud-paks/z-modernization-stack/2023.4?topic=introduction-concepts
2024-09-16T06:08:05Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651676.3/warc/CC-MAIN-20240916044225-20240916074225-00261.warc.gz
en
0.819768
1,265
3.0625
3
Politicians in the United States have met significant resistance in their efforts to craft effective cybersecurity legislation. While most officials realize the need for regulatory guidelines to address digital threats, disagreements over the role of government often present a challenge. One of the key issue centers around cybersecurity information sharing. On one hand, many experts believe it is necessary for the U.S. government to design laws that would force companies to share information about cyber criminal activity. In return, the government would share some of its information. However, critics of this strategy believe information sharing should be voluntary and encouraged through incentives. As legislative failures in the U.S. have shown, it isn’t easy to reconcile these two views in one country. And the problem gets even more complex on the international level, according to recent AOL article. The article highlighted comments from Major General David Neasmith, an Afghanistan veteran who now heads information management on the Canadian joint defense staff. Neasmith said that truly effective information sharing would have to be automated, so that cyber criminal activity is passed from computer to computer. However, setting up such a protocol to protect privacy between government entities and private businesses is not going to be easy. NATO’s manual for cyberwarfare Despite the struggle surrounding cybersecurity information sharing, NATO has made some progress in outlining guidelines for cyberwarfare. The draft of the Tallinn Manual is designed to provide guidelines for what cyber operations may constitute “use of force” between two nation states. Under most circumstances, use of force from one state against another is prohibited by international law. The manual does not dictate guidelines for responding to cyber criminal activity and is only designed to clarify acts of war. For example, a cyberattack that does comparable damage to a physical use of force is also considered unlawful. However, attacks that only result in psychological effects, such as reduced morale, are not considered uses of force. In addition, the manual dictates that a victim of such an attack has the ability to defend itself, using force if necessary. What role should government entities play in improving national cybersecurity? Should an international law be crafted that better defines cyber criminal activity?
<urn:uuid:fca3951b-e910-4ce2-ba2e-69a5ecd38f4a>
CC-MAIN-2024-38
https://www.faronics.com/news/blog/the-struggle-for-international-cybersecurity
2024-09-18T18:48:09Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651931.60/warc/CC-MAIN-20240918165253-20240918195253-00061.warc.gz
en
0.95357
437
2.984375
3
The benefits of artificial intelligence (AI) and robotics in medicine can be divided into three groups: increased efficiency and productivity by optimizing clinical processes and allowing doctors to concentrate their full attention on the patients; improved patient experience by avoiding unnecessary procedures; continuous development of care, based on accurate diagnosis and optimized treatments. The availability of new technologies is increasing in hospitals around the world and companies are developing new solutions that promise fewer errors and even fewer risks. At the same time, international organizations are keen on encouraging healthcare innovation. Thus, the future of these new technologies for the healthcare industry might not mean that doctors will be soon replaced by robots, but rather that AI and robotics will help them with their work. But the use of AI and robotics in medicine also has its limitations, just like any other technology. For example, algorithms learn to do certain tasks, such as recognizing diseases, by “feeding” on as much patient data as possible. However, in some cases, such as rare diseases, there is simply not enough data available. Machine learning cannot be used in order to recognize these conditions. So while the new technologies promise to bring new effective answers to old problems in the field, they also come with problems of their own. The other side of the story Doctors seem to agree in saying that new technologies are essential for the future of medicine, but they may also want to take note of their side effects. For instance, while robotic surgeries do cause significantly less pain than conventional interventions, it is not yet clear how these assisted interventions will impact on surgeons’ skills. Technology is getting better and better at a very fast pace and there seems to be a growing need for new studies when discussing its effects. “More frequent use of digital media may be associated with the development of ADHD symptoms,” according to a report published in the Journal of the American Medical Association. The new study involved roughly 2,500 teens over a two-year period. While the subjects had no significant symptoms of ADHD at the beginning of the study, teenagers who used social media heavily were found to be twice as likely as infrequent users to show symptoms of ADHD. While the authors mention that “further research is needed to assess whether this association is causal,” we cannot help but ponder on the implications. Could the extensive use of some specific technologies make us prone to psychiatric problems? Will doctors’ ability to concentrate grow in the future or could AI and robotics have an opposite effect? What have we learned so far? AI and robotics are now used in medical centers around the world that believe in true innovation and have the resources to sustain it. The medical community is also showing a growing interest in the new solutions. At the same time, people everywhere are becoming more interested in preserving and managing their own health. “Prevention” is a buzzword in healthcare, just like AI and robotics. Will that be enough to build a better future? Considering the benefits, one may argue that the future of healthcare looks brighter than ever. The introduction of digitalized health record has proved to be a game changer in terms of efficiency and productivity, and research shows that it also reduced the costs of outpatient care. Surgeons put their trust in robotic surgery, as it causes less pain than conventional interventions and provides patients with better chances for a speedy recovery. Cloud computing has provided the healthcare industry with better and safer ways to store and access big data. Breakthroughs in communication and especially telemedicine are now key factors that connect healthcare professionals with people, especially in rural areas. Despite the apparent risks, the impact of AI and robotics on healthcare seems to be positive, with the advantages compensating for the short list of side effects. The complete impact of AI and robotics on medicine may be hard to asses now. It may be reflected by doctors’ tendency to use robots for certain tasks, making precision medicine grow and evolve. Technology may also prove useful in building better health systems with large hospitals that provide patients with the same standards of care. Even the common meaning of the word “doctor” may change, as more and more healthcare professionals will probably have to acquire new digital skills. As new technologies can diagnose and treat old issues for the healthcare industry, the list of disadvantages seems rather short for the moment.
<urn:uuid:e89bf9fc-8f9a-40e7-85b3-9bbb69f2b35b>
CC-MAIN-2024-38
https://healthcarecurated.com/editorial/can-new-technologies-diagnose-old-issues-for-the-healthcare-industry/
2024-09-07T19:51:35Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650920.0/warc/CC-MAIN-20240907193650-20240907223650-00161.warc.gz
en
0.964515
872
3.53125
4
The digital revolution in healthcare has brought unparalleled advancements such as Electronic Health Records (EHRs), telemedicine, and sophisticated data-driven care. However, these innovations come with the heavy responsibility of safeguarding patient privacy. The significance of protecting personal health information cannot be overstated, as breaches can lead to devastating consequences, including identity theft and loss of public trust. PHI breaches not only jeopardize patient confidentiality but may also result in legal penalties and financial losses for healthcare providers. To combat these risks, robust data security measures are imperative. This entails encrypting patient data, training healthcare staff in data privacy, and implementing strict access controls. Additionally, regular audits and compliance with regulations like the Health Insurance Portability and Accountability Act (HIPAA) in the U.S. are crucial in maintaining data integrity. Promoting awareness among patients about their data rights and how they can protect their PHI is also a key strategy. In the interconnected landscape of digital healthcare, the collective effort of healthcare entities, IT professionals, and patients is necessary to secure sensitive health information. As technology continues to evolve, vigilance and constant adaptation of security measures must keep pace to safeguard the privacy that is fundamental to patient care. Importance of Healthcare Data Privacy Maintaining the confidentiality and security of sensitive health data is not merely a legal requirement but a critical aspect of patient trust and safety. Let’s delve into the reasons why healthcare data privacy is a fundamental concern for both patients and healthcare organizations. Building Patient Trust through Data Security Patient trust is the cornerstone of the healthcare provider-patient relationship. When patients confide their personal and sensitive health information to healthcare providers, they inherently entrust them with their well-being. The assurance that this information remains confidential is vital. A breach can lead to a profound loss of trust, causing patients to withhold vital information or even avoid seeking care altogether, potentially leading to poorer health outcomes. It is imperative for healthcare organizations to implement robust security measures to maintain this trust. Healthcare providers must also recognize the broader implications of data breaches. The disclosure of PHI can harm patients beyond the healthcare setting, affecting their employment, insurability, and privacy. Consequently, healthcare providers have a moral obligation to protect this data diligently. Legal Obligations and the Impact of Non-Compliance Healthcare entities are subject to a complex web of regulations that aim to protect the personal health information (PHI) of individuals. In the United States, healthcare providers adhere to the Health Insurance Portability and Accountability Act (HIPAA), which sets the bar for PHI security, and the Health Information Technology for Economic and Clinical Health (HITECH) Act, which imposes more stringent regulations on electronic records. Across the pond, the General Data Protection Regulation (GDPR) governs the handling of personal data for individuals within the EU, impacting any organization that deals with this information. Falling afoul of these regulations isn’t just a slap on the wrist; it carries severe repercussions. Heavy financial penalties, legal repercussions, and a tarnished reputation are among the dire consequences healthcare organizations can face for non-compliance. Moreover, if a data breach does occur, the costs of managing the fallout often reach astronomical figures. Through providing clear guidelines, these legal frameworks serve a dual purpose: they enforce the protection of patient information while also equipping healthcare providers with the means to minimize the risk of potential data breaches. The Risks and Relationships Inherent in Health Data With the growing interconnectivity of healthcare systems, the data captured and stored by healthcare providers can highlight sensitive relationships that pose privacy risks. This section offers insight into the threats associated with improper data handling. The Treasure Trove of Relational Health Data Health data’s complexity transcends mere medical specifics; it unveils deep personal narratives. Such data holds clues to a person’s financial status, close connections, and whereabouts at specific times. In adverse scenarios, this could lead to severe violations like identity theft or blackmail. The data’s sensitive nature goes beyond the individual, entwining with family health histories and possibly affecting kin. It’s this web of direct and indirect information that makes safeguarding PHI imperative. Privacy measures must be robust enough to shield against any leaks that may expose personal details, both outright and through association. As such, protection protocols for health data must be comprehensive, shielding the apparent and the subtle threads that make up the fabric of an individual’s health narrative. Cyber Threats to Healthcare Data The rate at which cyber threats evolve consistently outpaces the defenses many healthcare organizations have in place. As a result, healthcare data has become a prime target for cybercriminals due to its high value on the black market. Ransomware attacks, in particular, have escalated, with attackers locking healthcare providers out of vital systems and demanding significant ransoms. Phishing schemes can also trick unsuspecting employees into divulging login credentials or downloading malware. The interconnectedness of modern healthcare systems means that a breach in one area can quickly spread throughout an entire network, substantially increasing the potential damages. Healthcare organizations must, therefore, adopt a multi-layered approach to cybersecurity, with regular updates to protect against the latest threats. Strategies and Solutions for Protecting PHI As the healthcare industry continues to embrace digital technologies, it faces the daunting task of implementing and maintaining robust security measures. This section will cover current and emerging solutions for protecting health information. Employing Advanced Privacy Technologies The concept of differential privacy is becoming increasingly important in the realm of healthcare data. By adding a level of uncertainty to data, differential privacy algorithms allow the data to be used for analysis without exposing identifying details. While this technique introduces a trade-off between privacy protection and data accuracy, researchers and developers are continuously working to refine these algorithms to balance the two concerns effectively. Moreover, healthcare can take advantage of other advanced technologies such as encryption, pseudonymization, and tokenization. These methods transform sensitive data into unreadable formats for unauthorized users while allowing legitimate access for healthcare operations and research. The right combination of these advanced technologies can provide rigorous data protection for PHI without hampering its availability for the valuable insights it can provide. Training and Compliance: The Human Factor Humans are often the weakest link in the data privacy chain. It is therefore essential that all healthcare employees receive comprehensive training on the importance of PHI protection and the specific processes in place at their institutions. This training should cover the various ways data breaches can occur, the importance of strong passwords, recognizing phishing attempts, and proper responses to suspected breaches. Moreover, ongoing compliance is necessary to adapt to the constantly evolving landscape of data protection regulations. It is crucial for healthcare organizations to regularly evaluate their practices against the latest legal requirements and implement the necessary updates. Failure to maintain compliance can result in severe penalties, but more importantly, it can lead to vulnerabilities that might put patient data at risk. Continuous Vigilance: The Path Forward in Healthcare Privacy The future of healthcare privacy is an ongoing journey marked by continuous advances in cybersecurity measures and the evolving landscape of data protection regulations. This section discusses how the healthcare industry must adapt to ensure the ongoing integrity and confidentiality of patient information. Innovations in Privacy Technology Innovative solutions like Tonic are setting the stage for a revolution in the way healthcare data is managed and safeguarded. By generating synthetic data that closely mirrors real-world information without revealing any personal health information (PHI), these tools are carving out new paths for development and research. This artificial data is an invaluable asset for software testing and analytical purposes, ensuring that actual patient details remain uncompromised. The surge in investment towards privacy technologies indicates a promising future for healthcare organizations, which will likely have access to a suite of advanced tools aimed at protecting patient data against new and evolving threats. Furthermore, these tools pave the way for the creation of systems that incorporate ‘privacy by design’. This concept integrates security into the technology landscape from the very beginning, embedding it into the foundation of healthcare IT infrastructure. With security as a cornerstone, these innovations promise a safer, more secure environment for handling sensitive healthcare information. As a result, they offer peace of mind for patients and healthcare providers alike, and stand to significantly reduce the risk of data breaches that have long plagued the industry. This approach not only ensures compliance with rigorous privacy standards but also enhances the overall trust in healthcare systems. Proactive Breach Risk Management In safeguarding Patient Health Information (PHI), proactive defensive measures are paramount. Healthcare entities regularly carry out risk assessments to pinpoint and shore up weak spots—these include software vulnerabilities, scrutinizing employee access, and fortifying the locations where data is physically stored. A comprehensive preemptive scan not only identifies risks but leads to critical enhancements in safeguarding PHI before any actual damage arises. It’s a process of continuous improvement and vigilance against potential security threats. In the face of an actual data breach, healthcare providers must have effective response strategies in place. These are meticulously crafted plans that outline immediate actions—stopping the breach from spreading, evaluating the damage done, and promptly notifying the necessary parties as dictated by legal guidelines. By being well-prepared, healthcare organizations are better positioned to curtail the negative repercussions of data breaches. This level of readiness is essential not just for the protection of patient information but also to maintain the trust and integrity of the healthcare institution. Adapting to Regulatory Shifts In the ever-shifting landscape of technology and security, healthcare organizations must continually adapt to changing data protection regulations. It’s not enough to comply once; healthcare entities must engage in a dynamic process of continual compliance to ensure they are up-to-date with the latest standards. By keeping abreast of regulatory changes, healthcare providers can avoid hefty penalties and legal entanglements. More importantly, compliance signifies a commitment to preserving patient confidentiality, reflecting the core values of patient care. When an organization prioritizes data security, it fosters a relationship of trust with its patients, assuring them that their sensitive health information is in safe hands. Healthcare professionals must, therefore, view compliance as an ongoing responsibility—a critical aspect of their operations that demands regular attention and proactive management. In doing so, they not only fulfill their legal obligations but also demonstrate to their patients that their privacy and well-being are of utmost priority. This diligent approach to data protection is crucial in maintaining the integrity of patient care and upholding the reputation of healthcare institutions in the modern world.
<urn:uuid:a22c64d2-01d2-4893-9548-20cb77bf4eba>
CC-MAIN-2024-38
https://healthcarecurated.com/management-and-administration/safeguarding-patient-data-navigating-healthcare-privacy-challenges/
2024-09-07T20:58:52Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650920.0/warc/CC-MAIN-20240907193650-20240907223650-00161.warc.gz
en
0.922308
2,148
2.921875
3
What is ELT? ELT (Extract, Load, Transform) is a data integration process that has gained great importance in modern data handling and analytics. This is an evolution from the traditional ETL (Extract, Transform, Load), which follows a reverse order of operations. In ELT, the first step involves extracting data from different sources before loading it onto one central storage system and finally transforming it into an analysis-friendly format. It has several advantages compared to traditional ETL, especially when dealing with big data sets effectively and adapting to business changes. Key Components of ELT - Extract: The extraction phase entails obtaining information from various sources such as databases, web services, and spreadsheets. It is critical to ensure that all relevant data is captured during this stage to not miss any important aspects necessary for analysis. - Load: The load phase involves transferring the extracted data into a central storage system like a data warehouse. The choice of storage solution is critical in ELT because it directly affects its performance and scalability. - Transform: During the transformation phase raw information must be put in proper format that can enable it to be analyzed properly. The latter depends on such different actions as aggregating or filtering, cleansing etc., whose aim is just to make sure that the given statistics are accurate and ready for usage. Process Flow in ELT ELT process involves the following steps: - Data Extraction: This stage involves using tools such as SQL-based systems and cloud warehouses like Snowflake or BigQuery, as well as other Data Integration Tools to facilitate retrieval of data from various sources. - Data Loading: The tools used for this stage include SQL-based systems, cloud warehouses, or even Data Integration Platforms, among others, to enable the transfer of extracted records to any centralized repository, for instance, a Data Warehouse. - data-contrast=”none”>Data Transformation: This stage involves aggregating, filtering out redundant records, and data cleansing the loaded data to convert it into an easily analyzed format. - Scheduling and Automation: It constantly updates and analyzes the information flow. Advantages of ELT - Performance: There are several reasons why modern business intelligence applications use Extract-Load-Transform rather than traditional Extract-Transform-Load techniques, even though both involve facilitating performance. It can achieve query acceleration mechanisms by taking advantage of parallel processing architectures. - Flexibility: ELT is very flexible with changes in industry requirements; it allows for modifying transformations without extracting them again. - Cost Efficiency: Using cloud infrastructure-based ELTs reduces firms’ costs, as they do not need to maintain expensive hardware and infrastructure. - Scalability: ELT always ensures an efficient and effective DI&A process regardless of the incremental rise in the volume of data, hence leading to scaling up. Use Cases & Applications ELT offers prompt data uploading and effective data conversion, making it ideal for data warehousing. - Business Intelligence: ELT makes it easier to derive insights from the information by providing a flexible and scalable process of integrating data. - Data Integration: ELT unifies the data of different departments, resulting in a common understanding of operations and enabling organizations to make decisions based on them. - Machine Learning and AI: To support model training and predictions, ELT provides a scalable and efficient integration process for machine learning and AI applications. Challenges of ELT - Data Quality: Data quality assurance remains one of the significant bottlenecks in ELT, involving ensuring both precision and usefulness. - Security: During this loading process, where confidential information must be handled, sensitive information must be ensures safety. - Skill Requirements: Setting up an ELT pipeline requires specialization, particularly in skills required for processes like integration or transformation. ETL tools and software ETL can extract, load, or transform your data using different tools. However, some tools cover all these steps while integrating those into just one place called ETL/ELT. When choosing an ETL tool, users should identify whether they can read multiple sources, including what they are using now and any future sources their company may adopt. Most tools support source/target database systems and files compatible with many types. Some may look for tools that can do both ETL and ELT, as it is possible to require both in the future. Today, there are several market leaders such as IBM, Informatica, Microsoft, Oracle, SAS, Talend, Teradata, etc., Data Store helps manage the target data market, data warehouse, and/or data lake. NoSQL DMS and Hadoop are good candidates for an ELT approach, as are purpose-built DW appliances. In some situations, a classic relational DBMS will be relevant in practice. ELT has redefined modern data management by offering a customizable, scalable, and efficient method for integrating data. The size of data is increasing dramatically, making ELT critical in handling big datasets and supporting decision-making based on them. Emerging trends such as real-time ELT and cloud-native ELT are expected to further enhance the capabilities of ELF, thereby making it an essential tool for businesses seeking competitive advantages from their respective databases. Share this glossary
<urn:uuid:94cc617d-9fe3-4bd1-8a7d-8c8fec5facc6>
CC-MAIN-2024-38
https://kanerika.com/glossary/elt/
2024-09-07T21:03:27Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650920.0/warc/CC-MAIN-20240907193650-20240907223650-00161.warc.gz
en
0.918921
1,082
2.953125
3
Generative Pretrained Transformer 3 (GPT-3) is a highly intuitive cutting-edge language processing AI model developed by Open AI – an American research laboratory. It can generate human-like text, opens itself up to a wide range of applications. The AI stands out in its ability as a highly capable chatbot. GPT-3 allows the user to interact with the trained AI with a wide range of worded prompts in the form of questions, requests, or several other worded requests. GPT-3 has about 175 billion built-in parameters, making it almost impossible to narrow down its scope. However, this model is restricted to language alone and hence cannot produce video, sound, or images but has an in-depth understanding of both spoken and written words. According to McKinsey, generative AI is helping to bring down the development time and offering powerful capabilities to users that can transform the business world. Businesses must revisit their processes and services offered with the advent of new technology; AIs like ChatGPT can enhance workflow automation and deliver superior customer experiences. Chat-based AI can augment human efforts to provide more engaging interactions with users. - Compiling detailed research on various domains. - Can help create compelling marketing content. - Can be used as a collaborative tool among employees to brainstorm ideas. - Automate the query-related aspects of the sales process. - Easy onboarding of customers and work on providing customized instructions for after-care services. - ChatGPT cannot write complex code, such as what’s required for banking applications, but it has every potential to become a proficient coder within the next decade, - Help handle customer engagement, leading to improved loyalty and retention of customers. - It may mark the beginning of AI-creating automation that can rise above what humans can create alone. Let us look at these applications in a detailed manner. - Handling unstructured data: Unstructured data is difficult to sort, manage and organize to derive use from. ChatGPT can easily convert unstructured data into structured data. - Generating SQL queries with ease: ChatGPT can generate SQL queries, an essential tool for any data scientist. ChatGPT helps gain a solid understanding of SQL for those unversed with it. - Provide superior customer service: ChatGPT helps businesses generate automated responses to all the customer wants by training it on a large dataset of customer interactions. This brings down the workload of service teams and improves customer satisfaction. - Language translation for the business: ChatGPT can automatically translate text from one language to another as it is exposed to a large dataset of text in multiple languages. This is particularly useful for businesses with international operations. - A to Z of text generation: ChatGPT can be used to create anything from social media posts to product descriptions. This is helpful to enterprises looking at expanding their customer base to conduct targeted campaigns effectively. - Content personalization and virtual writing assistant: ChatGPT can also be used to tailor content to specific users in the form of emails, social media posts, and product recommendations. It can also generate complete articles or blog posts helping bloggers and writers who produce a large amount of content on a regular basis. It can also generate summaries of lengthy articles or documents. - Customer sentiment analysis: ChatGPT can also be used to analyse sentiment by training it on large datasets of texts. This helps businesses monitor customer sentiment on social media or to analyze customer feedback. - Classification of Text into various groups: Businesses can use ChatGPT to automatically classify new text into the appropriate category. This can be done by training it on a dataset of text marked under different categories. While AI seems to be taking over the industrial world, it is critical for users to clearly understand what an AI tool can and cannot do and whether it is the right fit for a business. To know more about the potential and challenges associated with AI tools, contact us. ChatGPT and Digital Transformation ChatGPT can disrupt the field of software engineering. It has significant potential to accelerate digital transformation programs for enterprises. Developers can fast track software development, debug code quicker, and write in languages that were not known earlier. ChatGPT can resolve simple comprehension queries of developers, thereby, helping developers equip themselves with more skills. IT teams may begin to consider ChatGPT as a productivity tool, allowing them to spend more time combatting architectural issues and engaging with the CX team for better outcomes. Businesses can reimagine a digital future beyond the confines of conventional rules. It highlights that this future can be realized if there is enough preparation to tackle challenges that may crop up due to the usage of AI. It is essential to note that there are limitations to what these tools can do and the associated ethical concerns of working with AI. The ability of ChatGPT to generate highly coherent and contextually appropriate text can help businesses improve customer service, reduce costs, and increase overall efficiency. Implementing ChatGPT in a business setting can be complex and may require a substantial investment in terms of time and resources. This is a major challenge. Overall, ChatGPT has the potential to bring significant benefits to businesses but jumping into implementing one requires a lot more careful consideration and planning.
<urn:uuid:f9ba39cb-d32d-49d5-b75d-3432f3007bfe>
CC-MAIN-2024-38
https://www.gemini-us.com/ai/chatgpt-transform-your-enterprise-with-new-ai-model
2024-09-10T03:41:09Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651196.36/warc/CC-MAIN-20240910025651-20240910055651-00861.warc.gz
en
0.931073
1,088
2.515625
3
The Internet of Things (IoT) has shown significant growth and promise, with data generated by IoT devices alone expected to reach 73.1 zettabytes by 2025. Moving this data away from its point of creation to a centralized data center or cloud would contradict the application’s purpose. Thus, edge computing was born. Fast forwarding to 2024, edge computing is now being paired with recent advances in AI to intelligently process data at the edge, leading to faster speeds, reduced latency and improved privacy and security. In sectors like manufacturing and healthcare, where efficiency and accuracy are key, AI at the edge is changing the game. In manufacturing, where there are approximately 15 billion connected devices globally, the milliseconds lost in sending data to the cloud for processing can mean the difference between detecting a flaw immediately or letting it slip through quality control. In healthcare, the immediacy with which patient data is analyzed can affect the accuracy of diagnoses and the effectiveness of treatment, especially with the risk of decentralized healthcare and wearable devices. By processing data on the spot, AI at the edge eliminates the latency that cloud computing introduces, leading to more timely, informed decisions. The global market for edge computing technologies is estimated to increase from $46.3 billion in 2022 to $124.7 billion by 2027 at a compound annual growth rate (CAGR) of 21.9 percent from 2022 through 2027. Implementing AI at the edge will result in tangible benefits to all industries, enabling businesses to unlock new possibilities and achieve greater levels of performance. The Shift To Smaller Models In the past year, the conversation around AI models has begun to change. Large models with extensive parameter counts have started giving way to smaller, more focused models. This includes both the utilization of smaller models as well as the utilization of efficiency techniques, such as quantization, sparsity and pruning, to make large models smaller. These smaller models are easier to deploy and manage while being significantly more cost-effective and explainable, yielding similar performance with a fraction of the computational resources. These smaller models can also be used in many task-specific domains. Pre-trained models can be optimized for specific task performance using techniques such as inferencing and finetuning, making them perfect candidates for the stringent requirements of edge computing. These smaller models are not only beneficial to the logistical challenges of deploying hardware at the edge, but they also meet the nuanced needs of specific applications. In manufacturing, a small, specialized AI model can continuously monitor the auditory signatures of machines to predict maintenance needs before a breakdown occurs. In healthcare, a similar model can provide continuous, real-time monitoring of patient vitals, alerting medical staff to changes that may indicate an emerging condition. Mastering Model Optimization And Inferencing Techniques Optimization at the edge is not just about making AI models smaller; it’s a balancing act to make models as small as possible while still retaining performance. Techniques such as pruning convert larger models into smaller models by reducing the number of unimportant connections and, more recently, layers. Pruning aims to create more memory and energy-efficient systems that retain the performance of their original larger counterparts. Successful pruning techniques include pruning by filter, pruning by channel, and pruning by layer (where optimal blocks of layers to prune are considered via a similarity search and then model recovery is achieved through fine-tuning via parameter-efficient fine-tuning (PEFT) and quantized low-rank adapters (QLORA). Another technique used to make models smaller is quantization, a process for reducing model size by reducing the precision of model weights, parameters and activations so they have a smaller memory footprint. The memory requirement to store 32-bit or 16-bit floating point values is very high, but with quantization, these weights, parameters and activations can be converted to 8-bit, 4-bit and occasionally smaller integers that can run at the edge. For example, depending on the technique, a Llama 2 7B model can be reduced from 13.5 GB to 3.9 GB, and a Llama 13B can be reduced from 26.1 GB to 7.3 GB by FP16 to INT4 conversion. Quantization can be accomplished both through post-training and during training. To maintain performance, however, mixing precision techniques or mixing precision with pruning may need to be considered. Other efficiency techniques, such as low-rank adaption (LORA), allow for parameter-efficient finetuning by reducing computational costs and memory and increasing speed while maintaining accuracy. This technique focuses on modifying a subset of parameters rather than the entire model. This is done by keeping the original model weights frozen and applying changes to a separate set of weights, which can be added. As large models inherently possess a low dimensional structure, model parameters can be transformed into a low rank, which is a dimensionality reduction process that finds a matrix’s rank or the number of linearly independent rows or columns (this implies that no row or column can be the result of a combination of other rows or columns). Popularly, LORA is now commonly combined with quantization, and QLORA can be used for finetuning models. Additional techniques to consider include federated learning, matrix decomposition, weight sharing, memory optimization and knowledge distillation. In practice, such optimized models can provide critical insights with minimal delay. For instance, in manufacturing, AI optimized for the edge using these techniques could analyze equipment vibrations to detect early signs of wear and tear, scheduling maintenance before a failure occurs. In healthcare, edge-optimized AI can process real-time videos of patients to alert caregivers when a patient has fallen. Hardware, Software, And Learning Optimizations Dell Technologies is at the forefront of AI at the edge, optimizing both hardware and software to support AI workload deployments at the edge. With initiatives like NativeEdge, Dell Technologies ensures that AI models at the edge are not just operational but powerful, whether on a CPU or GPU. Through a nuanced approach, Dell Technologies centralizes the deployment and management of edge infrastructure and applications across geo-distributed locations, helping enterprises securely scale their edge operations using automation, open design, Zero Trust security principles and multicloud connectivity. Regardless of the intended use case, the key to successful edge and AI implementation lies in the synergy between optimized models and the hardware they run on. The Dell NativeEdge platform exemplifies the integration needed to manage these advanced AI systems effectively. Using NativeEdge, organizations can seamlessly deploy and control their edge AI applications, ensuring that the insights gleaned are timely and actionable. Putting The AI At The Edge Vision Into Practice The journey toward implementing AI at the edge is marked by real-world applications that demonstrate its transformative impact. AI at the edge has the potential to provide organizations with valuable, real-time insights to better navigate complex data and opportunities. It represents a significant shift toward a new era of business where organizations thrive on immediacy and adaptability. The time to put these theories into practice is now. Organizations that solve for integrating the value of their data through AI where it is most valuable, at the edge, will gain a competitive advantage in the months and years ahead. To learn more about AI at the edge, visit the Dell Technologies Resource Library. Ananta Nair is an artificial intelligence engineer at Dell Technologies. This article was commissioned by Dell Technologies.
<urn:uuid:29489f50-7106-4a44-8541-56d98423009b>
CC-MAIN-2024-38
https://www.nextplatform.com/2024/04/14/ai-at-the-edge-from-theory-to-practice/
2024-09-10T04:07:10Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651196.36/warc/CC-MAIN-20240910025651-20240910055651-00861.warc.gz
en
0.928337
1,525
2.609375
3
You may have heard about the recent root certificate expiration that’s been affecting a large number of sites. Root certificates are a necessary part of the certificate chain, but when they need to be replaced it affects the entire chain. Just as the roots of a tree provide life to the leaves and branches, root certificates are the base of the certificate chain. What is a root certificate? In the chain of trust, a root certificate is the first link. Unlike other certificates, it is self-signed, meaning the issuer and subject are the same. It is a kind of X.509 certificate that can be used to issue other certificates. Certificate authorities (CAs) adhere to strict requirements to merit the trust of having a root certificate. Root certificates also typically have long periods of validity, compared to intermediate certificates. They will often last for 10 or 20 years, which gives enough time to prepare for when they expire. However, there still can be hiccups in the process of switching to the new root certificate. Fixing error due to an expired root certificate When a root certificate expires, operating systems may flag the certificate as invalid even if you have the new root certificate. You may be able to fix the problem by deleting the expired root certificate.
<urn:uuid:9f79899b-1b7a-4a28-b544-46772bcda475>
CC-MAIN-2024-38
https://www.digicert.com/blog/impacts-of-root-certificate-expiration
2024-09-11T11:05:11Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651383.5/warc/CC-MAIN-20240911084051-20240911114051-00761.warc.gz
en
0.941339
257
2.640625
3
But about half said a cyberattack or lack of a hack notification would erode that trust nCipher Security, the provider of trust, integrity and control for business-critical information and applications, reveals new research indicating that people trust banks and other financial entities to safeguard their personal data more than other organizations. The findings also illustrate how easily that trust can be eroded, along with Americans’ personal data protection concerns relative to banking and digital payments. Consumers trust banks most The nCipher survey results show that people trust the financial sector in general and their banks in particular more than any other industry vertical or organizations that touch their data. A third of those surveyed said they trust financial services organizations most to protect their personal data. More than half (52%) said they trust their banks specifically to protect their data. This indicates that people place much higher trust in banks and other financial institutions than they do in other business verticals and the public sector. Less than a quarter (23%) of those surveyed trust organizations in the legal profession to handle their data. And only about 20% of nCipher survey participants said they trust their cellular provider or the government to secure their personal data. But trust can be fleeting While the nCipher survey results indicate people have relatively high trust in banks, the research also illustrates that trust can be fleeting. It’s easy for trust to be eroded or disappear completely. Financial institutions can retain that trust by building customer confidence that they have things under control. Part of that involves sharing relevant information with customers, so consumers feel more in control. More than a third (34%) of Americans surveyed said they trust companies that provide them with a feeling of control. In addition, nearly half (49%) of those surveyed said they would lose trust in their bank if it didn’t seem in control of data security. And more than half (53%) said their trust in digital payments would erode if their bank didn’t notify them of a hack within 24 hours. Digital payments are a sore point Trust related to digital payments is a particularly thorny issue. People who have been hacked in the past are especially sensitive. Nearly 60% of respondents to the nCipher survey said they would distrust digital payments if their account was compromised and showed charges they didn’t make. A third (33%) said such a scenario would prompt them to stop using digital payments entirely. Trust in online retail purchases is low Almost half (46%) of survey participants said they worry about cybersecurity the most when they buy something online. When asked what organizations they trust most to protect their personal data, the survey group ranked ecommerce from brick-and-mortar retailers lowest (6%). Even the government (20%) and cellular providers (19%) are trusted more. When asked about their trust in individual companies, popular social media companies came in last. Overall anxiety is high These specific trends are part of a larger picture illustrating the public’s high anxiety related to identity theft and personal data security. The nCipher research also revealed that: - 68% of Americans said they fear identity theft - 62% said they want companies to make security more reliable - 38% of those surveyed said they believe a hack should be a federal offense - 20% of individuals said they suffer from security “fatigue” and don’t trust anyone to protect their personal information Peter Galvin, Chief Strategy Officer at nCipher Security says “Banks and other financial organizations are in a relatively strong position when it comes to consumer trust related to personal data privacy and security. But, as the nCipher research illustrates, people are generally on edge when it comes to protecting their personal data. Organizations that want to maintain the trust they’ve worked so hard to build – and those that need to win back customer confidence – must make sure they have the protection in place to safeguard customer data. That involves creating cybersecurity strategies that employ well implemented privacy-by-design mechanisms such as encryption to keep customers’ personal data safe and retain their trust.”
<urn:uuid:dcaf4155-a72a-4926-9034-e311f3baefe8>
CC-MAIN-2024-38
https://www.entrust.com/company/newsroom/ncipher-survey-reveals-americans-trust-banks-most-their-personal
2024-09-12T15:44:08Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651460.54/warc/CC-MAIN-20240912142729-20240912172729-00661.warc.gz
en
0.962823
841
2.59375
3
Keeping Elections Safe from Cyberattack Safeguarding elections infrastructure falls to local jurisdictions—here’s what to watch out for. In a tight election, a tiny number of votes can make a huge difference, making voter suppression a very big deal. Think of Rutherford Hayes, who won the 1876 presidential election by .09 percent of total votes cast. Of course, voter suppression tools were a little less sophisticated back then—literacy tests, poll taxes, and grandfather clauses. Today, technology plays a huge role in suppressing votes. And while much has been said about the 2016 misinformation campaigns and hacking from nation state actors such as Russia, distributed denial of service (DDoS) attacks also pose a real threat. “While we see a lot of election-based security coverage on disinformation campaigns, DDoS can shut down information availability, which can be just as—if not more—dangerous,” says Tom Bienkowski, director of product marketing at NETSCOUT. Where will this play out? Two top contenders are voter registration databases and election-night results displays: - Voter registration systems. According to the Center for Internet Security’s (CSI) Handbook for Elections Infrastructure Security, the ability to access voter registration systems through the internet has increased their vulnerability to remote attacks aimed at manipulating voter registration systems. Nation-states, for example, could access and disrupt voter registration databases in order to deny legitimated registered voters the ability to vote on election day—a concept that has already translated into reality. The FBI reported in February 2020 that state-level voter registration and voter information website received anomalous Domain Name System (DNS) server requests consistent with a Pseudo Random Subdomain (PRSD) attack. According to the report, “PRSD attacks are a type of DDoS attack used by threat actors to disrupt DNS record lookups by flooding a DNS server with large amounts of DNS queries against non-existing subdomains.” Dangers such as this, the handbook notes, “makes them a priority for strengthening of the security resilience of these components.” - Election-night results displays. The CSI handbook also identifies this as an area of concern, and with good reason. There’s already evidence that attackers have targeted such systems—witness the 2018 attack that took the Knox County Election Commission site displaying results of the county mayoral primary offline during Tuesday. But attackers could take it even further. For example, an adversary could access election night vote displays to change the displayed results such that the real winner of the election is now the reported loser in the election, eroding voter confidence. There is no way around it: protecting democracy starts with protecting elections, a job that primarily falls to local jurisdictions. According to the CSI handbook, “Although states are heavily involved in setting the rules and policies for administering elections and in choosing election technology, in most states local jurisdictions administer and conduct the processes of an election.” Ideally, these local election administrators will follow recommended best practices by planning and establishing a DDoS mitigation strategy. NETSCOUT recommends the following: - Assess your election infrastructure DDoS attack landscape. - Create and implement a plan for emergency mitigation of a DDoS attack. - Establish and maintain effective partnerships with your upstream network service provider and know what assistance they can provide in the event of a DDoS attack. - Consider establishing relationships with companies that offer DDoS mitigation services, including managing your DDoS strategy as a service. The bottom line is that anyone responsible for the security of election infrastructure needs to be focused on potential technology disruptions. Find out more about hybrid and automated DDoS attack mitigation Wetherbee is a product marketing manager at NETSCOUT
<urn:uuid:cee88265-e087-4222-a725-9be319173054>
CC-MAIN-2024-38
https://www.netscout.com/blog/keeping-elections-safe-cyberattack
2024-09-17T16:03:03Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651800.83/warc/CC-MAIN-20240917140525-20240917170525-00261.warc.gz
en
0.943315
768
2.734375
3
Cybersecurity In-Depth: Feature articles on security strategy, latest trends, and people to know. Understanding TCP/IP Stack Vulnerabilities in the IoT Internet of Things devices are highly susceptible to attacks, breaches, and flaws emanating from issues within the TCP/IP network communications architecture. Here's an overview of what you need to know to mitigate risks. Although the Internet of Things (IoT) introduces remarkable ways to collect, manage, and apply data, it's also a huge vector for cyberattacks. One of the biggest vulnerabilities lies in embedded TCP/IP stacks, which combine applications, transport, network, and physical components. In many respects, this architecture was never designed for the IoT. Although engineers and developers have attempted to modify and add extensions to the TCP/IP stack— and many pieces are now open source — the complexity of the environment, combined with the reality that it was never designed with security in mind, has introduced numerous security challenges, along with real-world problems. "What makes the TCP/IP stack vulnerabilities notable is the sheer number of devices that are affected. The TCP/IP stack is a fundamental software component in every IoT device," explains Benson Chan, senior partner at Strategy of Things, a consulting and IoT implementation firm located in Hayward, Calif. Why Is TCP/IP Such a Threat for the IoT? At the most basic level, the TCP/IP architecture enables IoT devices to communicate with the network and each other. These stacks are open source and freely used by most embedded devices and IoT module manufacturers. "IoT device manufacturers then buy the chips and modules with the TCP/IP stack code already embedded from these suppliers to create IoT products," Chan explains. However, many of these manufacturers aren't aware that their devices are vulnerable, since they have no visibility into what stacks are used in the chips and modules that become part of IoT devices. What's more, it's not feasible or cost effective to analyze every single device to find and patch programming errors or other problems within the TCP/IP stack. As a result, all devices are highly susceptible to attacks, breaches, and flaws. These can lead to performance failures, data loss or corruption, and brand damage. It can also increase cybersecurity costs. "TCP/IP stack vulnerability management is becoming a real challenge for the security community," says Daniel dos Santos, research manager at Forescout. What Threats Exist? The extent of the problem is significant. Last year, a set of vulnerabilities dubbed URGENT/11 and RIPPLE20 made headlines. This year it's AMNESIA:33, with 33 zero-day vulnerabilities impacting four widely used open source TCP/IP stacks – uIP, FNET, picoTCP, and Nut/Net – that serve as foundational connectivity components for millions of IoT, OT, networking, and IT devices, including medical devices, industrial control systems, routers, switches, and smart home components. Attackers could use remote code execution, a denial-of-service (DoS) attack, or simply commandeer a device. Devices from upward of 150 vendors are at risk, according to Forescout, which reported the vulnerabilities last month. Flaws can reside in both commercial and open source components. Embedded components can include systems-on-a-chip (SoCs), connectivity modules, and OEM boards. IoT devices may span smart plugs, smartphones, sensors, and game consoles. OT systems comprise access controls, IP cameras, protocol gateways, and HVACs. Network and IT devices include printers, routers, and servers. "AMNESIA:33 changes the stakes not just because of the large number and critical nature of the vulnerabilities found, but also for several other reasons," dos Santos points out. This includes the widespread and heavy reliance on open source components and the deeply embedded nature of the flaws within hardware. Code from these stacks intersect with every network packet that touches the device, thus allowing vulnerabilities to affect idle devices. Since source code is reused in 88% of embedded projects, it acts as a force multiplier for vulnerabilities such as AMNESIA:33, de Santos says. Thus, attackers can use remote code execution (RCE) to take control of a target device and DoS to impair functionality and impact business operations. Attackers can also exploit an information leak to acquire potentially sensitive information and tap DNS cache poisoning to point a device to a malicious website, Forescout reports. "The widespread nature of these vulnerabilities means that many organizations around the world may be affected by AMNESIA:33," according to Forescout. How Can Organizations Address the Risk of TCP/IP Stack Vulnerabilities? Experts point to three foundational steps for dealing with TCP/IP stack vulnerabilities: identifying all devices on a network to understand which are vulnerable; assessing the risks introduced by these devices, which include their business context, criticality, and Internet exposure; and mitigating the assessed risks. "The last point can be achieved in several ways: patching devices when possible, segmenting the network and isolating critical devices, enforcing security compliance, and monitoring the network for malicious traffic," dos Santos explains. In regard to AMNESIA:33, he recommends disabling or blocking IPv6 traffic and relying on internal DNS servers whenever possible. "Several of the vulnerabilities affect these specific protocols in the stacks," he adds. It's also wise to tap cybersecurity solutions that can automate and optimize best practices. This includes taking a more proactive approach, "such as segmenting and isolating critical devices — whether or not they have known vulnerabilities — to reduce exposure and limit the impact of breaches," dos Santos says. An organization can also mitigate risk and potential damage by deploying IoT devices in segmented or isolated networks; staying on top of patches, policy updates, and device replacements; and implementing tighter and more granular controls over components and code, Chan says. Among the key questions security teams should ask: "What is the legacy of the code? Who is or has worked on it, and are there people still working on it?," Chan explains. "[Open source] libraries have simplified coding, but at the same time developers need to also understand what is in it. It is too easy to link to a library without knowing the code in it." To be sure, AMNESIA:33 and IoT vulnerabilities related to TCP/IP aren't going away. "Most of the vulnerabilities in the Amnesia:33 TCP/IP stack are caused by poor software development and management practices," Chan says. "Updating the software will address the vulnerabilities. But the real problem is knowing which devices have the affected stacks. IoT device manufacturers buy the chips and modules from suppliers, and the actual software stacks used are typically not specified or known by them." About the Author You May Also Like How to Evaluate Hybrid-Cloud Network Policies and Enhance Security Sep 18, 2024DORA and PCI DSS 4.0: Scale Your Mainframe Security Strategy Among Evolving Regulations Sep 26, 2024Harnessing the Power of Automation to Boost Enterprise Cybersecurity Oct 3, 202410 Emerging Vulnerabilities Every Enterprise Should Know Oct 30, 2024
<urn:uuid:b4a6a2e2-01c9-447a-886f-1356bdef9ecf>
CC-MAIN-2024-38
https://www.darkreading.com/iot/understanding-tcp-ip-stack-vulnerabilities-in-the-iot
2024-09-18T22:41:29Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651941.8/warc/CC-MAIN-20240918201359-20240918231359-00161.warc.gz
en
0.943999
1,480
2.796875
3
Cybersecurity is complex, confusing and continually changing. It can be hard to tell where to start and where to go. Even the terminology used for cybersecurity can vary from person to person, vendor to vendor or regulation to regulation. To address this problem, the National Institute of Standards and Technology created a set of guidelines and best practices called the Cybersecurity Framework. NIST first released the framework in 2014, and security experts across a broad range of industries, including the financial sector, began applying it. Read full article at Independent Banker.
<urn:uuid:d8530e12-6f04-4f32-b361-4e0b2930f3ac>
CC-MAIN-2024-38
https://www.mariakorolov.com/2016/strengthening-standards/
2024-09-09T05:47:45Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651072.23/warc/CC-MAIN-20240909040201-20240909070201-00161.warc.gz
en
0.940561
110
2.546875
3
Creator: University of Colorado Boulder Category: Software > Computer Software > Educational Software Tag: analysis, Coursera, engineering, management, technical Availability: In stock Price: USD 99.00 This course describes what it means to be an entrepreneur and clarifies some of the distinguishing characteristics of technology entrepreneurship. Entrepreneurial concepts, processes, & support systems are introduced from a holistic perspective. Methods and strategies for idea and opportunity recognition and development are provided to help identify new product ideas and solutions. Core concepts introduced include market gap analysis, sustainable competitive advantage, vision, mission, and customer value proposition. Getting Started with Startups can be taken for academic credit as part of CU Boulder's Master of Engineering in Engineering Management (ME-EM) degree offered on the Coursera platform. Interested in what the future will bring? Download our 2024 Technology Trends eBook for free. The ME-EM is designed to help engineers, scientists, and technical professionals move into leadership and management roles in the engineering and technical sectors. With performance-based admissions and no application process, the ME-EM is ideal for individuals with a broad range of undergraduate education and/or professional experience. Learn more about the ME-EM program at https://www.coursera.org/degrees/me-engineering-management-boulder.
<urn:uuid:6283fcdc-e92f-44de-bf95-d36a61eafd3d>
CC-MAIN-2024-38
https://datafloq.com/course/getting-started-with-startups/
2024-09-11T12:27:24Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651387.63/warc/CC-MAIN-20240911120037-20240911150037-00861.warc.gz
en
0.914738
271
2.609375
3
The world of web development is growing rapidly and serves as a tool for business development. Its services help companies improve product knowledge, sell products or services, and increase the company’s popularity rapidly. It also helps keep good communication between the company and the customer. The various trends and aspects of web development have different purposes in reaching particular audiences. The development of web apps is about to undergo a major revolution caused by the increase in data science. Web developers are now using Artificial Intelligence (AI) to make sense of all the data points and incorporate the results into apps from the design stage. This method helps companies save time and costs by examining their target groups’ specific behaviors and preferences. The web industry has been largely revolutionized by data science. With the help of data science, software development currently involves the coding or reprocessing of existing modules to create a working application that meets certain preset requirements. Advanced analytical tools can predict the features or functions common to users. These data science facilities help to satisfy the customers’ needs completely. Data science is helpful in predicting the users’ perceptions of how they are satisfied with the website and their requirements on the website, which helps to improve their needs. This helps to recognize emerging patterns in consumer behavior and to learn about the websites at regular intervals to meet the growing demands of end customers. AI-powered apps will soon be reliable user assistants. Smartphones can provide people with geolocation tips, past preferences, and interaction with specific brands even now. This makes it easy for programmers to customize sites so that they meet the specific requirements of a particular customer segment. Web applications that recall user preferences can save time and energy. Here, AI can learn about user behavior and time spent on certain websites. Thus, it can provide personalized advice and make a choice easier by reducing the users’ bad experiences. These apps can become a kind of private assistant, reliable partner, smart databases, or smart repositories. The changes made by data science in web app development will impact both consumers and developers. There are many advantages in using user data for web design, from improving the iterative design to quick information about how people interact with design and information. Check out: Top Web Security Solution Companies
<urn:uuid:4997acc1-6279-4193-a07e-caf2f4337dbe>
CC-MAIN-2024-38
https://www.enterprisetechnologyreview.com/news/leveraging-data-science-in-web-development-nwid-278.html
2024-09-11T12:42:30Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651387.63/warc/CC-MAIN-20240911120037-20240911150037-00861.warc.gz
en
0.936404
453
2.84375
3
10 Most Famous Indian Personalities In History The term "Indian" may relate to a member of the native, indigenous or aboriginal Americans commonly referred to "Red Indians" and also to the "West Indians" or the "Indian Indians". An assumption is that on visiting the East Indies, Columbus called its inhabitants as "Indios." A root of "Indians" is Hindu literally meaning "black," but mainly standing for the Hindu religion. From it has been coined "Hindustan" which is another name for "India". It is not known exactly if the great River Indus has something to do with the name "India". By far the most commonly understood meaning of "Indians" is the inhabitants of the subcontinent of India before 1947 and the citizens of the State of India after it. It is in this perspective that "Famous Indians of India" in history will be reviewed. Chanakya was born to Canin and Canesvari in the village Chanaka in 350 BC. He was the teacher of Chandragupta, the first Emperor of Maurya. In the recorded history of the subcontinent, he was the emperor of the entire kingdom. Known also as Kautilya and Vishnu Gupta, Chanakya wrote the oldest Indian book on politics called "Arthsastra". Although he was born centuries before Machiavelli, on account of his thoughts, he is sometimes referred to as the Indian Machiavelli. Having first been taught at Taksasila, Chanakya taught in the same five centuries old academy. His book Arthasastra was related to economics, welfare, international relations, and war strategies. In another book Neetisastra he enlightened on the social aspects of the Indian lifestyle and also extended some guidance for the rulers. He wrote 216 Neeti Sutras as do's and don'ts for the rulers. Chanakya died in 283 BC. 2. Abul Fazal Abul Fazal was born as the second son to Sheikh Mubarak in Agra, India on January 14, 1551. He was assassinated by Vir Singh Bundela during a trip. His severed head was sent to the Mogul Prince Salim whose accession he opposed. Abul Fazal was the vizier of the mogul Emperor Akber, who himself was not a learned person but liked the company of learned scholars. He had gathered in his court the top nine scholars from different fields, and they were known as "Nauratan" or nine jewels. Abul Fazal was the brightest among them. He had written the official history of Akbar's reign, known as Akber Nama comprising three volumes. Its third volume, Ain-i-Akbari has special importance in contemporary writings. He had also translated the Bible into Persian. 3. Mahatma Gandhi Mahatma Gandhi, also known as "Bapu" and "Gandhiji" was born to Putlibai (mother) and Karamchand Gandhi (father) in Porbandar, Kathiawar Agency, British-ruled India on October 2, 1869. He was assassinated by a Hindu Nationalist Nathuram Godsey on January 30, 1948. Mahatma Gandhi was a Hindu leader of the highest renown, more known as one of the two most prominent figures of the Indian Independence movement. His philosophy of Satyagrah and Ahinsa advocated non-violence. He led a simple life. He is honored officially as the Father of the Nation, and his birthday is commemorated as a national holiday. 4. Jamsetji Nasarwanji Tata Jamsetji Nasarwanji Tata was born to Nasarwanji and Jeevanbai Tata in Navsari, Gujarat, India on March 3, 1839, and died in Bad Nauheim, German Empire on May 19, 1904, at the age of 65. He was educated at Elphinstone College, Bombay, India. He considered three basic principles for the industrial progress of India being: steel as the mother of heavy industry, hydroelectric power as the cheapest source of energy, and technical education as essential for industrial advancement. He was an influential person not only in India but also being well-known in other countries too. George Washington encouraged him to visit Niagara Falls to review hydroelectric power generation. His son, Sir Dorabji Tata, fulfilled his dream by establishing the Tata Hydro-Electric Power Supply Company in 1910. 5. Amitabh Bachchan Amitabh Harivansh Bachchan, popularly known as Amitabh Bachchan, was born to Harivansh Rai Bachchan and Teji Bachchan in Allahabad, United Provinces in British India on October 11, 1942. He is the most famous Indian actor, singer, producer, and television presenter. He became known for the first time with his appearance in the Hindi Film Angry Young Man in the 1970s. He has appeared in more than 180 films since then. On account of his outstanding career, he has been honored with many awards a few of them being: National Film Award, Film Fare Award, Padam Shiri, and Padma Bhushan Civilian Awards from the Indian Government, Film Fare Lifetime Achievement Award, Superstar of the Millennium 2000, French Highest Civilian Award of France, Knight of the Legion of Honor. The BBC voted him "Greatest Star of Stage or Screen" in 1999. He was the first living Asian to be displayed as a wax statue in Madame Tussauds Wax Museum in London and later on in New York and Hong Kong. 6. Sachin Tendulkar Sachin Tendulkar was born to Ramesh Tendulkar and Rajni Tendulkar in Bombay, Maharashtra, India on April 24, 1973. Standing 5'5" (1.65 m) high, popularly known as "Little Master," right-handed Sachin Tendulkar is regarded as one of the Greatest Batsman of All Time. Next only to Donald Bradman, Sachin is the second greatest batsman. He is the second greatest all-time One Day International (ODI) batsman. Sachin Tendulkar broke Brian Lara's record for the most runs in test cricket on October 17, 2008, and became the first batsman to score 15,000 runs in test cricket on November 8, 2011. Tendulkar has been honored with the highest Indian Civil Award, Padam Vibhushan, and the second-highest civil award, Rajiv Gandhi Khel Ratna. He has been honored with many honorary degrees and other awards. 7. Kalpana Chawla Kalpna Chawla was born to Banarsi Lal Chawla and Sanjogta Kharbanda in Karnal, Haryana India on July 1, 1961, and she was one of the seven crew members who died on February 1, 2003, after the disintegration of the Space Shuttle Columbia on reentry into Earth's atmosphere during mission STS-107. She received the Congressional Space Medal of Honor. She received her early education from Tagore Public School Karnal. She earned a bachelor's degree in Aeronautical Engineering from Punjab Engineering College, Chandigarh in 1982, and a master's from the University of Texas, Arlington, and a Ph.D. in aerospace engineering from the University of Colorado at Boulder. As an Indian American Astronaut with NASA, she is a permanent chapter in the history of space engineering. She was posthumously awarded the NASA Space Flight Medal and the NASA Distinguished Service Medal. 8. Lakshmi Mittal Lakshmi Niwas Mittal, commonly known as Lakshmi Mittal, was born in Sadulpur, Rajasthan, India on June 15, 1950. He is the chairman and Chief Executive Officer of ArcelorMittal, the world's largest steel manufacturing company. He was educated at St. Xavier's College, Calcutta. He is best known as a steel tycoon. In addition to being Chairman and CEO of ArcelorMittal, he is also Director of Goldman Sachs, owner of Karrick Limited, Co-owner of Queens Park Rangers F.C. He is the richest man in India, the U.K., and Asia, the second richest in Europe, and the sixth richest person in the world with personal wealth over $20.7 billion. He is the 44th most powerful person of the Forbes list of 68 people. Financial Times named him "Person of the Year" in 2006 and the Times counted him as one of the "100 Most Influential Persons in the World." In the recorded history of the world, the wedding of his daughter Vanisha Mittal was the most expensive. 9. Mirza Ghalib Mirza Asadullah Khan Ghalib was born to Abdullah Baig and Izzat-ul-Nisa Begum in Agra, Mughal Empire on December 27, 1797, and died in Delhi, Punjab in British India on February 15, 1869, at the age of 72. He was the most famous and greatest poet of Urdu and Persian. Last Moghul Emperor of India, Bahadur Shah Zafar, was a poet himself and was tutored by the famous poet "Zauq" who was Ghalib's rival. Both Ghalib and Zauq, however, agreed upon the greatness of Meer Taqi Meer, best known for his simplicity and meaningfulness of his "Ghazal" the most popular genre of Urdu and Persian poetry. The Emperor honored Mirza Ghalib with three titles; Najm-ud-daulah, Dabir-ul-Mulq, and Mirza Nosha. Diwan-i-Ghalib is the most known book of poetry in Pakistan and India. 10. Kundan Lal Saigal Kundan Lal Saigal was born to Amarchand and Kesar Bai in Jammu, British India on April 11, 1904, and died in Jalandhar, Punjab, India on January 18, 1947, at the age of 42. He was the most famous vocalist, actor, and playback singer of his time. He was a school dropout and did some odd jobs like timekeeping for the railway and salesmanship for Remington Typewriters prior to being introduced to the film industry. In 1933, his four Bhajans in the film "Puran Bhagat" made him known throughout India. Later on, he emerged as the most remembered superstar of the Indian film industry. Saigal was an idol for the post-independence generation by virtue of his unique voice. Among those who idolized his voice were the superstars like Lata Mangeshkar, Mohammad Rafi, Mukesh, and Kishore Kumar.
<urn:uuid:2aa3c4c6-5fd4-4f07-ab69-ccd454b569d9>
CC-MAIN-2024-38
https://www.knowledgepublisher.com/article/1122/10-most-famous-indian-personalities-in-history.html
2024-09-14T00:10:52Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651540.77/warc/CC-MAIN-20240913233654-20240914023654-00661.warc.gz
en
0.98377
2,252
3.109375
3
Elevate the power of your work Get a FREE consultation today! The United Nations designated International Holocaust Remembrance Day-January 27, the anniversary of the liberation of Auschwitz-Birkenau-to remember the six million Jewish victims and millions of other victims of the Holocaust. International Holocaust Remembrance Day was designated in 2005 on the 60th anniversary of the liberation of the Nazi concentration camps, the largest Nazi concentration and death camp, and the end of the Holocaust. Based in Washington, D.C., the United States Holocaust Memorial Museum is a physical and digital reminder of the Holocaust and the fragility of democracy. Iron Mountain is proud to support the Museum, specifically for the discoverability and digitization of archival materials from the Nuremberg Trials, where Nazi leadership were prosecuted after WWII. The Nuremberg Trials, and the US role in those trials, stands as a crucial part of that story and one that is in great demand for further scholarship and refuting the Holocaust. The Museum anticipates that once digitized, these collections will be accessed worldwide. Over 250,000 pages containing post-war documentation will be organized and digitized that share the atrocities of the Holocaust and set the stage for international law. “History has much to teach us, both about where we come from and where we want to go,” said Alisha Perdue, manager community engagement. “Iron Mountain’s Living Legacy Initiative is committed to helping preserve and make accessible cultural and historical information and artifacts. We are honored to support the Museum and pay our respects on International Holocaust Remembrance Day, which serves as a reminder of the atrocities of war, the evils of man, and the cost of prejudice and hate.” The Nuremberg Trials records include those of U.S. prosecutors Benjamin Ferencz and Robert M.W. Kempner as well as the papers of Judge William C. Christianson. Iron Mountain’s support will fund selections from the collection’s cataloging and finding aids for research and digitization. Once discoverability is implemented, materials can be accessed by scholars of many different disciplines, exhibition teams and other educators and academics for educational use. "The digitization of these selections from our extensive Nuremberg Trials collections will help reveal more details of the Allied prosecution of these war crimes and the behavior of the accused perpetrators, a very critical part of Holocaust history," said Dr. Rebecca Boehling, director of the Museum's National Institute for Holocaust Documentation. "We are so grateful to Iron Mountain for this generous investment in our collections, making them available to not only our visiting scholars and researchers, but to the general public through our online collections search." A living memorial to the Holocaust, the Museum inspires citizens and leaders worldwide to confront hatred, prevent genocide, and promote human dignity. Iron Mountain's IT data management and protection services can help you overcome the challenges of data with lifecycle management amd ensure data protection. The Living Legacy Initiative is Iron Mountain’s commitment to preserve and make accessible cultural and historical information and artifacts. Information Management has changed over the years and so have we. Discover how we've adapted every step of the way in our 70-year history. Gain flexible, affordable, and secure vital records storage in our underground vault Get a FREE consultation today!
<urn:uuid:a5db6b61-2515-47ce-9d63-23915126a101>
CC-MAIN-2024-38
https://www.ironmountain.com/about-us/sustainability/stories/i/international-holocaust-remembrance-day
2024-09-15T07:17:22Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651616.56/warc/CC-MAIN-20240915052902-20240915082902-00561.warc.gz
en
0.91879
682
3.234375
3
The 3D animations bring to life the technologies that make wireless networks possible. Watch how signal encoding, radio transmission and data encryption deliver the internet to devices all throughout your home. The basic 4 parts of WIFI - (WiFi stands for Wireless Fidelity) Frequency band - Most Wi-Fi uses either 2.4Ghz or 5Ghz frequency bands Channels - The main 2.4Ghz and 5Ghz signals are divided into sub-frequencies called channels. Packets - Channels carry packets, or small groups of data. Waves - Packets are made up of sets of waves. Author: Sam Carson - Sam graduated from NC State University with a degree in Meteorology. He loves to take casual hikes to Crowders Mountain and is often told he looks like the actor Woody Harrelson." This was brought to our attention by - Rob Mitchell a Communication Specialist for The Bookmark on VerizonInternet.com, a Verizon authorized retailer.
<urn:uuid:7ee95353-7e99-4240-a75d-9637ef890611>
CC-MAIN-2024-38
https://www.networkdatapedia.com/post/2018/03/20/how-wi-fi-works-a-visual
2024-09-17T18:29:29Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651829.57/warc/CC-MAIN-20240917172631-20240917202631-00361.warc.gz
en
0.925696
204
2.796875
3
While most AI in use today can be classified as early-stage advanced analytics, some enterprises have built large data science teams to apply machine learning and deep learning algorithms to business processes. Many of these enterprises have built the necessary support infrastructure to train these algorithms on large data sets, deploying the resulting AI models to production to generate business insights. However, many consumer and business consumption patterns changed dramatically in 2020, causing these advanced AI models to fail or behave erratically. Many of these models that have been trained to make predictions based on historical data have not been able to deal with the data anomalies created by disruptive business conditions and changing preferences. Companies using AI for insights had a hard time making use of existing models in production. Many companies have reverted to using the standard analytics dashboards and tools that were in place before AI adoption. The impact of macroeconomic conditions on AI models has been greatest in the customer service area, where next-best action and other recommendations are based on predictions from historical data. With data about changing economic patterns, sentiment analysis models predict negative customer expectations. What this demonstrates is that machine learning techniques, such as supervised and unsupervised learning that depend on massive amounts of data (Figure 1), do not work when economic patterns change dramatically because the existing data is no longer relevant and useful for predicting meaningful business insights. As machine and deep learning techniques evolve, two promising approaches are reinforcement learning and deep reinforcement learning. These learning techniques are less dependent on data than traditional machine learning and offer greater insights when conditions are changing. What is reinforcement learning? Reinforcement learning is the process of a program learning as it accounts for the consequences (both good and bad) of actions it undertakes, such as predicting an insight. Based on these consequences, the program adjusts its logic, effectively learning to solve a problem over a period of time rather than by trying to produce insights from current data sources. Instead, the process is generating the data it needs from the actions it takes. Using this data and knowledge of the consequences of previous actions (collectively called state space), the program performs new actions to generate a beneficial consequence, which in turn produces more actionable insights. Reinforcement learning is not dependent on input data sets as is the case with other machine learning algorithms. The actions and their consequences determine the generation of training data that helps the algorithm learn. How to generate useful machine learning-based insights To take advantage of machine learning in order to generate useful insights in a rapidly changed macroeconomic environment, enterprises will need to make some significant modifications to their approach. One way forward is to combine retraining current models with the adoption of reinforcement learning models and commit greater resources to employing experienced data science teams. To start with, AI models currently in use need to be redeployed with new versions that should be trained with current business data. The models should be versioned along with the data sets used to train a version of the model. Versioning should include parameters used to tune the model. Next, enterprises could use the opportunity to train these models on once in 100- to 500-year business disruption events. Thanks to the fact that data providers have been collecting data on business disruptions during unexpected events, a plethora of data is now available to train existing models. At the same time, the parameters used to tune the model with this additional information should account for anomalies in the data. Trained data sciences team — integral to the process Integral to the whole process are data science teams. First, enterprises need to ensure they have an experienced team in place, and these data scientists need to be focused on correcting any anomalies in the insights that the models are predicting. This means data scientists should not be pushed aside in favor of “citizen data scientists,” users who are not employed as statisticians or analysts but use tools that automate some aspects of data science. While there is certainly an important role for the citizen data scientist, these individuals are not trained to identify anomalies, and this skill set is vital to addressing the problem with using AI models in unpredictable environments. In addition, enterprises should increase the use of external data to train the models since having a more diverse set of data will help to normalize the dependence of insights from internal data. However, which data and how much of this data will be relevant to the given situation is something only data science experts can identify. The final step is to make use of reinforcement learning in parallel with current models to get another set of data points that will help in making decisions. Pricing optimization — an insurance industry example The benefits of reinforcement learning in generating useful insights is well illustrated in the insurance industry (Figure 2). Insurers are increasingly using predictive algorithms to price life insurance policies — both new and renewals — as part of a broader model that supplements the traditional role of actuaries. Actuaries, whose knowledge and expertise in statistics is integral to projecting risk and pricing in the industry, are now working with data scientists, who are increasingly being brought in to collaborate on initiatives that will enable the business to optimize premiums. The current price optimization technique involves using data sets that include competitor prices, individual customer information — their buying patterns and socio-demographic profiles — market prices for similar insurance products and data about macrotrends in the economy. This data needs to be prepared — from data cleansing, to moving the data into the AI model and then managing it — so machine learning algorithms can process it. To better prepare the data, data scientists use various algorithms and techniques — from linear models to more sophisticated neural network models — to understand changes in premiums and their impact on sales and other business metrics such as combined ratio. Based on predictions from these models and after adhering to local regulatory requirements, they are able to price premiums for both new business and renewals, which is increasingly resulting in personalized premiums for individual customers. However, changing macroeconomic conditions and lower demand have made these models more erratic. To continue to gather valuable insights for price optimization, machine learning algorithms that depend on data sets to make predictions need to be tuned based on newly identified parameters, and then redeployed. This is where reinforcement learning excels, helping insurers make sense of the data in the context of changing economic circumstances. The way a reinforcement learning algorithm learns in this context is by setting a baseline strategy of increasing premiums where there is new business and retaining a defined threshold level in the case of renewals. The algorithm would start with a set of customers who face an increase in premium (based on business strategy and multiple financial metrics) and learn from the consequence — either increased or decreased sales or retention at a threshold level or below. This consequence results in the algorithm learning the best action for the next set of customers. Over time, it learns what works and what does not work in terms of meeting business strategy and related financial metrics. The result of the algorithm could be used by the insurer in conjunction with prediction from the current machine learning models to make a final decision on setting premium prices for new policies and renewals. Once that decision is known, the insurer would feed the result back to the algorithm for it to improve its logic for future decision making. To adjust to changing economic circumstances in AI modeling, enterprises should expand their analytics and decision-making capabilities beyond traditional predictive analytics that rely on large, historical data sets. Adding reinforced learning algorithms will help models learn over time with much less input data. Combined with current analytics and machine learning capabilities, reinforced learning provides enterprises with valuable input, helping them to gather relevant insights and make good business decisions even during sustained unexpected business disruption. Learn more about the insurance industry.
<urn:uuid:5de11659-c38a-4577-8bea-2c32d0e82bda>
CC-MAIN-2024-38
https://dxc.com/us/en/insights/perspectives/paper/improving-efficacy-of-ai-models-during-times-of-business-disruption
2024-09-19T01:45:21Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651944.55/warc/CC-MAIN-20240918233405-20240919023405-00261.warc.gz
en
0.941045
1,555
3.109375
3
Cloud computing is changing the way we do business. It’s estimated that 60% of all corporate data is stored on the cloud ”“ twice as much as in 2015 ”“ while global cloud infrastructure spending is projected to reach $118 billion by 2025. Cloud computing means a vast increase in storage capabilities and increased flexibility, opening the door to unprecedented levels of productivity and innovation. But for many, the cloud also brings uncertainty. That’s because this new landscape brings about a whole new way of thinking when it comes to privacy and security. Cloud security management involves a different level of complexity than before, with new tools, techniques, and technologies, along with new processes and methods. In this article, we break down the key aspects of cloud security, from the different types of clouds to the main risks and the challenges companies will face today and in the future. Cloud security is a system of rules, processes, and technologies used to protect the cloud services used by a company and the data and applications stored and used within the cloud. While adoption is steadily increasing, there’s still some apprehension when it comes to migrating to the cloud. There are many exaggerated fears about the nature of cloud computing, causing some companies to be left behind. However, as we’ll show in this article, the question isn’t whether the cloud itself is safe, it’s whether it is being used in a safe way. The most common type of cloud, public clouds are owned and operated by third-party providers and created specifically for individuals or businesses. Today, the majority of the public cloud market share belongs to its five largest providers: Amazon Web Services (AWS), Google Cloud, IBM Cloud, Microsoft Azure, and Alibaba Cloud. These companies essentially lease their cloud services for public use. The public cloud is cost-effective and hassle-free ”“ companies don’t have to worry about managing or maintaining servers or infrastructure, while security issues are dealt with by the provider. However, due to its multi-tenant nature and a high number of access points, public clouds have a larger threat surface. This means that, while the best public cloud providers offer a robust infrastructure, with expert security measures in place, there’s a larger risk of security threats. Also, while the providers themselves are responsible for cloud security monitoring, cloud security management, and the overall infrastructure, the data security within the cloud and how it’s used is the responsibility of your company. Unlike public clouds, private clouds are created for and used by a single dedicated end user or group. This means they have completely isolated access to the cloud. Until relatively recently, most private clouds were sourced from on-premise IT infrastructure. However, they can now be created in rented data centers. This leads to cloud security management provided by a third party ”“ a managed private cloud. Private clouds are often considered to be more secure than public clouds. After all, they have better isolation, unlike public clouds, which are based within an environment that’s shared by many. Nevertheless, this all depends on the strength of the company’s security and the expertise of its cybersecurity team. Private cloud security can also be more costly due to leasing expenses and extra hardware costs, although most private clouds provide a similar level of security to public ones for a similar price, aside from the infrastructure management that the provider offers. Seen as an ideal blend between public and private clouds, hybrid clouds are essentially a single IT environment made up of several distinct groups. They’re kind of like several smaller clouds within a larger cloud, where data can be shifted from public to private and vice-versa, depending on specific security requirements. Hybrid clouds ensure a higher level of flexibility and scalability. However, they also bring a higher level of complexity, are harder to implement, and make cloud security management more difficult. Still, the use of hybrid clouds is on the rise, particularly for larger organizations. In 2020, the hybrid cloud market had an estimated value of $52 billion. With a compound annual growth rate (CAGR) of 18.73%, it’s expected to reach $262 billion by 2027. In fact, according to a survey by Cisco, 82% of large IT organizations are already using some form of hybrid cloud. These are similar to hybrid clouds but with a key difference ”“ multi-clouds generally contain more than one public cloud, while hybrid clouds only contain one. The different clouds within a multi-cloud approach are generally not orchestrated or integrated. Instead, each cloud within a multi-cloud environment handles specific tasks. Multi clouds enable next-level availability, as an entire workload can be shifted from one vendor to another in the event of a temporary issue. This saves on downtime. They can also be very cost-effective, as companies aren’t liable to pay for off-premise or in-house data centers. Also, taking a multi-cloud approach means taking the best points from different providers, giving companies holistic, tailored cloud security solutions. This approach is favored by large companies that want to use the best parts from different vendors for optimal cloud security. Like hybrid clouds, multi-clouds require a high level of cloud security management expertise to implement and manage. IaaS means a cloud provider does all the cloud management for you ”“ from the servers and the network to the actual data store and the connection to the internet. You rent the cloud infrastructure from the provider. IaaS is rapidly gaining in popularity. According to a study by Gartner, the IaaS market grew from $64.3 billion in 2020 to $90.9 billion in 2021. It’s a monopolized market though, with the top five providers ”“ Amazon, Microsoft, Alibaba, Google, and Huawei ”“ taking up an 80% market share. Typically used by professional developers or programmers, PaaS provides a shared environment for app development and management. It allows developers to work together and share data without the need to build and maintain separate infrastructures. A key part of DevOps, the PaaS market is projected to reach $319 billion by 2030, with a CAGR of 22% between 2021 and 2030. CaaS enables each service or team to manage, run, scale, and operate their own package of software, or ”˜container.’ This is ideal for a microservices approach to software development that consists of several separate services. CaaS makes it a lot easier to deploy and scale distributed systems, without the need to manage or monitor the actual cloud architecture. SaaS enables the use of a dedicated application to a group or team, along with maintenance and management services of that application. This takes away the need for apps to be installed on individual users’ machines ”“ instead, the team can access the app through a cloud, where they can collaborate and share information. Examples of SaaS platforms include Microsoft Teams, Slack, Salesforce, Dropbox, and Creative Cloud. Migrating to the cloud means thinking differently about cybersecurity, as it’s a completely different environment from traditional IT systems. The three main things to consider are: Dealing with data becomes more complex with the introduction of a cloud environment, which means having to understand new methods to protect it. Of course, the complexity depends on the architecture of the cloud, whether public, private, hybrid, or multi-cloud, as well as the platform type, whether SaaS, PaaS, CaaS, or IaaS. This involves analyzing the new environment to identify potential weaknesses, thereby mitigating any threats before they happen. It’s a continuous process and is crucial to good cloud security management. Examples of cloud vulnerabilities are poor access management, security gaps, regular data loss, and non-compliance. Having control of who has permission and access to the cloud is a big part of cloud security management. IAM, or identity and access management, helps to pinpoint and mitigate security risks, reduces service costs, and boosts productivity. Having robust critical infrastructure protection in place is a major part of cloud security management. It provides visibility to attack threats, protects hardware and software assets, and increases your company’s security posture. Infrastructure protection is generally implemented by experts in cloud security management. A set of processes, controls, and policies within a cloud app, application security protects those apps from threats. Application security is an ongoing process and involves developing, testing, and adapting features within the cloud application. When it comes to cybersecurity, preparation is key. That’s why having an incident response plan (IRP) in place is so essential. It acts as a roadmap for how to respond when an attack takes place. A well-designed, multifaceted IRP saves on operational downtime, protects company revenue and data, and takes a proactive approach to cybersecurity. Cloud security is a complex topic that, understandably, makes people apprehensive. But, while moving towards cloud technology can be daunting, it’s important to understand that the way you use the cloud is key. That’s why experts in cloud security solutions like Ackcent are increasingly sought after. We’re passionate about helping people transition onto the cloud, but our work doesn’t end there. Once we’ve created a robust infrastructure based on your company’s precise requirements, we help your company become more productive and achieve the kind of growth that wasn’t previously achievable. So be it public, private, hybrid, or multi-cloud, and whether via IaaS, PaaS, CaaS or SaaS; we’re here to help you transition to the cloud with ease, then build towards a better, brighter, more secure future. Are you ready to find out more about cloud security? Contact Ackcent today! Get resources in your mailbox for free
<urn:uuid:91775e69-07fd-40ee-9fd8-c00e51483623>
CC-MAIN-2024-38
https://ackcent.com/a-beginners-guide-to-cloud-security/
2024-09-20T06:36:40Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652138.47/warc/CC-MAIN-20240920054402-20240920084402-00161.warc.gz
en
0.942439
2,071
2.609375
3
The fact that businesses can face fines up to $22.07 million or 4% of their annual global revenues, whichever is higher, has also sent companies scurrying to come into GDPR compliance—or at least make a genuine attempt to comply with the regulations. But going through the official legal text will only end up glazing your eyes with the sheer number of dense legal jargon terms filling the document. Don’t worry. Today, we’ll translate one of the most important chapters of GDPR, Article 25, to make it easier for you to understand and implement. What is GDPR Article 25 Anyway? Article 25 of the GDPR wants you to consider data protection and security risks before you do anything. According to its requirements, you shouldn’t collect any more data than you need, and whatever data you do collect, you should pseudonymize. Below, we’ll show you what the legal text says about Article 25, along with an easy-to-understand explanation so that your brain doesn’t get repulsed by the legalese. Note: Before getting into the nitty-gritty details, you should know what a controller is to make it easier for you to follow the text. Controllers are individuals or entities that determine the purpose and means of processing personal data either alone or in association with others. Now, let’s break down the different clauses. 1) Taking into account the state of the art, the cost of implementation and the nature, scope, context and purposes of processing as well as the risks of varying likelihood and severity for rights and freedoms of natural persons posed by the processing, the controller shall, both at the time of the determination of the means for processing and at the time of the processing itself, implement appropriate technical and organisational measures, such as pseudonymisation, which are designed to implement data-protection principles, such as data minimisation, in an effective manner and to integrate the necessary safeguards into the processing in order to meet the requirements of this Regulation and protect the rights of data subjects. Article 25(1) specifies the requirement for data protection by design. It requires the controller to implement appropriate technical and organizational measures to effectively implement data protection principles. The controller is also responsible for integrating the necessary safeguards into the processing to protect the rights of data subjects. In other words, the controller is responsible for the design and implementation of their company’s data protection. 2) The controller shall implement appropriate technical and organisational measures for ensuring that, by default, only personal data which are necessary for each specific purpose of the processing are processed. That obligation applies to the amount of personal data collected, the extent of their processing, the period of their storage and their accessibility. In particular, such measures shall ensure that by default personal data are not made accessible without the individual’s intervention to an indefinite number of natural persons. Article 25(2) specifies the requirement for data protection by default. It requires the controller to implement appropriate technical and organizational measures so that only personal data that’s necessary for a specific purpose is processed by default. This obligation extends to all the collected data, the extent of processing, the storage period, and the accessibility of the stored personal data. 3) An approved certification mechanism pursuant to Article 42 may be used as an element to demonstrate compliance with the requirements set out in paragraphs 1 and 2 of this Article. Article 25(3) states that if the controller adheres to an approved certification under Article 42, they can use this certification as a way of demonstrating compliance with the above requirements. How GDPR Article 25 Works Article 25 has two key principles—privacy by design and privacy by default—that underlie the entire GDPR. For instance, Article 5(1) requires data processing to be limited to what is absolutely necessary given the purpose for which the data is collected (privacy by design) and be limited to only those who need to access the data (privacy by default). Similarly, Article 32(1)(b) outlines businesses to consider the ongoing confidentiality, integrity, and availability of processing systems and services. Despite the mandatory pseudonymization and data minimization technical measures, Article 25 allows controllers to determine which additional technical measures to take based on their respective data security and privacy requirements. While they do have flexibility in that aspect, the controller has to follow certain organizational and technical strategies to comply with Article 25. Here’s a rundown of what these strategies can include: - Anonymizing or pseudonymizing data by not copying production databases for development, testing, or analytics. - Avoiding storing spreadsheets and other data sources in a local folder or in cloud-based storage applications (Eg: Dropbox, OneDrive, Google Drive, etc.). - Making encryption necessary for emails containing identifiable personal data. - Restricting email archive access to privileged users and monitoring their activity regularly. - Creating and implementing policies about bringing your own devices to see secured data. - Protecting personal data at rest, in motion, and in use using an existing database format. - Performing security activities like training staff, carrying out internal audits of processing activities, reviewing policies, and documentation of compliance. - Having ethical walls: Maintaining strict separation between business groups to comply with M&A requirements, government clearance, and so on. - Data masking: Anonymizing data via encryption/hashing, generalization, etc. Here, all the sensitive data is replaced by fictional yet realistic data in a bid to maintain operational and statistical accuracy. - User rights management: Identifying excessive, inappropriate, unused, and/or unnecessary privileges. - Privileged user monitoring (PUM): Monitoring privileged user databases access and activities, and blocking any suspicious access or activity, if needed. - User tracking: Mapping web application end-user to the shared application/database user to the final data accessed. - VIP data privacy: Maintaining strict access control on critical data, such as data stored in multi-tier enterprise applications like SAP. Example #1: Pseudonymization to Implement Data Protection by Design You can replace personally identifiable material with artificial identifiers–this is what pseudonymization means. To boost the level of security further, encrypt messages and emails in a way that only authorized people can read them. Example #2: Ensuring Limited Access to Implement Data Protection by Default Social media platforms should be encouraged to set user profile settings in the most privacy-friendly settings. For instance, accessibility to every user’s profile on a social media channel should be limited from the very beginning so that it isn’t accessible by default to just anyone. How to Get Started With GDPR Article 25 Article 25 of GDPR is formulated to prepare companies considering data privacy and data protection in every aspect of their business, ranging from product development to operations to rendering services. To get started, you only need to remember and implement the three clauses of Article 25. Here’s a step-by-step rundown of how you can comply with the requirements of this specific GDPR chapter: Step 1: Understand What to Do You must set up appropriate technical and organizational measures to implement data protection principles while safeguarding individual rights. Since there is no one-size-fits-all solution or method, you’ll have to select a set of measures based on your unique circumstances. The important thing here is to consider data protection issues from the beginning of any processing activity. You can then adopt appropriate policies and measures that meet the requirements of data protection by design and by default. Below are a few examples of how you can do this: - Ensuring complete transparency with regard to functions and processing of personal data - Minimizing the processing of personal data - Prompt pseudonymization of personal data - Creating and improving security features - Allowing individuals to monitor the processing of personal data Remember, this isn’t an exhaustive list. There are so many other measures to comply with data protection by design and default. So keep analyzing and improving to figure out what works best for you. Step Two: Understand When to Do It For implementing data protection by design, you have to start at the initial phase of any system, service, product, or process. You can begin by considering your intended processing activities, the risk to individuals, the possible measures you can take to comply with the data protection principles, and what you can do to protect individual rights. These considerations should include the following: - The costs of implementation of any measures - The nature, scope, context, and purpose of your processing - The risks your processing poses to the rights and freedoms of data subjects Think of it as an information risk assessment for your security measures. You can then take these considerations and move on to the next step, where you implement the data protection principles by applying actual technical and organizational measures and integrating safeguards into your processing. It’s due to this that there’s no standard or uniform solution or process that you could potentially apply to every organization or processing activity. That said, you can apply measures: - When you are at the design phase of any processing activity - During the life-cycle of your processing activity Step 3: Understand How to Put It Into Practice You have to develop a set of actionable and practical guidelines for your organization, framed by your assessment of the posed risks and measures you can take. Putting these concepts into practice, though, will depend on your respective circumstances—who you are, what you do, what resources you have available, and the nature of the data you process. Of course, whatever organizational approach you take should achieve specific outcomes. To ensure this, you can: - Consider data protection issues as part of the design and implementation of systems, services, business practices, and products - Make data protection an essential component of the core functionality of your processing systems and services - Process only the necessary personal data in relation to your purpose, and make sure you use it for that purpose only - Adopt a ‘plain language’ policy for public documents so that individuals can understand what you plan on doing with their personal data - Make the identity and contact information of those responsible for data protection public so it’s freely available to individuals as well as within your organization - Give individuals tools that help them understand how you’re using their personal data, and whether you’re enforcing your policies properly - Offer strong privacy defaults, user-friendly options and controls, and respect user preferences
<urn:uuid:e909027e-5cc4-4901-ac6f-93fe7d7747fd>
CC-MAIN-2024-38
https://nira.com/gdpr-article-25/
2024-09-20T07:42:13Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652138.47/warc/CC-MAIN-20240920054402-20240920084402-00161.warc.gz
en
0.907214
2,189
2.71875
3
Cloud computing is upon us this year in 2010, showing us how much better our data-center universe can be. Applications and processing technologies are advancing so quickly that they are difficult to keep up with and even more difficult to plan for. Only those companies that have lots of capital and a fast track to market will be able to effectively deploy these new platforms and lead us into the next generations of computing. Clouds will accelerate the operating and energy efficiencies of the computing industry across the board. For starters, they allow a business to better manage and control IT demands resulting in the downsizing and more efficient deployment of their IT systems. According to Jason Stowe of Cycle Computing, the cloud-computing model enables three fundamental forms of efficiency: - Managing the Peaks: Organizations no longer have to purchase and power infrastructure to handle their peak needs for an application, only to see it sit idle. - Economies-of-Scale Efficiencies: By provisioning large-scale server capacity, used by multiple applications only when it is needed, and charging only for the resources an application uses, cloud helps drive costs down and make deployment more cost efficient. - Increased Reliability, Lower Costs: Properly handling disaster-recovery scenarios requires provisioning of multiple data centers and networking contracts, which increases costs and lowers efficiency. Clouds reduce this by removing complexity and enabling access to multiple compute and storage sources with large economies of scale for increased reliability and lower costs. These forms of IT efficiency all result in reduced demand for lower power and cooling, and subsequently the downsizing and more efficient deployment of mechanical and electrical support systems, as well. How exactly will these advances in applications technology change the way we design and build the networks and facilities that will support the cloud? And, what will they do to our now possibly outdated plans to virtualize, consolidate, and improve our operating and energy efficiencies? Do we already need to start planning for the move of our facilities into the next generation of cloud computing? Different Strokes for Different Folks The first thing that should become evident over the next year or so is the type of cloud computing that can succeed for different business models. Risks associated with the security and service delivery of cloud computing will quickly differentiate the operating environments acceptable for corporate enterprise computing from those for Internet search and networking. The risk tolerance defining the design basis of the networks and facilities for those two environments will differ as much as the cloud environments that they support. Internet search and network businesses will thrive on flexible and inexpensive architectures like the Amazon Elastic Compute Cloud EC2. This type of “public” cloud environment provides redundancy across a broad network of data centers and requires lower levels of redundancy and reliability in individual data centers. Because of that inter-regional network redundancy, the risks associated with the operation of individual servers are more tolerable. This allows for environmental conditions that approach extremes and the minimization of power and cooling systems redundancy. Yahoo’s chillerless designs and Google’s battery-on-board servers both help their data centers achieve PUEs of about 1.1 and are good examples of the efficiencies that can be achieved with “public” cloud architectures.The enterprise environment, however, will prove to be less tolerant of security and service delivery risks and will demand that cloud environments be restricted to “internal” clouds comprised of very secure private networks and individual facilities. And, although redundancies and efficiencies will be provided within the cloud, each data center will be need to be more fault tolerant and the reliability of power and cooling will continue to be of a high priority to the operating environment. On the other hand, the cloud redundancy may allow these facilities to take advantage of component efficiency improvements such as the newly developed “by-pass” UPS systems offered by Eaton and the now more commonly accepted data center air economizers, and allow the enterprise spaces to approach PUEs similar to those achieved by search engine data centers. Sun Microsystems manages much of its cloud operations out of the 1,000,000 square foot SuperNAP data center complex in Las Vegas, where its high-performance servers are housed in high-density computing pods known as a T-SCIFs (thermal separate compartment in facility). The racks are packed top to bottom with servers, creating a power load of up to 24 kilowatts per rack, all of which are cooled by outside air and sophisticated airflow controls that minimize cooling costs year round. The facility is destined to support the delivery of “complete, engineered, and integrated systems,” with an emphasis on the new Oracle stack that spans applications, middleware, database, and hardware. Microsoft’s new $500 million data center in Chicago is one of the largest independent data centers ever built, spanning more than 700,000 square feet. But it’s also one of the most unusual, with its garage-like lower level optimized for 40-foot shipping containers packed with web servers, while a second story houses traditional raised-floor data center space. Containers packed with servers and, in some cases, equipment to power and cool the servers occupy about a dozen parking spaces in their “container canyon.” Each row of containers plugs into a spine providing hookups on the lower level, with power distribution equipment on a mezzanine level. With double-stacked containers, portable stairs can be wheeled in to provide access for maintenance. The data center increases hardware utilization, reduces use of resources like water and electricity, and reduces waste material. The data center is the next evolutionary step in Microsoft’s commitment to thoughtfully building its cloud computing capacity and network infrastructure. IBM just recently announced the opening of a new data center designed to support cloud computing in order to help clients from around the world operate smarter businesses, organizations, and cities. The new data center reduces technology infrastructure costs and complexity for clients while improving quality and speeding the deployment of services--using only half the energy required of a similar facility its size. The data center will ultimately total 100,000 square feet at IBM’s Research Triangle Park (RTP) campus and is part of a $362 million investment by the corporation to build the new data center in North Carolina. IBM owns or operates more than 450 data centers worldwide. IBM has engineered the data center to help its clients use new Internet technologies and services to meet the business challenges of an environment marked by an exponential rise in computational power, a proliferation of connected devices and an imperative to manage energy costs. Path to the Cloud Amongst others, the U.S. Government is making similar commitments to cloud computing operations in a very big way. After receiving a $2.4 billion increase in its 2009 budget to improve IT systems and services, the National Aeronautics and Space Administration (NASA) just announced a halt of its New Enterprise Data Center (NEDC) plans for data center consolidation, an upgrade of its wide area fiber backbone, and the implementation of outsourced data center capabilities, which the agency hoped would improve the efficiency and security of its data center services. NASA has decided to rethink its strategy to better represent the philosophy of its new IT leadership as well as new federal requirements related to cloud computing, greening it, virtualization, and federal data center consolidation guidance. NASA reexamined the NEDC acquisition strategy and concluded it did not fully address future NASA enterprise requirements. Based on commercial best practices and other federal data center consolidation lessons learned, NASA intends to create a data center consolidation plan to incorporate all data centers, systems, and applications. The new data center plan will include a data center architecture and full enterprise assessment, which will allow NASA to design an infrastructure strategy to address all business requirements while taking advantage of opportunities to reduce energy costs and utilize innovations like cloud computing. The plan will also allow for optimal NASA data center consolidation and make appropriate decisions about utilization of data center facilities, modernizing existing NASA facilities or outsourcing. Critical Facilities Round Table In the first week of February 2010, the Critical Facilities Round Table (CFRT) supported the DOE Energy Star and The Green Grid at conferences in San Jose. Several CFRT members helped found The Green Grid and now sit on its Board of Directors. And, many other CFRT members have contributed operating data to DOE’s Server and Data Center Energy Star development programs. CFRT is a non-profit organization based in the Silicon Valley that is dedicated to the open sharing of information and solutions amongst our members made up of critical facilities owners and operators. Please visit our Web site at www.cfroundtable.org or contact us at 415-748-0515 for more information.
<urn:uuid:bd1de664-f4fd-424e-869d-826ca1d1542b>
CC-MAIN-2024-38
https://www.missioncriticalmagazine.com/articles/83185-cloud-computing-is-here-to-stay-until-something-better-comes-along
2024-09-08T04:04:37Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650958.30/warc/CC-MAIN-20240908020844-20240908050844-00361.warc.gz
en
0.931656
1,768
2.609375
3
There’s an unceasing buzz around big data and AI, the opportunities and threats of these technologies and concerns about their future. Meanwhile, companies are installing more and more sensors hoping to improve efficiency and cut costs. However, machine learning consultants from InData Labs say that without proper data management and analysis strategy, these technologies are just creating more noise and filling up more servers without actually being used to their potential. Is there a way to convert simple sensor recordings into actionable industrial insights? The simple answer is yes, and it lies in machine learning (ML). Machine Learning Capabilities The scope of ML is to mimic the way the human brain processes inputs to generate logical responses. If people rely on learning, training or experience, machines need an algorithm. Also, as each of us learns more, we adapt our reactions, become more skilled and start to apply our efforts selectively. Replicating this self-regulatory behavior in machines is the finish line of ML development. To learn, a computer is presented with raw data which it tries to make sense of. As it progresses, it gets more and more experienced, producing ever more sophisticated feedback. Under the broad umbrella of the Internet of Things (IoT), we can find anything ranging from your smartphone to a smart fridge to sensors monitoring industrial processes. Yet, there are at least four essential concerns related to IoT implementation, which need to be addressed: - Security and Privacy: Any algorithm that processes this kind of data needs to embed ways to keep all communication safe, especially if we’re talking about personal data such as that collected by medical sensors. - Accuracy of Operation: Sensors implemented in harsh conditions can send faulty data, or no data, disrupting the algorithm. - The 3 Vs of Big Data: Most IoT devices generate what can be classified as big data because it checks the 3Vs: volume, velocity, and variety. Tackling the 3Vs means finding the best algorithms for the type of data you’re using and the problem you’re trying to solve. - Interconnectivity: The value of IoT is in making disconnected items and tools “talk” to each other. However, since these are all created differently, they need to have a common language, which is usually the smallest common denominator. If computers already have protocols like TCP/IP, how would your fridge talk to your coffee machine? Why Use Machine Learning for IoT? There are at least two main reasons why machine learning is the appropriate solution for the IoT universe. The first has to do with the volume of data and the automation opportunities. The second is related to predictive analysis. Data Analysis Automation Let’s take car sensors as an example. When a car is moving, the sensors record thousands of data points which need to be processed in real time to prevent accidents and offer comfort to passengers. There’s no way for a human analyst to perform such a task for each car, so automation is the only solution. Through machine learning, the central computer of the vehicle can learn about dangerous situations, like speed and friction parameters, which can be hazardous to the driver, and engage safety systems on the spot. The Predictive Power of ML Coming back to the car example, the real power of IoT lies not only in detecting current dangers but identifying more general patters. For example, the system could learn about the driver who takes turns too tightly or has difficulties with parallel parking, and help him or her by providing additional guidance in these matters. The most useful feature of ML for IoT is that it can detect outliers and abnormal activity and trigger the necessary red flags. As it learns more and more about a phenomenon, it becomes more accurate and efficient. A great example is what Google did with its HVAC system, reducing energy consumption significantly. Last but not least, there’s also the opportunity to create models which predict future events very accurately by identifying the factors leading to a particular result. This offers a chance to play with the inputs and control results. How Should It Work? It’s vital to understand that, when an IoT system depends on human input, it can fail miserably. It needs the support of machine learning to become a perfectly aligned system resistant to human errors. In an interconnected world, human mistakes are quickly corrected by algorithms. This helps optimize the entire process through feedback mechanisms. The predictive component of the system can identify the correct input to get the expected output. When powered by ML, IoT can work flawlessly both at an individual level, so that you don’t mess up your morning routine, for example, and at a collective level. The latter case can be illustrated with interconnected cars that can communicate with each other and perform dynamic rerouting to avoid traffic jams. From Big Data to Smart Data The “work smarter, not harder” advice is a good fit for managing IoT-generated data and turning it into useful insights. While big data is all about overcoming the challenges posed by the 3 Vs, smart data can refer to: - Clean-up of sensor data on the spot before sending it to the cloud for analysis - Pre-processed batches of sensor information, ready to be turned into actionable insights The added value of machine learning in both cases is that it can take smart data and make ML models work faster and more accurately.
<urn:uuid:a01689e3-f445-4f46-9d34-619555d23e95>
CC-MAIN-2024-38
https://www.iotforall.com/why-iot-needs-machine-learning
2024-09-09T08:35:12Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651092.31/warc/CC-MAIN-20240909071529-20240909101529-00261.warc.gz
en
0.940599
1,109
2.71875
3
Today, I read an interesting story in Datanews, a Belgian IT newspapers. To briefly resume, “Company A”, the customer, complains about “Company B”, the telecom operator, which installed a telephone central at the first one premises. During a weekend, hackers took control of the system and used it perform calls to very-high-rate-countries. The result was a huge bill of more than 7000 EUR! The customer refused to pay the bill and complained about a nonexistent pro-active monitoring of the telecom operator network usage. On the other side, the telco had also a killing argument: one of the customer employee configured a very simple password on his voice mail (“1111”). Read the full article in French here (English translation). Despite the fact that this user password was not strong enough (but isa PIN code strong enough? Only 2^4 combinations of numbers), the customer submitted an interesting question: why didn’t the telecom company perform “behavioral monitoring“? Let’s review some types of monitoring… The first type, I call it “dumb monitoring, is just comparing a key indicator against a predefined value or an interval of values. A good example is the Nagios monitoring tool. Each key indicators are compared to two values: a warning and a critical thresholds. Such indicators may be the CPU usage of a server or the used percentage of a file system: $disk_usage=get_disk_usage("/var"); $warning=90; $critical=90; if ($disk_usage < $warning) print "Disk usage OK"; elseif ($disk_usage < $critical) print "Disk usage WARNING"; else print 'Disk usage CRITICAL"; The main issues with dumb monitoring are false alarms: some jobs executed during the night or weekend can fill the file system up to 98% during a few minutes then clean up temporary files. If the monitoring tools polls the file system usage during the job execution, a false alarm will be generated followed by a recovery. Another phenomena is when the CPU or disk usage "flaps", going always from a low to a high usage (example: a job executed every hour which comsumes 100% of the CPU). The second type of monitoring is called trending or forecast monitoring. Based on algorithms such as Holt-Winters, we try to estimate when the key indicator will be in trouble. Applied to our first example (a file system usage), we should be able to eliminate false positive alerts (when the file system is temporary almost full) but also perform forecasting. Like in the first type of monitoring, the file system usage polling is performed but instead of checking the returned value, we pass it to the algorithm which estimates when the file system will be full and draws nice graphs. Check out the example below: Note that trending is a very interesting tool for managers. Generated graphs are excellent evidences. It will be easier to defend a purchase order for extra disks in the company SAN or more bandwidth with the help of the graphs. Finally, the state-of-the-art: behavioral monitoring. The oldest gamers will remember the Lemmings game? It was based on the behavior of real lemmings which are known to perform mass actions in the same time, even if they are dangerous for the whole community. To setup the behavioral monitoring, we must first define an activity profile of our key indicator. As for the two first methods, polling is performed and received values will be compared to the defined profile but also to previous alerts! Back to our file system usage, a behavioral monitoring solution will be able to detect if a disk full event is "normal" or not. Example: every Sunday night, the disk is full due to a backup procedure but it must be cleaned up a few hours later. A disk still full on Monday morning will generate an alert! It is also possible to mix multiple sources of data to build the policy: A nice example of the usage correlation of a user access badge and his session on the network. If the user swipe his badge to enter building "A", he must open a session on a computer located in the same building; otherwise, his badge might been stolen or his account compromised. Detection of unknown protocols on your network or during non business hours is also part of the behavioral monitoring. Example: is it normal to detect VoIP traffic during the weekend? Yes if a valid user used his access card to enter the building and logged on the network! To resume, behavioral monitoring makes the difference between normal and abnormal activities. Behavioral monitoring is already used in a lot of business. Your bank monitors your credit card activity. If you just paid something in a restaurant in France and a few minutes later, the same card is used in Japan, something is wrong! (or you have some nice teleportation facilities ;-). The same applies with your mobile phone SIMM card. If your mobile is connected to your local operator and, in the same time, a connection request is forwarded from a foreign country operator (roaming), same issue, your SIMM card might be stolen and duplicated. Now, back to our story described above. I fully agree with the "customer". The telecom operator has all required data to setup behavioral monitoring in place! Maybe one exception in this case, it's dangerous to stop the service in case of suspicious activity (telephone can be used to call 911 in case of emergency!). Why not notify the customer BEFORE sending the bill?
<urn:uuid:dc7499e1-9c47-4566-aed6-d643613b9291>
CC-MAIN-2024-38
https://blog.rootshell.be/2009/04/22/tell-me-how-you-work-and-ill-monitor-you/
2024-09-10T15:11:06Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651255.81/warc/CC-MAIN-20240910125411-20240910155411-00161.warc.gz
en
0.93879
1,134
2.5625
3
Cybersecurity criminals have always kept their list of targets wide and diverse. Yet, they recently started focusing on schools and kindergartens as their primary target. Recent data shows that 57% of the latest ransomware incidents involved K-12 students. It shows that cybersecurity awareness and training are not well implemented in the schooling system. Remote learning is one of the reasons why it’s more important than ever to talk about cybersecurity. The Security Challenges of Remote Learning The current global pandemic has changed the way schools and educational facilities operate. Most classes are now taking place online, which exposes children to more cybersecurity risks daily. Remote learning seems like the perfect solution for students during a global pandemic. Yet, it creates more room for hackers to carry out malicious attacks and steal personal data. The risk has become so high that the U.S. government issued a statement to warn schools and kindergartens of the potential cybersecurity threats. The statement explains how criminals are now targeting distance-learning programs. This shows how hackers have expanded their reach from colleges and universities to lower grade educational facilities. This has exposed way more students to potential security risks. Their most common approach in this field is the ransomware attack method. It is also the most profitable option on their end. Teachers Can’t Carry Out Online Classes Most simply put, during a ransomware attack, hackers breach the security system of a device or a network of devices to encrypt their content. They can also restrict access by locking users out of the network. Once they take over the targeted network, they request a certain amount of money to be paid as ransom. Upon making the payment, the user then gains access to the network. However, this doesn’t always happen. Hackers rarely carry out their end of the deal. Besides ransomware attacks, schools and kindergartens are often the targets of DDoS attacks. They prevent teachers from carrying out online classes on remote platforms. During a DDoS attack, hackers use multiple devices or networks of devices to overwhelm servers and online platforms. Thus, they cause them to crash. During the crash, students can’t take part in remote learning, and their personal data and online accounts are put at risk as well. How to Respond to Potential Security Threats Most classes are taking place online. Hackers have shifted from corporations to kindergartens and schools that seem to be the easiest prey for these attacks. The Wall Street Journal reported a recent event at the Las Vegas’ Clark County School District where students’ private data was leaked after the school refused to pay the ransom. This goes to show that hackers are more than serious in their intent to leak personal data. This whole situation with the global pandemic is not close to an end. Students will have to stick to remote learning for at least the next six months or even further. With no end to remote learning, the cybersecurity risk will remain one of the number one problems for students and teachers around the world. Thus, it is crucial to learn how to respond to potential security threats such as ransomware and DDoS attacks. On top of securing networks and devices with proper antivirus tools, both teachers and students should use virtual private networks during remote learning practices. It is clear that children of this age are not aware of online security risks, let alone the security measures. That is why teachers and educators should encourage parents to take action. Also, don’t forget to install the right security tools on their children’s devices. The Unforeseeable Future With no clear end to the pandemic, we will continue the remote working/learning trend in 2021. Hopefully, educational facilities will take online security threats more seriously. It’s essential to have better security measures and raise awareness of VPN services and similar security tools. It’s important to be aware of these threats and protect ourselves. One of the best ways to ensure safety is to educate kids about various cyber threats. Only a mutual effort by students and teachers can prevent the consequences of common cybersecurity attacks.
<urn:uuid:52939b06-37ee-4c9e-b21b-12e609da59f9>
CC-MAIN-2024-38
https://www.cybersecurity-insiders.com/schools-and-kindergartens-a-new-target-for-hackers/
2024-09-12T21:48:19Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651498.46/warc/CC-MAIN-20240912210501-20240913000501-00861.warc.gz
en
0.959394
825
3.0625
3
Authenticating individual users for security purposes within organizations has never been more important to enterprises across all sectors. These days, it's not only banks and law enforcement offices that have to think about better means of authentication, but also educational institutions, small businesses and online retailers, among many others. "The heightened threat atmosphere of today is demanding greater security measures across the board." The heightened threat atmosphere of today is demanding greater security measures across the board, and that is leading companies to pursue new and evolving means of protection. Increasingly, biometric authentication solutions are becoming part of the answer to the growing threats facing enterprise infrastructures. One of the things that makes biometric authentication so effective and appealing to a broad user base is its simplicity: The technology is built around the idea of verifying identity using unique biological traits. Chiefly, this takes the form of fingerprint-based authentication, but biometric technology can also encompass things like retina scans, finger vein IDs and voice identification. The business move toward biometric-based security As many corporate breaches have illustrated - and continue to show - traditional passwords are no match for hackers who are adept at breaking into password databases and stealing privileged login credentials. For many enterprises out there, the inherent vulnerability of legacy authentication solutions is fueling a push toward biometric authentication, which is seen as a technology that is more uniquely suited than passwords to match today's cyber threat atmosphere. "The concept of using unique biological characteristics as an authoritative stamp dates back at least 31,000 years." The concept of using unique biological characteristics as an authoritative stamp dates back at least 31,000 years, to the first evidence of prehistoric men leaving handprints as signatures next to their cave paintings. It wasn't until 1858, however, that the world's first system of using hand images for identification was leveraged by a Civil Service leader who wanted a formal system of distinguishing people legitimately in his employ versus imposters. By 1903, prisons in New York were harnessing fingerprints to keep track of offenders. And by the 1960s, biometrics had branched out to encompass rudimentary versions of face and signature recognition. But it's only been in relatively recent years that biometric technology has undergone an evolution into a mainstream means of identification and authentication. This is perhaps most visibly highlighted by the CIA's use of facial recognition technology, combined with DNA, to verify the identity of the body of Osama bin Laden. These days, biometrics represents the future of authentication. As a sector, the biometrics system market is projected to reach a value of $24.44 billion by 2020, according to MarketsandMarkets. The significance of that statistic prompts the question: What will biometric authentication look like by 2020? Here are some ideas: - It will be the primary means of banking identity verification: Fewer enterprise sectors carry a greater responsibility when it comes to identity guarding than banks. Financial institutions guard some of the most privileged data out there, and therefore must be at the forefront of authentication technology. It's significant, then, that by 2020 biometrics will be the primary means of identity authentication leveraged by banking patrons, according to a Goode Intelligence report."There is a growing desire from the banking industry to adopt convenient methods to verify the identity of their customers and this is creating the conditions to drive the adoption of biometrics in banking even higher," Goode Intelligence founder Alan Goode said. "The adoption for banking purposes is a major contributor to this growth." With banks projected to flock to biometrics in big numbers, it's very likely that this move will set a precedent for identity security among other enterprise sectors, which helps to explain why the biometrics sector is anticipated to attain such a large value by 2020. - It will encompass developing means of identity vetting: When most people think of biometrics, fingerprint scanning is the first thing to come to mind. But by 2020, other means of biological characteristics-based verification will likely be the norm as well. According to FCW, traditional biometrics methods like fingerprint and iris scanning will be joined by evolving methods like measuring respiration and pulse as well as cognitive vetting, such as evaluating reactions to certain stimuli. These methods may prove even more effective than the current ones, but they also bring up important privacy questions that will need to be considered. As a leader in the field of biometric authentication, Entrust provides all business sectors with cutting-edge biometric technology that allows users to guard networks with the most robust tools out there. That technology comes in the form of Entrust IdentityGuard Biometrics, a solution that leverages fingerprint-based authentication to secure access to privileged networks. IdentityGuard Biometrics is designed to guard sensitive information with one of the most powerful authentication mechanisms on the market. It's the kind of solution that's built to handle the security needs of the present and the future.
<urn:uuid:3367fd63-4bb4-40d6-aefd-bfc07c1f7de5>
CC-MAIN-2024-38
https://www.entrust.com/blog/2015/08/what-will-biometric-authentication-look-like-in-2020
2024-09-15T09:47:39Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651622.79/warc/CC-MAIN-20240915084859-20240915114859-00661.warc.gz
en
0.955657
997
2.75
3
Let’s discuss Easy Methods to Know the Version of Installed Apps in Windows 11. You must be wondering why I should know the version number of Installed Apps. But it’s necessary to know the version and build numbers for some problem-solving and updates in Windows 11. Sometimes, you may encounter problems with an Application while you are working. We’ll have to troubleshoot to solve that problem. For that process, it is essential to know the version number of the working application. The version number refers to the feature version of the Windows update. A version number is a unique number or set of numberers representing the specific release on the OS. For example, Windows 11 versions are 22H2 and 23H2. Check the version number of the software you installed on your computer with the latest version number and release date to see if your Windows is updated. Normally, Microsoft releases its new feature updates once a year. What are the Advantages of Knowing the Version of Installed Apps in Windows 11? As we mentioned above, knowing the Version of installed Apps in Windows 11 is quite advantageous. Let’s explain how: 1. Software Compatibility: Knowing your Windows 11 version ensures that you can install and run compatible software. 2. Driver Updates: To ensure the best performance and stability, you should know the version that helps you to find and install the right drivers for your hardware components. 3. Security Patches: Knowing the Windows 11 version can help you identify any security issues and take the necessary steps to patch them. 4. Feature Access: Some specific Windows 11 have certain features and functionalities exclusively. Suppose you know your version number, that ensures you which features are accessible to you. 5. Troubleshooting: Knowing your Windows 11 helps troubleshoot issues and errors effectively. - Windows 11 Version Numbers Build Numbers Major Minor Build Rev - FIX Windows 11 Version 23H2 Download Error – 0x80246019 Easy Methods to Know the Version of Installed Apps in Windows 11 The following are the easy methods to know the current version number of installed applications or programs that are installed on your Windows 11. Different Methods to Know the Version Number of Installed Apps | 1. To know the Version of Installed Apps in Settings | 2. To know the Version of Installed Apps in the Microsoft Store | 3. To know the Version of Installed Apps in the Control Panel | 4. To know the Version of Installed Apps in Command | 1. To Know the Version of Installed Apps in Settings The easiest method is to know the Version number of Installed Apps via the Settings application. For this Click on the Start Menu and open Settings Application. You can also access settings by pressing Win+I or right-clicking on the Start menu and selecting Settings. In the Settings Application, you can see different settings. Select the Apps settings on the left side and select the Installed Apps in the right window pane. To see the app’s current version, click on the 3 dots on the right side of the right window pane. For example, to see the version of the Snipping Tool, tap on the 3 dots and click on Advanced Options. Here, under the Specifications, you can see the current version of the Snipping Tool. It also shows the Publisher name, App, Data, and Total Usage. - Easy Methods to Check Hardware Specifications on Windows 11 - 3 Best Ways to Disable Taskbar Thumbnail Preview on Windows 11 2. To know the Version of Installed Apps in the Microsoft Store To know the version of the app installed on Windows 11 from the Microsoft Store, you should first update your Microsoft Store’s latest version. Here’s how you can check it: - Click Start menu - Search Microsoft Store and Open it. Click Library on the left side to get a list of apps you have installed and downloaded on your PC. Tap on the App (for example, Snipping Tool) that you want to know the version details. Open the App (Snipping Tool) and you will see the app download page where you must scroll down the Additional Information section. Under Additional Information you have the Installed Version Number of that particular App (here snipping Tool). - Best Usage of Windows Memory Diagnostic Tool in Windows 11 - Enable Disable Settings Homepage in Windows 11 3. To know the Version of Installed Apps in the Control Panel Use the start menu to search Control Panel to launch it. In another way, you can Press the WIN+ R keys together to Run and type control on the Run Command Box. Hit OK to execute it. In the Control Panel there are different computer settings like Color Management, Keyboard, Mouse etc. Here we want to click on Programs and Features to know the version number of installed apps. There you can see the list of installed apps and programs on the left side and the version number in the right-side version column. Here we take an example of Microsoft Edge. Its version number is 123.0.2420.65 - 2023 December KB5033375 KB5033369 Windows 11 Patches and 1 Zero-Day Vulnerability - How to Use Windows Device Portal in Windows 11 4. To Know the Version of Installed Apps in Command With a single command, you can know the version number of installed apps on your computer. For that open Windows Terminal and select PowerShell. Then Copy and paste the command given below into Windows Terminal, and press Enter. Note:- (Apps installed for current user only) Note:– Get-ItemProperty HKCU:\Software\Microsoft\Windows\Currentversion\Uninstall\* | select-object DisplayName, DisplayVersion We are on WhatsApp. To get the latest step-by-step guides and news updates, Join our Channel. Click here –HTMD WhatsApp. Smitha Madathil is a computer enthusiast. She loves writing about Windows 11 and Intune-related technologies and sharing her knowledge, quick tips, and tricks about Windows 11 or 10 with the community.
<urn:uuid:77fe09f9-e638-4e3c-bbbb-a4e9fd24c399>
CC-MAIN-2024-38
https://www.anoopcnair.com/version-of-installed-apps-in-windows-11/
2024-09-16T16:15:41Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651697.45/warc/CC-MAIN-20240916144317-20240916174317-00561.warc.gz
en
0.854472
1,265
2.5625
3
Swap space is a part of secondary storage that is used as virtual memory. When the main memory (or primary storage) is full, inactive pages (data in the main memory that is not required by currently running processes) are moved to virtual memory, and then moved back when required. This swapping frees up the main memory for frequently processed data. Using secondary storage as virtual memory can increase performance for systems that have little main memory available. However, swap space does not replace or increase physical memory, because secondary storage is much slower to read from and write to. It should be used only for infrequent swapping of data that does not require a lot of processing. Swap usage refers to the percentage of virtual memory that is currently being used to temporarily store inactive pages from the main physical memory. It is crucial to monitor swap usage, because swap space is your “safety net” for when you run out of RAM. While swapping, processes may get slower, data is still nevertheless processed. However, when you run out of virtual memory, processes are queued and stalled until some memory is freed up. The amount of data swapped is not as important as the rate at which swapping occurs. An efficient system may swap-out a lot of inactive pages at a slow rate, leaving RAM space for frequently processed data. A more important metric is the swap-out rate, which refers to the rate (in kilobytes per second, KB/s) at which memory is swapped from main memory to virtual memory. Swap-out rate is measured over a sample period of time. The higher this metric is, the more frequently the disk is accessed for swapping. Since the disk is secondary storage, which is slow, higher rates mean slower data processing. Even when swap space is not particularly full, continuously high swap-out rates indicate that the pages that are swapped are actively processed. This means that your system requires more fast physical memory for your workloads, and swap is not helping. Monitoring swap usage and swap rate can help analyze the workloads on your server. For example, if you determine that most of the processes require fast processing, you need to have enough main memory, while virtual memory will not help much. On the other hand, if a lot of processing is fairly slow with large chunks of data stored in memory between calculations, you can use swapping to free up the main memory for faster client request processing. Balancing the workloads both across servers and on each separate server is the key to providing efficient services. Understanding the benefits of swapping can help with this balancing. Using the Anturis Console, you can set up monitoring of swap usage and swap rate for any hardware component (a server computer) in your infrastructure by adding the SwapUsage monitor to the component.
<urn:uuid:62cedc57-4788-4d67-b24b-88c6a3c49cf0>
CC-MAIN-2024-38
https://anturis.com/swap-usage-monitor/
2024-09-09T10:49:33Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651098.19/warc/CC-MAIN-20240909103148-20240909133148-00361.warc.gz
en
0.941091
564
3.359375
3
What Is VoIP and Why Is It Good for Your Business You may have heard about VoIP, but you are not sure what it is. Understanding how it works will enable you to use this technology to your company’s advantage. What is VoIP? VoIP, which stands for voice over internet protocol, is a technology that allows you to make and receive telephone calls using the internet instead of a traditional phone line. VoIP converts analog – the traditional – voice calls into packets of data. These packets of data travel over the internet just like any other type of data, such as email. Using a VoIP service, you can call a landline or a cell phone. When calling a landline or a cell phone, the packets of data are converted back to telephone signal – the traditional voice signal – before they reach the person you are calling. You can also make or receive calls using landline telephones via your VoIP service. For this purpose, you need an analog telephone adapter to be connected to your network. In addition, you can call someone via computer-to-computer, with you – the caller – and the receiver speaking into computer microphones and listening through computer headsets or speakers. A basic VoIP system only requires a broadband internet connection and a VoIP-enabled phone; a computer with VoIP software and a headset; or a traditional phone connected to an adapter. VoIP versus Unified Communications Aside from VoIP, you may have heard about “unified communications”. VoIP refers to the basic internet-based telephony system. Unified communications, meanwhile, is a communication system that includes not just VoIP, but other communication services, including conferencing that combines video, data and desktop sharing. You can also instantly monitor the availability of your colleagues through this unified system. Benefits of Using VoIP Your company’s aging telephone system only causes productivity slowdowns, as well as loss of revenue due to poor quality and expensive maintenance. VoIP, on the other hand, gives your business the following benefits: - It allows for individual employee telephone numbers without the need for multiple physical landlines. - It reduces local and long-distance charges. - It reduces travel costs as on certain occasions your staff need not have to travel – thanks to online conferencing, a convenient way to use video calls and other collaboration tools. - It can easily make changes – adds or moves phone extensions and locations – saving your company money and giving your company more flexibility. - With the unified communications solution, your employees have more ways to collaborate – through voice calls, video chat, web conference and instant messaging. - Your customers can contact your staff more easily. Hosted VoIP versus On-site Installed VoIP Once you have made the decision to replace your aging telephone hardware with VoIP, you have to choose which solution is best for your small business: a hosted VoIP or an on-site installed VoIP. Both hosted VoIP and premise-based VoIP have their distinct advantages and disadvantages. Some small businesses favor the greater customization and control of premise-based VoIP, while other small businesses favor the scalability and ease of hosted VoIP. Your decision on what VoIP solution to choose will depend on how your company views VoIP – whether as an operating expenditure (OpEx) or capital expenditure (CapEx). Your organization’s growth plans, as well as the availability of in-house experts to manage the VoIP will also be factored in when deciding to choose between hosted VoIP and premise-based VoIP. The following are some of the major differences between hosted VoIP and premise-based VoIP: 1. Installation and Management One of the main differences between the two is that hosted VoIP can be accessed over the internet as a hosted service, while premise-based VoIP is installed on your local network. A premise-based VoIP runs on your I.T. infrastructure and connects to the public switched telephone network (PSTN). If you choose this path, your IT staff or IT partner will be responsible for the installation of the VoIP system and the upgrading of the routers needed as voice gateway to support the system. With a hosted VoIP, there is no need to upgrade the router as the voice getaway is part and parcel of the network of the hosted VoIP. Both hosted VoIP and premise-based VoIP require a fast internet connection to transmit voice traffic. A premise-based VoIP needs on-site resources and experts to manage the system, a hosted solution, on the other hand, does not need them. This management issue is one of the main reasons why small businesses pick hosted VoIP instead of premise-based VoIP. Some larger small businesses may opt for premise-based VoIP if they have in-house experts as they can have more control over the system. For instance, with premise-based VoIP, your company can upgrade the system anytime, instead of relying on the hosted VoIP provider. A premise-based solution also enables your company to exercise control over which features should be enabled for an enhanced VoIP system. Hosted solution, on the hand, offers VoIP features as bundles or packages, as such your organization cannot select the features that you want. If for instance, your company hires a large pool of temporary workers during holidays and you need more phone lines, it makes sense to choose the hosted solution. It is simpler to add more phone lines with hosted solution compared to premise-based solution. With hosted solution, adding more users can be enabled with just a few clicks of the mouse. On the other hand, adding users to the premise-based solution will involve installing network and phone system equipment – an ordeal that makes it difficult to scale. 3. Cost and Pricing Models Companies often choose their VoIP solution based on the number of users that must be supported. The initial costs and regular costs differ between hosted and premise-based solutions. The premise-based solution needs money to buy the necessary hardware, software, as well as the impending installation fees and assistance from the VoIP vendor. Hosted VoIP, meanwhile, charge users a monthly fee for the service. This fee is dependent upon the number of phone lines your organization subscribes to. The bundles or packages you choose are also factored into the fee, for instance, if your organization opts to avail of advanced features such as the unified communications or video conferencing. At GenX, we ensure that your organization gets the best VoIP system by using the following 5-step approach: - Planning and Assessment - Cost Analysis - Long Distance Requirements - Audio and Videoconferencing - Implementation and Training Contact us today to get started – (416) 920-3000.
<urn:uuid:488eecdd-d259-4d26-affc-4959a6716732>
CC-MAIN-2024-38
https://www.genx.ca/voip-good-business
2024-09-12T00:07:47Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651405.61/warc/CC-MAIN-20240911215612-20240912005612-00161.warc.gz
en
0.942352
1,405
2.921875
3
Both employees and leaders in the IT field are familiar with the IT skills gap. For digital enterprises today, IT not only supports the business – it often is the business. The IT skills gap illustrates that companies sometimes have a difficult time finding, hiring, and retaining employees with the IT skills they need, to ensure quality products. Industry experts are split. Some argue that the IT skills gap doesn’t exist (in the United States, at least). More experts, however, argue that the gap exists, though to varying degrees. Either way, the biggest concern is how companies address these current issues in the short-term to ensure that they don’t grow into larger issues down the road. In this article, we’re examining the IT skills gap – how it exists today and what we can do to improve the situation. Defining Skills Gaps and IT Skills Forbes defines a skills gap as the difference between what companies need or want their employees to do and what employees are actually able to do. A skills gap can apply to any skills area: hard skills like sales, business, finance, and IT, as well as soft skills like interpersonal communication and time management. When we talk about the IT skills gap, needed skills vary widely based on different companies’ needs as well as the needs of specific jobs within one company. The State of Skills Gap in 2019 A 2017 Forbes report on IT service management says that C-level executive report a lack of IT skills as the biggest issue in aligning IT with business services. Many CIOs report that the IT skills gap applies to employees and candidates within the higher echelon of their IT teams. Leaders are able to staff and retain help desk and entry- and mid-level programmers, but as technology silos narrow down evermore into special areas, companies say it’s harder to find enough people with the precise qualifications needed. In 2019, the IT skills gap continues to expand with rising demand and inadequate supply. According to a World Economic Forum report, 133 million new roles may emerge globally by the year 2022. The primary growth drivers include the proliferation of Artificial Intelligence and Machine Learning technologies. Despite the advancement of technology, organizations will continue to rely on the human workforce for creative problem solving. Intelligent computing systems are expected to complement the human workforce and not entirely replace them, which in turn requires organizations to re-skill their workforce to take advantage of the technology advancements. According to another research survey conducted across over 1000 business executives in the US, organizations are struggling recruit the right candidate due to the skills gap and the education system is not doing enough to address the cause. The research finds that 75% of organizations believe that the skills shortage among applicants is the primary cause for the difficulty in hiring. 51% believe that the education system is doing little or nothing to solve the problem. Cause of IT Skills Gap Industry insiders cite many reasons for this IT skills gap, including: - The demand is increasing. As companies need to both maintain and optimize ever increasing technology, they need more people to support this mission. Many organizations are pursuing digital transformation initiatives, only to realize the projects are inherently dependent upon onboarding individuals with the right skillset. As a result, a sudden increase of IT skills has emerged in recent years. - The supply is decreasing. Jobs in IT grew by 13 percent between 2002 and 2013, yet those who graduated with IT degrees shrunk by 11 percent over the same period. IT skills requirements also change rapidly through the years. In the period 2018 to 2022, an average of 42% of shift in IT skills is expected. Hiring managers are therefore increasingly focusing on recent graduates equipped with the latest digital skills most relevant to their digital transformation initiatives. - The need for IT is spreading. Early on, only traditional businesses were implementing technology. Today, however, industries beyond business require vast amount of IT, especially education and healthcare. Cloud computing and artificial intelligence solutions are the primary drivers for this trend. Not every organization needs to develop and deploy a technology solution in-house. Instead, they can leverage SaaS services readily available as an affordable OpEx. However, IT professionals with the right skills to take advantage of these latest digital solutions are still needed across all organizations. - The way we learn IT skills is changing. More people are re-tooling outside of traditional degree-granting colleges and universities, opting instead for short-term, intensive developer boot camps that provide the hard IT skills, but leaving this population short of degrees. This lack of a formal degree may mean that without a specific degree, some companies are missing out on this talent pool. Further, the boot camps focus primarily on programming skills – and considerably less time on business know-how, which could affect a candidate’s soft skills for a position. - The specialization within IT is double-edged. With IT booming, candidates can specialize in just one or two areas. But, companies that require too narrow or specific a specialization will often come at the expense of a candidate with broader technology and business knowledge. Additionally, the technology landscape is evolving continuously. Individuals specialized in a select few skills will be required to learn additional skills to leverage the advancing technologies. - The reliance on IT as an emergency responder continues. Despite companies continuing to transition towards a proactive IT and business approach, IT still acts as emergency responders too often. Many organizations continue to spend a majority of their budget on maintenance instead of on project development, innovation, and improving the organization’s competitiveness. The Severity of the IT Skills Gap While there is a significant amount of agreement around the existence of the IT skills gap, opponents point to the current unemployment rate in the U.S., under 5% in Q4 of 2017, as some proof that the IT skills gap either can’t exist or can’t be as serious as some argue. Indeed, whether the skills gap is industry-threatening or merely one of many issues facing businesses today depends significantly on who you are. A recent TEK Systems study surveyed more than 1,300 leaders and professionals in IT in the U.S. Interestingly, leaders and those working in the field share some beliefs on the IT skills gap. Variations among them, however, may illustrate lack of understanding, lack of communication, or lack of standardization. Across the board, more than 75% of those surveyed agreed that the perceived IT skills gap is real. They further agree that only one-third of companies have the talent in-house to meet their IT needs. But, compared to IT professionals, leadership believes that this gap has a significant negative impact. This could be due to their macro view, which often underscores negatives more than positives. Perhaps some leaders have diverging understandings of what IT does and what IT needs. This bears out further in comparing responses to why candidates and open IT positions don’t match. More than half of surveyed IT leaders reported unqualified candidate pools (similarly, candidate pools that don’t align to business needs), small pools of candidates, and budgetary constraints – preventing them from hiring more ideal candidates. At the same time, surveyed IT professionals reported that many job requirements are unrealistic, and that company’s expectations do not align with compensation budgets. Of the IT professionals surveyed, 41% further reported that the experience and/or expertise that companies demand is too niche. Interestingly, while some blame a lack of extensive education, only a quarter of IT professionals currently at work in the field agree. The severity of the IT skills gap is also illustrated in the impact on IT teams. Seventy-one percent of IT leaders surveyed reported decreased efficiency, and 63% of IT workers agree. The reasons for this inefficiency? Leaders see that it negatively affects project completion and time-to-market 69% of the time, while professionals say this is an issue only half the time. Remedying the IT skills gap Without a strategic plan for the IT workforce, companies will continue having difficulties finding quality IT candidates. When positions go unfilled, there is the immediate impact of inefficiency, but there’s collateral damage, too. Decreased employee morale leads towards attrition. This means companies risk losing existing talent. So, what measures are companies taking to fix this gap? Surveyed companies rarely have a long-term strategic workforce plan, but more than half of them reported short-term fixes including training and developing the current workforce, outsourcing current projects, and increasing the use of contract/temporary workers. There was a major drop-off when it came to employee benefits, with only one-third of companies investing in education, 26% offering flexible hours and telecommuting options, and a mere 9% offering more lucrative compensation packages. Longer-term, companies need to adapt new practices to both attract new and retain existing talent. Indeed, communication is a major issue, and the perceived differences between IT leaders and works indicates this. There is little clarity around hiring: why one candidate was chosen over another, or what a candidate was lacking that he could then improve upon. Until this feedback loop is closed, the skills gap cannot be bridged. There’s also a lack of standardization. Surely, a majority of jobs in the IT sector fall within a certain range: help desk, entry- and mid-level programmers, high-end programmers, security specialists, management with programming experience, etc. With consensus general job titles, candidates can better prepare for the industry. Further, when jobs are standardized to a realistic extent, job descriptions themselves can sound more realistic, instead of the big-ticket wish list many read as now. This will help candidates find “perfectly good” candidates for a position instead of the one “perfect” candidate. For most companies, this talent gap is self-inflicted. Addressing the skills gap will take time, but by planning strategically and increasing communication, clarity, and standardization in job expectations, candidates can better prepare. Responding with Transforming Workforce Development Another facet of the IT skills supply-demand gap concerns the response of business organizations – how they utilize their existing HR resources, attract the right workforce to address the right functional requirements and match the pace of technology change. Salesforce recently conducted a survey of 750 hiring managers and found that the Fourth Industrial Revolution necessitates transformative workforce development strategies in response to the changing landscape of the IT talent market. The report suggests that new hires alone will not suffice in filling the talent void. The competition for talent will become even more intense, by 55%, while 58% of the respondents believe that transformative new workforce development methodologies will be required to equip the existing workforce with the new skills. According to another research report by the World Economic Forum, emergence of new technologies is driving growing skills instability. The skills relevant today may not be entirely sufficient tomorrow. Around 54% of all employees will require significant reskilling by the year 2022. The transformation in IT landscape has been rapid and the skills gap has emerged as a consequence of decisions made by business organizations, education system and IT professionals. According to the survey, hiring managers believe that an overhaul of training at the academic level is required along with formalized education programs at the workplace. The proportion of IT professionals opting for hard skills in data science and business intelligence is small and inadequate to fulfil the growing demand in the industry. Bootcamps and training programs at the workplace, as well as online courses in IT skills for the fourth industrial revolution teach the necessary hard skills but the necessary soft skills lack still. Around 70% of hiring managers suggested that soft skills such as collaboration and team work, judgement and decision-making, creative/abstract thinking and emotional intelligence will become increasingly essential. Once hired, both the employers and employees are expected to contribute toward the overall workforce skill development initiatives. Employers need to retain skilled workforce over the long term. Employees need to broaden their skillset with both technical and human skills such as creativity, critical thinking and negotiation to keep up with advancing technologies.
<urn:uuid:6c6004a9-8b43-4bc5-ae11-4485cafc6eb0>
CC-MAIN-2024-38
https://www.bmc.com/blogs/the-it-skill-gap-explained/
2024-09-13T04:50:30Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651507.67/warc/CC-MAIN-20240913034233-20240913064233-00061.warc.gz
en
0.953635
2,466
2.65625
3
A petaflop is the ability of a computer to do one quadrillion floating point operations per second (FLOPS). Additionally, a petaflop can be measured as one thousand teraflops. A petaflop computer requires a massive number of computers working in parallel on the same problem. Applications might include real-time nuclear magnetic resonance imaging during surgery or even astrophysical simulation. Nvidia and the National Energy Research Scientific Computing Center (NERSC) have flipped the “on” switch for Perlmutter, billed as the world’s fastest supercomputer for AI workloads. Named for astrophysicist Saul Perlmutter, the new supercomputer boasts 6,144 NVIDIA A100 Tensor Core GPUs and will be tasked with stitching together the largest ever 3D map of the visible universe, among other projects, Venture Beat reported. Perlmutter is “the fastest system on the planet” at processing workloads with the 16-bit and 32-bit mixed-precision math used in artificial intelligence (AI) applications, said Nvidia global HPC/AI product marketing lead Dion Harris during a press briefing earlier this week. Adv. This site uses Akismet to reduce spam. Learn how your comment data is processed.
<urn:uuid:d1e3747a-0329-4eaa-a435-b8a12be6b8d6>
CC-MAIN-2024-38
https://www.abijita.com/what-is-petaflop/
2024-09-14T07:08:44Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651559.58/warc/CC-MAIN-20240914061427-20240914091427-00861.warc.gz
en
0.876784
265
2.84375
3
North Korea has continued to advance its nuclear weapons program, demonstrating significant progress and resilience despite international sanctions and diplomatic efforts aimed at curbing its development. Since 2006, North Korea has announced and carried out six nuclear weapon tests. These tests occurred in 2006, 2009, 2013, 2016, and 2017, with yields ranging from less than 1 kiloton in 2006 to as much as 250 kilotons in 2017. In addition, North Korea declared that it was closing and dismantling the Punggye-ri test site, a significant move that drew international attention. The nuclear explosive devices designed and tested by North Korea have been the subject of extensive analysis and scrutiny. The devices’ design, yield, and the materials used are crucial to understanding the capabilities and intentions of North Korea’s nuclear program. The modeling and measurement of North Korea’s nuclear warhead design have provided valuable insights into the country’s nuclear capabilities. In a presentation at Princeton University’s Program on Science and Global Security’s Brown Bag Seminar on October 23, 2019, Senior Research Associate and Nonproliferation Review Editor Joshua H. Pollack discussed three distinct streams of evidence regarding North Korea’s nuclear-weapon program: seismic and other observational data, insider accounts, and official North Korean statements. By comparing these streams, a largely consistent picture emerges, suggesting that North Korea likely has four designs: three implosion devices and one two-stage device. The Program on Science and Global Security (SGS), based in the Woodrow Wilson School of Public and International Affairs, conducts scientific, technical, and policy research to advance national and international policies for a safer and more peaceful world. SGS has a long history of working on nuclear arms control, nonproliferation, and disarmament to reduce the dangers from nuclear weapons and nuclear power. In recent years, SGS has also focused on biosecurity issues and the challenges posed by disruptive technologies such as new biotechnologies, information and communications technologies, autonomous weapons, artificial intelligence, and space-based systems. Joshua Pollack, who joined the Middlebury Institute of International Studies (MIIS) in April 2016, has an extensive background in issues related to weapons of mass destruction, including proliferation, arms control, and deterrence. Before joining MIIS, Pollack served as a consultant to the US government and worked as a defense policy analyst at DFI International, Science Applications International Corporation, and Constellation West. His clients included the Defense Threat Reduction Agency, the Office of the Under Secretary of Defense for Policy, the Department of Homeland Security, the National Nuclear Security Administration, and the Plans and Policy Directorate (J5) of U.S. Strategic Command. In 2015, he was named an Associate Fellow of the Royal United Services Institute and has been a frequent commentator in major media outlets. The six nuclear tests conducted by North Korea provide a timeline of the country’s nuclear development. The first test in 2006 had an estimated yield of less than 1 kiloton, indicating a relatively rudimentary device. The second test in 2009 showed an increased yield of approximately 2 to 6 kilotons, suggesting improvements in design and efficiency. The 2013 test further increased the yield to about 6 to 16 kilotons, demonstrating continued progress. The tests in 2016 and 2017 were particularly significant, with yields estimated at 20 to 30 kilotons and up to 250 kilotons, respectively. The 2017 test, in particular, marked a significant milestone, as it was believed to be a hydrogen bomb, a much more powerful two-stage thermonuclear device. The Punggye-ri test site, located in the mountainous region of North Hamgyong Province, was the site for all six of North Korea’s nuclear tests. The site’s closure and dismantling in 2018 were viewed as a positive step towards denuclearization. However, the extent to which the site was rendered unusable remains unclear, and some analysts suggest that North Korea could still resume testing at other locations if it chooses to do so. The design and development of North Korea’s nuclear warheads have been informed by various sources of evidence. Seismic data from the nuclear tests provide information about the yield and characteristics of the explosions. Insider accounts, though difficult to verify, offer insights into the internal workings of North Korea’s nuclear program. Official statements and media releases from North Korea also provide clues about the country’s nuclear capabilities and intentions. The three implosion devices reportedly developed by North Korea include a single-stage fission device with a yield of around 10 to 20 kilotons, similar to the bombs dropped on Hiroshima and Nagasaki. The two-stage thermonuclear device, or hydrogen bomb, represents a significant advancement, with a yield potentially exceeding 250 kilotons. This design involves a primary fission explosion that triggers a secondary fusion reaction, resulting in a much more powerful explosion. North Korea’s choice of fissile material and its boosting capabilities are still unknown. Fissile material, such as highly enriched uranium or plutonium, is essential for sustaining a nuclear chain reaction. Boosting involves injecting a small amount of fusion fuel, such as deuterium and tritium, into the core of a fission bomb to increase its yield. If North Korea has developed boosted fission weapons, it could significantly enhance the destructive power of its arsenal. The international community continues to monitor North Korea’s nuclear program closely. Diplomatic efforts, such as the summits between North Korean leader Kim Jong-un and US President Donald Trump in 2018 and 2019, have sought to address the issue through dialogue and negotiation. However, these efforts have yielded limited results, and North Korea’s commitment to denuclearization remains uncertain. As of July 15, 2024, the Federation of American Scientists (FAS) released the latest edition of the Nuclear Notebook, published in the Bulletin of the Atomic Scientists. The authors, Hans Kristensen, Matt Korda, Eliana Johns, and Mackenzie Knight, estimate that North Korea may have produced enough fissile material to build up to 90 nuclear warheads. However, the country has likely assembled fewer than that, potentially around 50. This estimation follows the trendline researchers have been tracking over the years, with no significant change from previous estimates in 2021 and 2022. North Korea’s nuclear policy has undergone notable shifts, particularly with the abandonment of its no-first-use policy. This change aligns with the country’s recent efforts to develop tactical nuclear weapons, indicating a more aggressive nuclear posture. The development of tactical nuclear weapons suggests that North Korea is looking to enhance its flexibility and options for nuclear use, which could lower the threshold for nuclear conflict. Warhead Preparation and Delivery North Korea has developed a highly diverse missile force across all major range categories. This includes short-range tactical missiles, sea-based missiles, and new launch platforms such as silo-based and underwater platforms. The advancements in solid-fuel missile technology are particularly significant as they improve the survivability and mobility of North Korea’s missile force. Solid-fuel missiles can be launched more quickly than liquid-fuel missiles, which require longer preparation times and are more vulnerable to pre-emptive strikes. Since 2006, North Korea has detonated six nuclear devices, updated its nuclear doctrine to reflect the irreversible role of nuclear weapons for its national security, and continued to introduce a variety of new missiles test-flown from new launch platforms. Hans Kristensen, director of FAS’s Nuclear Information Project, notes that the size and composition of North Korea’s nuclear stockpile depend on warhead design and the number and types of launchers that can deliver them. Disco Balls, Peanuts, and Olives North Korea’s nuclear warhead designs have been informally referred to as the disco ball, peanut, and olive, based on their appearance in North Korean state media. These nicknames describe the visual characteristics of the supposed warhead designs. The “disco ball” is a single-stage implosion device, the “peanut” is a two-stage thermonuclear warhead, and the “olive” is a new miniaturized warhead called the Hwasan-31. The Hwasan-31, first showcased in 2023, demonstrates North Korea’s progress towards developing and fielding short-range, or tactical, nuclear weapons. While North Korea’s warhead design and stockpile makeup are not verifiable, it is possible that most weapons are single-stage fission weapons with yields between 10 and 20 kilotons of TNT equivalent, similar to those demonstrated in the 2013 and 2016 tests. A smaller number of warheads could be composite-core single-stage warheads with a higher yield. The development of the Hwasan-31 and other tactical nuclear weapons indicates North Korea’s intent to strengthen its regional deterrence posture. Tactical nuclear weapons provide options for nuclear use below the strategic level, which can complicate the security calculus for neighboring countries and the United States. The pursuit of these weapons suggests that North Korea is preparing for potential conflict scenarios where limited nuclear use could be considered. North Korea’s missile force includes a range of delivery systems capable of carrying nuclear warheads. This includes short-range ballistic missiles (SRBMs), medium-range ballistic missiles (MRBMs), intermediate-range ballistic missiles (IRBMs), and intercontinental ballistic missiles (ICBMs). The diversity of North Korea’s missile force enhances its ability to threaten targets at various distances, from regional adversaries like South Korea and Japan to the continental United States. The development of sea-based missile platforms, such as submarine-launched ballistic missiles (SLBMs), adds another layer of complexity to North Korea’s nuclear strategy. SLBMs can be launched from submarines, making them more difficult to detect and pre-empt. This capability enhances the survivability of North Korea’s nuclear forces and strengthens its second-strike capability. North Korea has also invested in mobile missile launchers, which can be moved to different locations to evade detection and increase the survivability of its missile forces. Mobile launchers can launch missiles from various locations, making it more challenging for adversaries to target and destroy them before they are launched. The advancements in solid-fuel missile technology are a critical component of North Korea’s missile development program. Solid-fuel missiles have several advantages over liquid-fuel missiles, including shorter launch preparation times, greater mobility, and enhanced survivability. These characteristics make solid-fuel missiles a more credible and responsive component of North Korea’s missile force. North Korea’s nuclear and missile programs have been the subject of numerous international sanctions and diplomatic efforts aimed at curbing their development. Despite these efforts, North Korea has continued to advance its capabilities, demonstrating a high level of technical expertise and determination. The international community remains deeply concerned about North Korea’s nuclear and missile programs. The potential for miscalculation or escalation in the region poses significant risks to global security. Diplomatic efforts to address North Korea’s nuclear ambitions have been met with varying degrees of success, but the path to denuclearization remains fraught with challenges. In conclusion, North Korea’s ongoing modernization and expansion of its nuclear weapons arsenal highlight the persistent and evolving threat posed by the country’s nuclear and missile programs. The development of new warhead designs, advancements in missile technology, and shifts in nuclear doctrine all underscore the complexity and urgency of addressing this issue. The international community must continue to work towards a peaceful resolution while remaining vigilant against the potential threats posed by North Korea’s growing nuclear capabilities.
<urn:uuid:dc2a3186-5fdd-414c-be8e-77d264776c09>
CC-MAIN-2024-38
https://debuglies.com/2024/07/26/north-korea-continues-to-modernize-and-grow-its-nuclear-weapons-arsenal/
2024-09-13T07:58:09Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651510.65/warc/CC-MAIN-20240913070112-20240913100112-00161.warc.gz
en
0.939331
2,402
2.90625
3
Worldwide, policymakers, in response to measurable threats and harms to children online and due to a desire to make the Internet safer for children, are updating existing and issuing new age verification and assurance regulations. Age verification is the practice of verifying an individual’s age. Age assurance is a process of verifying or estimating, with a level of certainty and accuracy, an individual’s age. The goal of this paper is to document and share the prevailing consensus on what is happening with AGE assurance regulations — how regulations and the field are most likely to evolve and what actions organizations will likely be required to take to comply with the new regulations and technical demands. Age assurance is the collective term used globally to refer to the processes and technologies used to verify an individual’s age, estimate an individual’s age, and determine with a level of certainty that an individual’s claimed or estimated age is accurate (i.e., truthful, substantiated, and verifiable). We discuss why age verification and assurance matters and what you and your organization should consider to help make the internet a safer place for children and to protect your organization. Contents in brief Online Child Safety: Overview and Harms On The Rise: Regulations and Government Initiatives Objections to Age Verification State of the Art: Age Verification and Assurance What The Future Holds: Predictions and Recommendations Key Terms Underpinning of the Future of Personal Data
<urn:uuid:e49ea134-0e41-4152-9045-8b40ca1ccddd>
CC-MAIN-2024-38
https://mobileecosystemforum.com/whitepaper-is-it-rude-to-ask-your-age/
2024-09-15T16:33:09Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651632.84/warc/CC-MAIN-20240915152239-20240915182239-00861.warc.gz
en
0.920072
299
3.1875
3
Quantum resistant encryption & why it’s critical to future cybersecurity (TheSSLStore) Quantum resistant cryptography will be a key part of cybersecurity in the future. Managing Editor Casey Crane has written an extensive discussion of quantum resistant encryption & why it’s critical to future cybersecurity and what to do to prepare. IQT-News summarizes below. Quantum resistant encryption refers to a set of algorithms that are anticipated to remain secure once quantum computing moves out of the lab and into the real world. (They will replace the public key cryptography algorithms currently used by billions of people around the world every day.) By the way, when people use any of the following terms, they’re typically talking about the same thing (in most cases): Quantum resistant encryption Quantum resistant cryptography (QRC) Quantum safe cryptography Post-quantum cryptography (PQC) Post quantum encryption All of the public key encryption algorithms we currently rely on today are expected to be broken once researchers succeed in building large enough quantum computer. Once that happens, quantum resistant encryption will need to be used everywhere (both by “normal” [i.e., “classical”] and quantum computers) so that attackers with quantum computers can’t break the encryption to steal data. Current public key cryptographic algorithms rely on complex mathematics. According to the National Security Agency (NSA), quantum resistant cryptography should be “resistant to cryptanalytic attacks from both classical and quantum computers.” With this in mind, these algorithms would be something that can be used both before and after quantum computers are put to use in real-world applications. They’re designed with quantum computing threats in mind, but they’re not limited to being used only after a cryptographically relevant quantum computer (CRQC) is created. NIST says that quantum resistant algorithms typically fall in one of three main camps: Code-based cryptography — These are algorithms that rely on “error-correcting codes.” Lattice-based cryptography — These algorithms involve matrices based on geometric structures. Multivariate public key cryptosystems — These types of algorithms vary based on the type of problems they’re trying to solve. There is a fourth category that some reference — stateful hashed-based signatures. But according to NIST’s PQC FAQs page: “It is expected that NIST will only approve a stateful hash-based signature standard for use in a limited range of signature applications, such as code signing, where most implementations will be able to securely deal with the requirement to keep state.” The National Institute of Standards and Technology (NIST) has been engaged in a large-scale cryptographic competition of sorts for the past several years. The competition is an opportunity for mathematicians, researchers, cryptographers, educators and scientists to submit algorithms for consideration as future federal standards. The standards body announced their selection of seven candidates and eight alternate algorithm candidates from the third round of submissions. However, no final decisions have been made regarding which algorithm(s) will be standardized: 4 public key encryption and key-enablement algorithms (Classic McEliece, CRYSTALS-KIBER, NTRU, SABER) 3 digital signature algorithms (CRYSTALS-DILITHIUM, FALCON, Rainbow) 5 alternate public key encryption and key enablement algorithms 3 alternate digital signature algorithms The truth is that the threats that quantum computing represents aren’t new concepts, nor do they represent threats to your business and customers right now. The concept of quantum computing — and all of its benefits and dangers — has been around for decades and isn’t expected to come to fruition yet. how long is all of this expected to take? The answer depends on who you ask and in what context: The National Institute of Standards and Technology (NIST) says it can take 10-20 years “from deciding a cryptosystem is good until we actually get it out there as a disseminated standard in products on the market.” The NSA says that “new cryptography can take 20 years or more to be fully developed to all National Security Systems.” On Jan. 19, 2022, the White House released a memorandum specifying that agencies have 180 days to “identify any instances of encryption not in compliance with NSA-approved Quantum-Resistant Algorithms or CNSA […]” and must report the following to the National Manager: –What systems are noncompliant (including those with exceptions or waivers) –A timeline for how these systems will transition to compliant encryption, and –Any reasons why any systems should be exempt from compliance What does all of this mean at the level of your organization or company? In reality, not much right now for everyday businesses. But let’s be realistic here — it’s virtually impossible to be compliant with rules that haven’t yet been implemented. You can take steps to stay ahead of the curve as much as possible by taking the time to research and plan your strategy now. Part of this planning should include: Prioritizing which systems to transition first, starting with the most sensitive and at-risk resources, as well as those that are integral in terms of your organization’s goals and needs Designating who is responsible for different aspects of the implementation The widespread use of quantum computing — and, therefore, the deployment of quantum resistant cryptography — is still on the horizon but is likely at least a good decade or so away. But that’s why now is the time to prepare for PQC to help your business stay ahead of the curve. You don’t want to be one of the organizations caught unprepared when quantum computers make their mainstream debut. Sandra K. Helsel, Ph.D. has been researching and reporting on frontier technologies since 1990. She has her Ph.D. from the University of Arizona.
<urn:uuid:f0333cca-9ef3-4777-addb-5e04b855d095>
CC-MAIN-2024-38
https://www.insidequantumtechnology.com/news-archive/quantum-resistant-encryption-why-its-critical-to-future-cybersecurity/
2024-09-16T23:27:55Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651714.51/warc/CC-MAIN-20240916212424-20240917002424-00761.warc.gz
en
0.936743
1,244
2.75
3
Injection attacks distinguish themselves among these threats as devastating exploitation. Businesses face a variety of cyber risks in today’s interconnected digital environment that might jeopardize their operations. This also compromises sensitive data and results in substantial monetary losses. India experienced an 18% increase in weekly attacks on average during the first quarter of 2023 compared to the same period in 2022, with 2,108 attacks per organization per week. For businesses of all sizes, it is vital to understand the seriousness of injection attacks. Legal liabilities, reputational harm, a loss of confidence from clients, and the disruption of vital activities can all have huge financial ramifications. Organizations may proactively set up robust safety measures. This is done to defend their digital assets and protect against the underlying risks involved with injection attacks. This blog would elaborate on the important topic of injection attacks by looking at their effects on businesses. You would understand explaining the factors that contribute to their prevalence and outlining practical ways to reduce the cyber risk. We will look at the most prevalent kinds of injection attacks, such as SQLi, XSS, code injection, and many more. We will explore how these attacks might affect businesses, presenting examples to demonstrate how serious the issue is. Additionally, we’ll go through reliable preventative measures and recommended practices to lessen the risks of injection attacks. This gives organizations the tools they need to strengthen their defenses and fend off possible attacks. Table of Content Dark Side of Injection Attacks on Business Due to a major flaw in their systems, businesses can seriously be threatened by injection threats. These attacks make use of flaws in operating systems, databases, or online applications. This is done to allow hostile actors to insert and run unauthorized code or commands. A successful injection attack can have disastrous results and wide-ranging repercussions for the targeted company and its clients. Leading cybersecurity organizations’ research has uncovered alarming figures on the impact of injection attacks on businesses. In a recent study, in which more than 60% of the firms polled, encountered a serious security incident. This took place because of injection attacks in the previous 12 months. These occurrences caused significant financial losses, unfavorable legal outcomes, and reputational damage to the concerned companies. The most prevalent suspects of injection attacks that could affect your application are SQL injection and Cross-Site Scripting. However, it’s important to be aware of the wider variety of injection attacks, each with a unique set of objectives and exploitation methods. The following are some of the most common injection attack types to be aware of: SQL Injection Attacks (SQLi) For businesses of all sizes, SQL injection continues to be one of the most prevalent and hazardous types of cyberattacks. SQL injection primarily entails taking advantage of flaws in input fields of online applications that interact with backend databases. Attackers can insert malicious SQL code into the application’s database query by changing user inputs. This will change the application’s intended behavior. Unauthorized access to sensitive data, data manipulation or deletion, and even total system compromise are all possible outcomes of this unauthorized code execution. Attacks using SQL injection can have devastating impacts on enterprises. One of the oldest and most dangerous web-based application vulnerabilities is SQL injection. SQL injection is ranked #6 in the CWE Top 25 for 2021 and is identified as CWE-89: Improper Neutralisation of Special Elements Used in a SQL Command. Cross-Site Scripting (XSS) XSS attacks are a sort of injection, in which malicious scripts are injected into websites that are otherwise safe and reliable. XSS attacks occur when a threat actor utilizes a web application to transmit malicious code to a separate end user, typically in the form of a browser-side script. To exploit an XSS vulnerability, the attacker delivers a text string with malicious code, such as by inserting it into the URL as the user ID parameter. The attack code is then executed by the victim’s browser rather than being viewed as a regular parameter value. Attackers can influence how the victim’s browser interprets and executes the code by inserting malicious scripts into the application’s output. This might result in unauthorized activities or data disclosure. The three basic categories of XSS attacks are stored (permanent), reflected (non-persistent), and DOM-based. XSS has been identified as the second most risky software vulnerability in the CWE Top 25 for 2021 and is classified as CWE-79: Improper Neutralisation of Input During Web Page Generation. Several other types of XSS vulnerabilities, such as stored (permanent) XSS and DOM-based XSS, are to be taken into account. OS Command Injection The reliability and functioning of online applications are put at risk. This is happening to the significant security vulnerability known as OS Command Injection. This type of attack happens when a user-supplied command may be directly executed on the underlying operating system. Malicious actors may inject and carry out unauthorized orders by taking advantage of this vulnerability, potentially compromising the entire system. OS command injection, also known as CWE-78: Improper Neutralisation of Special Elements Used in an OS Command, was ranked #5 on the CWE Top 25 list. Blind command injection and out-of-band command injection are two examples of the several command injection vulnerabilities that should be detected. Code Injection (Remote Code Execution) Code Injection, or more specifically Remote Code Execution (RCE), is a major security vulnerability. This jeopardizes the integrity and confidentiality of web applications. This kind of attack allows hostile actors complete access to and control over the target machine. This is done through remotely injecting and executing arbitrary code. In this section, we examine the workings of code injection (RCE), probable repercussions, and countermeasures to this serious issue. Under CWE-94: Improper Control of Code Generation, code injection is listed. Numerous programming languages and frameworks are covered by the detection of dozens of code execution and code evaluation vulnerabilities. Strong security controls must be put in place to avoid code injection. Validate and fix up all user input to make sure a threat to the code isn’t present. When communicating with databases, use prepared statements and parameterized queries. Strong security measures that reduce the possibility of code injection must be included in online applications to ensure robust security. A risk like this is known as XML External Entity (XXE) Injection. Here the attackers use the processing of XML data to inject malicious entities, possibly exposing data. This causes server-side request forgery (SSRF), or even execution of remote code. CWE-611: Improper Restriction of XML External Entity Reference is an indicator for attack vectors involving XML external entities. XXE injection was ranked #4 in the OWASP Top Ten for 2017. However, in 2021, it was combined with the Security Misconfiguration category. All user input, especially XML input, must be validated and sanitized before processing to prevent XXE Injection attempts. Strict checks should be included for input validation to spot and reject any potentially dangerous or incorrectly formatted XML entities. Preventive Measures for Injection Attacks Organizations may significantly reduce the risks of injection attacks by putting these preventive measures in place. Maintaining a robust and secure web application environment requires regular monitoring, testing, and keeping up with new threats. - Regular Security Audits and Updates: Conduct frequent security audits to find vulnerabilities. Install the appropriate patches and upgrades to the software components utilized in the application. - Steer clear of Dynamic Code Execution: Reduce or avoid the use of eval() or comparable functions. This enables the dynamic execution of user-supplied code. The probability of injection attacks dramatically increases when code is generated dynamically from user input. - Input Validation and Sanitization: Validate and sanitize all user input completely. This is done to make sure it follows the desired formats. The code should be free of malware and other special characters. Put strong input validation procedures in place, and clean up user input by eliminating or escaping potentially dangerous characters. - Utilize Parameterized Queries: Use parameterized queries or prepared statements when interacting with databases or running queries. By separating data from executable code, organizations can stop attackers from inserting malicious code directly into the query. Organizations must put strong security policies in place to reduce the risks brought on by injection attacks. To make sure that user input is free of malicious code, carefully input data validation and ensure sanitization. When working with databases, using parameterized queries and prepared statements can prevent attackers from inserting harmful code or commands. Kratikal, a CERT-In empanelled organization can help businesses avoid injection threats. We can assist in locating and fixing vulnerabilities in online applications with our expertise in vulnerability assessments. We also assist in penetration testing and secure development techniques. Security awareness programs and fostering security-conscious mindsets within organizations are also part of our solutions. Businesses may strengthen their defenses against injection attacks, and protect their digital assets, and client trust. This will also build their reputation by putting these preventative steps into place. Businesses can also collaborate with a reputable cybersecurity firm like ours. To maintain a continued defense against injection attacks and other cyber threats, be watchful. Make sure to have routine upgrade security procedures, and stay aware of emerging risks.
<urn:uuid:b77b624e-d0b8-4710-bbee-f38bfd31b73b>
CC-MAIN-2024-38
https://kratikal.com/blog/how-injection-attacks-can-cripple-your-business/
2024-09-19T10:37:04Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652028.28/warc/CC-MAIN-20240919093719-20240919123719-00561.warc.gz
en
0.929423
1,919
2.546875
3
I recently heard about a programming language developed at MIT called Scratch. According to the website (http://scratch.mit.edu/), “Scratch is designed to help young people (ages 8 and up) develop 21st century learning skills. As they create and share Scratch projects, young people learn important mathematical and computational ideas, while also learning to think creatively, reason systematically, and work collaboratively.” [more] The Scratch programming language makes the syntax easy for kids by using a building block method. There is also an option to share programs via the Internet. I am not sure what kind of content controls are in place when sharing programs, so be careful with the kids on that one. Scratch even has the ability to incorporate physical sensors into its programs using a circuit board you can purchase from them. Anyway, this is one of the better kid-friendly programming environments I’ve seen, so I thought I would pass it along for those of you with young ones.
<urn:uuid:abbbe867-e2e8-468e-9d25-deec627c63a6>
CC-MAIN-2024-38
https://conetrix.com/blog/scratch-programming-for-kids
2024-09-08T13:46:40Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651002.87/warc/CC-MAIN-20240908115103-20240908145103-00661.warc.gz
en
0.965269
206
3.078125
3
Political organisations and election candidates have become targets for threat actors that wish to disrupt and interfere in the democratic process. This can be part of a wider hybrid campaign to influence voters. Cyber attacks that target election candidates or organisations can be very damaging to the candidate themselves, the political party they represent, or to society’s overall trust in the democratic process. This guidance aims to raise awareness of cyber threats to democratic processes and institutions, and to help prevent attacks on both organisations and individuals. What does this document cover? This advisory is focused on cyber security issues which pose a risk to the security of the electoral process. It should not be considered as a comprehensive guide for the overall security of an individual candidate or political parties’ data or systems. This advisory recommends implementing identity and access management policies, enhancing website security, preventing spoofing, educating constituents about misinformation, and preparing for ransomware and deep-fake attacks to safeguard the integrity of elections. Securing your accounts You or your team’s personal accounts are prime targets for attackers. If compromised, attackers could access your stored information. If possible, you should prioritise the use of corporately managed accounts and devices for professional tasks, as they benefit from centralised management and enhanced security measures. If you do not have access to such software, implementing the following measures can still greatly reduce this risk. - Use of strong passwords: Create complex passwords by using a minimum of 12 characters which include numbers and symbols to increase complexity. Alternatively, use three random words and make each password unique for every account, especially for critical accounts like email, social media, and online banking. Consider using a password manager to help you remember these different passwords. - Multi-Factor Authentication (MFA): MFA should always be used when logging into internet facing accounts – this could be an app or website, or an email account. Use an authentication app like Google Authenticator or Microsoft Authenticator. MFA adds an extra layer of security, ensuring that even if an attacker knows your password, they still cannot access your account. If multi- factor authentication is not enabled, it is entirely possible for an attacker to use stolen credentials or to ‘brute force’ access to an account by simply guessing a password. - Social media use: You should check for unusual logins. Set up notifications to send a text or email alert when your account is accessed from a new device or location. Exercise caution when sharing personal information publicly. Review your security settings to control who can view your information. Avoid accepting message requests from unfamiliar accounts; consider calling to verify their identity first. Please refer to the individual social media guides linked to on the Coimisiún na Meán site.
<urn:uuid:12123eaa-f7f7-4141-9ca7-5761c9c00742>
CC-MAIN-2024-38
https://ciso2ciso.com/cybersecurity_political_organisations_election_candidates/
2024-09-10T23:12:25Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651323.1/warc/CC-MAIN-20240910224659-20240911014659-00461.warc.gz
en
0.911576
554
2.71875
3
Coronavirus Phishing & Other Covid-19 Scams | Resources And Security Alerts In the Coronavirus phishing campaigns, emails are disguised to look like they are being sent by medical authorities like the Centers for Disease Control and Prevention (CDC) and World Health Organization (WHO). An example is a fake email claiming to have been sent by the CDC. These emails are asking recipients to click on a link to learn about safety measures or new cases in their cities. Eventually, the recipient is asked to share financial details like the bank or credit card details or personal information like social security numbers. Other forms of coronavirus phishing emails include fake workplace policy emails that appear to be sent by organizations. They target remote employees and include links to log in to fake company pages. The motive is to obtain user credentials. So, how do you spot coronavirus phishing mails, other Covid scam attemps, and protect yourself? The first thing to remember is that the objective of these emails is the same as any other phishing email – to make you click on an embedded link or open an email attachment. Here are some tips to easily spot a Coronavirus phishing email: 1. Check the email address Hackers are good at creating fake email addresses that closely resemble legitimate ones. For example, email addresses with the website domain as “@cdc-gov.org” or “@who-pc.com.” First, determine if these domains are genuine and operational before acting upon the email. Here is an example of a phishing email with a fake email address: 2. Do not click on suspicious links in an email Cybercriminals include genuine-looking links in the phishing email to make people click on them. However, when the user clicks on them, they are redirected to a different target URL of a phishing site. Before clicking on any such links, hover your mouse over the link and view the URL link where it would lead. Here’s an example of a fake CDC phishing email containing a suspicious link: 3. Beware of phishing SMS Emails are not the sole medium through which hackers target users for phishing. Cybercriminals could also send an SMS with suspicious links. When you click on the link, you are taken to a website that could steal your financial information. These cases are far sneakier as our mobile phones hardly have the security measures we have on our work laptops or computers. 4. Beware of social media posts Cybercriminals know that people are on the lookout for virus-related information on social media, and therefore, social media is a fertile ground for these activities. 5. Do not submit any personal information As mentioned before, phishing emails aim to obtain your personal or financial information. As a rule, do not submit any confidential information as a response to such emails. Banks or any other legitimate agencies don’t ask for any such data or any login credentials. Additionally, delete any emails that ask for personal information on an immediate or urgent basis. Times of crisis require that we are well-informed and responsible. It is important that we secure ourselves and those around us. Just like we are participating in social distancing and self-quarantining offline to stop the spread of the Coronavirus, let us commit to stopping the spread of such Covid scams and misinformation. Request More Information ICS is an award winning Microsoft Certified Silver Partner and Small Business Specialist As a Microsoft Certified Partner, we are experts at managing and deploying Microsoft® Windows® networks. Our network solutions help you protect your data, increase productivity, and present a more professional image to customers.
<urn:uuid:280b5188-2753-44b2-a530-25f4efcf1dfd>
CC-MAIN-2024-38
https://www.icssnj.com/covid-scams-phishing-security.html
2024-09-10T23:31:27Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651323.1/warc/CC-MAIN-20240910224659-20240911014659-00461.warc.gz
en
0.929919
770
2.953125
3
Most people have heard the story of the Trojan horse. We’ve learned the moral of that story: don’t trust gifts from strangers. Today, nobody would use a giant horse statue – except maybe pranksters with a bad sense of humor – although similar tricks exist in digital form. People are wary of unknown .exe files and sudden downloads from corny pop-up ads. They trust more familiar file types such as .jpg. But in the latest incarnation of the Trojan horse, malicious code can be hidden in images, allowing cybercriminals to take over Android devices. Google moves quick and has already issued an Android update. Once those software changes occur on a device, this loophole will be resolved. Another positive, this tactic hasn’t been found in common use – it was a discovery by cybersecurity researcher Tim Strazzere. So ultimately, the photo hack is not a widespread threat. But people should be aware and educated nonetheless. Understanding how cybercriminals act is the first step to knowing how to stay safe online. Truth be told, hiding malicious code in other forms isn’t a new tactic. Crooks have concocted all manner of disguises. Infected Microsoft Word documents have even been found. When users launch these infected files on their PC, malware is installed to steal sensitive information. The photo method is only the latest in this context. What’s unique in this case is the caveat that villains can get their way without images being clicked on. Let’s jump into brief technical details to explain how. When photos are sent to your Android device, some of that data is processed before users open the file. This is known as parsing in computer terms: the unpacking and separating of information. See, smartphones need to know what data to process when users open a file. But first, they must look at other data to understand the nature of the file type. That’s where cybercriminals could package secret instructions to be triggered from the get-go. What happens then? Well, the malware can “brick” users’ devices, meaning they are rendered unusable. Cybercriminals then gain remote access to the smartphone. With all the sensitive information our devices now store, crooks could snatch quite a bit! Think of all those financial details, passwords for online accounts, and emails. Additionally, perpetrators could even use the phone’s applications to their gain, and the user’s expense. Now remember, Google already issued an Android update to combat this – so there’s no need to panic. The threat, however, should open some people’s eyes. A range of sneaky threats exist out there, and education is key. This won’t be the last reincarnation of the Trojan horse. But with the right security knowledge, people can help to keep themselves safe in our digital age. With that in mind, here are three cybersecurity tips to act on: - Don’t over-expose yourself on social media and chat apps. For this photo trick to hijack your Android device, someone has to first send you an image. How will they do so? Probably over Facebook message or another chat app. So think of security in layers: at the very perimeter, don’t allow just anyone to contact you. This also protects you against social media bullying. - Update your device as soon as possible. Large companies have cybersecurity teams working to protect customers, but updates may depend on users. In this case, once Google received the report, the latest Android update was tailored to resolve the vulnerability. Make sure to update your devices and applications as well. - Be wary where you browse and what you open. As threats evolve, cybercriminals continue to serve malware in advanced methods. Some people could receive an email with an infected Microsoft Word attachment, while others could accidentally end up on an unsecure website. Always make sure what’s on your screen is authentic. Double check email senders. Look for the official URL. Be mindful and aware.
<urn:uuid:dac540ab-a160-4382-99f0-d3608770ccdc>
CC-MAIN-2024-38
https://www.mcafee.com/blogs/mobile-security/android-photo-hack/
2024-09-10T23:43:39Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651323.1/warc/CC-MAIN-20240910224659-20240911014659-00461.warc.gz
en
0.935234
833
2.875
3
The Ultimate Guide to MySQL ODBC Connections July 13, 2023In this comprehensive guide, we delve into the world of Open Database Connectivity (ODBC), a universal data access method that plays a pivotal role in database connectivity. We’ll also explore the significance of MySQL, one of the most widely used open-source relational database management systems globally. By understanding the synergy between MySQL and ODBC, you can unlock new levels of efficiency and effectiveness in your data management tasks. Let’s embark on this journey to master MySQL ODBC connections. About MySQL and ODBC MySQL is a highly popular open-source relational database management system (RDBMS) used by developers worldwide. It’s known for its speed, reliability, and ease of use. MySQL supports a broad subset of ANSI SQL 99, as well as extensions and specific features for compatibility with MySQL Server versions. Key features of MySQL include its robust data security, scalability, and flexibility, making it a preferred choice for many web-based applications, data warehousing, and logging applications. Open Database Connectivity (ODBC) is a standard Application Programming Interface (API) that allows external applications to access data from a common database or cloud application. It provides maximum interoperability, meaning a single application, independent of any data store, can access data through an ODBC driver. This driver serves as an interface between an application and a data source. The ODBC API is independent of any DBMS, programming language, and operating system. The MySQL ODBC Connector, also known as MySQL Connector/ODBC, plays a crucial role in this context. It is an ODBC driver that enables applications and programming languages to access data in MySQL database management systems through the ODBC interface using SQL. This connector is easy to install and configure, and it allows applications to communicate with MySQL databases, retrieve data, and run queries. It is a key component in ensuring seamless interaction between applications and MySQL databases. Installation of MySQL ODBC Connector Follow the steps below to install the MySQL ODBC Connector on your Windows machine: - Download the Installer: Visit the MySQL official website or a trusted source to download the MySQL ODBC Connector installer. - Run the Installer: Locate the downloaded file on your computer and double-click it to run the installer. - Overwrite Warning: If you already have a version of the MySQL ODBC Connector installed, you might receive a warning about overwriting existing files. It’s recommended to uninstall the previous version before proceeding with the new installation. Click ‘Yes’ to proceed. - Select Components: During the installation process, you’ll reach a ‘Select Components’ page. Here, you can choose whether to install the 64-bit version of the driver. If you don’t need the 64-bit version, uncheck the box. There’s also an option to install the Help and Manual files. - License Information: In the ‘License Information’ dialog box, select the license type and activate the product. If you don’t have an activation key, select ‘Trial’ to use the driver for evaluation purposes. - Enter Activation Key: If you have an activation key, select the ‘Activation Key’ option. Copy the activation key from your registration email or your Customer Portal account and paste it into the ‘Activation Key’ edit box. - Load Activation Key File: If you have an activation key file, click the ‘Load Activation Key’ button and browse to the location of the file on your computer. - Proceed with Installation: Click ‘Next’ to proceed with the installation. - Install and Finish: Click ‘Install’ to begin the installation process. Once the installation is complete, click ‘Finish’. - Configure the Driver: After the installation is completed, you will need to configure the MySQL ODBC Connector to suit your specific needs. Remember to restart your system after the installation to ensure that all changes take effect properly. Step-by-Step Guide to Connect to a MySQL Database with MySQL Connector/ODBC Follow these steps to connect to a MySQL database using the MySQL Connector/ODBC: - Install MySQL Connector/ODBC: If you haven’t already, download and install the MySQL Connector/ODBC from the official MySQL website or a trusted source. Follow the installation instructions provided. - Open the ODBC Data Source Administrator: Type “ODBC Data Sources” in the Windows search box and choose the application that matches the bitness of your application (32-bit or 64-bit). Alternatively, you can open ODBC Data Sources from Control Panel > Administrative Tools. For a 32-bit DSN, you can run C:\Windows\SysWOW64\odbcad32.exe, and for a 64-bit DSN, run C:\Windows\System32\odbcad32.exe. - Select the DSN Type: In the ODBC Data Source Administrator, select either the User DSN or System DSN tab. Most applications work with both types, but some require a specific type of DSN. - Add New Data Source: Click ‘Add’ to create a new data source. The ‘Create New Data Source’ dialog will appear. - Select MySQL ODBC Driver: In the list of drivers, select ‘MySQL ODBC Driver’ (or ‘MySQL Connector/ODBC’ depending on your version) and click ‘Finish’. The driver setup dialog will open. - Enter Connection Information: In the setup dialog, enter the connection information for your MySQL database in the appropriate fields. This typically includes the server’s name or IP address, the database name, and your username and password. - Test the Connection: Click ‘Test Connection’ to verify that the settings are correct and a connection can be made to the MySQL database. - Save the DSN: Click ‘OK’ to save the Data Source Name (DSN). Your application can now use this DSN to connect to the MySQL database through the MySQL Connector/ODBC. Remember, the exact steps and options might vary slightly depending on your version of Windows and the MySQL Connector/ODBC. Alternative Method: Connecting to a MySQL Database with OData While MySQL ODBC Connector provides a robust and reliable way to connect to MySQL databases, there are alternative methods that can be used. One such method is using OData (Open Data Protocol), a widely accepted open standard for data access over the Internet. Skyvia, a cloud data platform, allows you to connect to any data source that provides an OData interface for data access over the web. This includes MySQL databases. Skyvia supports importing data to and from OData endpoints, exporting their data to CSV files, and replicating their data to relational databases. To connect to an OData endpoint, you need to specify its URL, and, if necessary, provide authentication parameters. Here’s how you can do it: - Specify Server: Enter the OData endpoint URL. - Select Authentication Method: Skyvia supports OData endpoints without authentication and with basic HTTP authentication. - Enter User and Password: If you’re using basic HTTP authentication, enter your username and password for the OData endpoint. - Specify URL Parameters: If necessary, you can specify custom query string parameters for HTTP requests. The value should be encoded as a part of a URL and have the usual query string format: param1=value1¶m2=value2¶m3=value3 Once you’ve set up the connection, you can use Skyvia’s data integration features to interact with your MySQL database. However, it’s important to note that Skyvia does not support OData endpoints in synchronization packages and does not support Replication with Incremental Updates selected for OData endpoints. When updating data via UPDATE statements in Query, you can only provide constant values for the columns. You cannot use expressions with current column values. In conclusion, while OData provides an alternative method for connecting to MySQL databases, it has its own set of features and limitations. Depending on your specific needs and circumstances, you might find that using OData is a more suitable option than using the MySQL ODBC Connector. The MySQL ODBC Connector is a powerful tool that offers a standardized and efficient way to connect your applications to MySQL databases. Its versatility and wide acceptance make it a valuable asset in the world of database management. By leveraging the MySQL ODBC Connector, you can unlock new levels of interoperability, efficiency, and data integration. Whether you’re a seasoned database professional or a beginner, we encourage you to explore the benefits of using the MySQL ODBC Connector. Start connecting your applications to MySQL databases using this tool and experience the difference it can make in your data management tasks.
<urn:uuid:1b941cf0-6f59-4c07-bf5b-f98b00d5b62d>
CC-MAIN-2024-38
https://cloudcow.com/content/the-ultimate-guide-to-mysql-odbc-connections/
2024-09-14T17:42:41Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651580.73/warc/CC-MAIN-20240914161327-20240914191327-00161.warc.gz
en
0.826571
1,891
2.828125
3
Raymond Chan: Okay. So thank you for watching this abstract presentation for the DFRWS 2021. So, I’m Raymond, and today I would like to present the topic about tracing a walking path using Smart Lighting System. So, just to give you a brief introduction of what actually the Smart Lighting System is. A Smart Lighting System actually is a subsystem under the building management system, which is called the BMS. So because the BMS actually is used to monitor and control the electronic devices within the building. So in our university, because we have an Advanced Cybersecurity Lab, so we also want to explore the possibilities of using IOT systems in our campus. So that’s why we installed a Smart Lighting System in our Advanced Cybersecurity Lab. Not only to figure out are there any security or vulnerability issues, but also we would like to explore the forensic capability for the Smart Lighting System, as well. So as you can see in this image, this is our Advanced Cybersecurity Lab. So all the light within the cybersecurity lab actually is controlled, or actually is, we installed the Smart Lighting System in one cybersecurity lab. So that when we go into the lab, the light will automatically turn on and turn off. So that actually is also saving the power bill, as well. So, this figure is our Smart Lighting System layout. So you can see, in fact, the Smart Lighting System consists of a web application as you can see here, the web server, and this is a Zigbee gateway that is wirelessly connected to the LED light. So the LED light actually activates the Zigbee controller, and there’s a sensor to detect whenever there’s a human walking through to it. So, the application actually is not only used to monitor the situations of the LED light, but also actually you can change the behavior of the LED light as well for better management of our lab environment. So the LED light actually is wirelessly connected using the Zigbee control. So that’s why there’s a Zigbee gateway right there to send them to the receiver and send the signal to the LED light directly. So this is another figure to show the Smart Lighting System layout. So you can see in here we have 15 lights and in total there are seven sensors. So when someone passes by the sensor you can see the light actually will automatically turn on. So, using these figures as an example, so, if there’s someone walking through sensor 1, then you can see that lights 2 and 3 will automatically turn on. So if he walks inside to the left, then the sensor, for example, sensor 7 will then automatically detect someone passed by it and lights 6 and 12 will also automatically turn on, and so on and so forth. So, because as you know that in the PUS diagram, I showed you that there are web applications that can retune the sensor signal. So in fact, we can also utilize that to see if we can make it as a forensic use, because it is where the information is very useful if I can know someone goes into the lab and his walking pattern, as well. So, actually we developed a Python program to reshift the information from the web applications by ourselves. And then we curate those sensors periodically to see whenever the light is now turning on or off. And we can also show then, we just put it in a CSV format to actually record the pattern and the behavior of the sensor, as well as the LED light on/off behavior. So, actually, the results are quite impressive because when we are trying to analyze those data, we can track when someone walks into the lab, and actually which location he is currently located at. So, for example, if you’re using this prediction as an example, you can see the first indicator in the bottom left corner is that the LED light 2 actually is being turned on at 12:49 for 15 seconds. And then from time to time, you can see that LED 6 and LED 11 actually turn on later on continuously, and then LED 7 is also being turned on, as well. So using this kind of information, initially, you will only know that, “Oh, okay. That light is turned on. So if light 7 turned on after 11, then actually it’s just the pattern of the LED light.” But, however, we can actually use this information to actually track if someone is walking to the PC. And then accordingly, so you can also see, so, if the LED light actually is being turned on and turned off, that means someone is actually walking far away from the sensor. So we also do this experiment and then we just go in to have a run through the lab. And then we found out that the LED light information is actually very useful and very accurate to detect when someone will pass by within our Advanced Cybersecurity Lab. So, to conclude, actually, we discovered that actually we can use the sensor data of the Smart Lighting System for forensic investigations, because in the old days we only relied on the CCTV, which in most of the cases, if sometimes the CCTV is not functioning, then we might not be able to check if someone is really walking near to some locations or not. So using the Smart Lighting System sensor data can become a cost reference check. If there is also CCTV that’s right there, then we can actually adjust and synchronize the time to actually have a stronger proof of if someone is walking near a machine or near other devices. Also, if we do not have CCTV, actually this information can also be considered as a very useful digital evidence for tracking if someone is walking within indoor locations or either outdoor locations, if possible. So in the future, basically we would like to actually set up a smart living lab on the system and we will actually test more subsystems and find out their forensic capabilities because we are actually trying to build up a so-called “Smart Campus”. So we feel that actually there will be a lot of useful sensor data that can provide small forensics capabilities, and it can also be considered as some kind of digital evidence. And that’s all from my presentation.
<urn:uuid:266660de-94e4-4136-b761-47cfeb837e43>
CC-MAIN-2024-38
https://www.forensicfocus.com/webinars/tracing-walking-path-using-smart-lighting-system/
2024-09-15T22:38:38Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651668.26/warc/CC-MAIN-20240915220324-20240916010324-00061.warc.gz
en
0.955908
1,279
3.203125
3
IoT Device Management vs. IoT Platform The Internet of Things (IoT) is a big word. The IoT ecosystem consists of devices, big or small, networks, protocols, connections, clouds, databases, etc. Essentially, the IoT ecosystem can be broken down into several significant components or building blocks, with thousands of vendors offering solutions for each of these blocks. Here are the major IoT building blocks: - Connectivity – devices or objects become IoT manageable after they connect to the Internet, directly to a cloud or via IoT gateway. - Device Management – the part of the IoT platform that enables provisioning and authentication of the devices, remote configuration and management, monitoring and diagnostics, and software (or firmware) updates and maintenance - Applications – IoT applications or dashboards that enable IoT visualization and control - Security – vendors that keep IoT solutions secure - Data and Analytics – IoT component that collects, stores, and analyzes data. When IoT Platform is not precisely a platform Often, when people say “IoT Platform,” they really mean “IoT Application”. IoT apps enabling limited IoT functionality are usually enough for basic IoT scenarios. They can connect devices and sensors to the cloud, monitor and collect data, and provide IoT project visualization. We see quite a few such IoT platforms offered by hardware manufacturers that introduce these platforms in addition to or as part of their hardware offering. Basic IoT Platforms are sufficient for small to medium-sized IoT projects, but large-scale IoT scenarios require advanced IoT functionality. Device Management component is a must when the IoT scenario requires a connection, management, and integration of thousands of IoT devices and sensors. Why IoT Device Management? The IoT Device Management enables provisioning and authentication of the devices, remote configuration and management, monitoring and diagnostics, and firmware or software updates and maintenance. Imagine a large IoT project. Thousands of devices, numerous hardware vendors, multiple connectivity protocols – how do you make all of it become IoT manageable? IoT Device Management providers usually provide advanced features that enable the management of various types of IoT devices and sensors on a unified platform. Friendly Technologies, for example, is utilizing its Smart Layer™ technology to translate between multiple protocols and manage all types of IoT devices via its unified One-IoT™ Device Management platform. IoT service providers are interested in generating revenue and maintaining profitability from selling their IoT services. When it comes to reaching and maintaining the profitability of IoT projects, the primary way to achieve this goal is to avoid the lock-in by a specific hardware vendor. IoT providers must have the freedom to choose the best of breed devices and sensors that do not generate dependency that can lead to profitability decrease in the future. Another example of a significant IoT deployment challenge that IoT providers face is dealing with loads of manual work to make devices and sensors IoT manageable – customization, creating configuration files, development, homologation, testing, etc. This manual work can quickly turn into immense costs. Automated provisioning of the IoT devices that helps avoid such expenses is another advanced feature that only experience IoT Device Management providers can solve. Here is a comparison of the basic (generic) IoT platform and advanced IoT Device Management Solution: At Friendly Technologies, we aim for maximum automation and ease of use with no or minimum required professional services. We have acquired extensive experience with large-scale deployments in low ARPU markets with large numbers of devices. Our carrier customers can profitably compete in the LPWaN low ARPU market. Our IoT Line of products includes IoT Device Management, a full Smart Home solution, and embedded IoT clients. Friendly Technologies applies an open standard approach, enabling its customers the freedom to choose any hardware vendor. Main Features of Friendly One-IoT™ Device Management Solution: - Automated provisioning - Sensor and gateway management - Manage multiple protocols on one platform: LwM2M, MQTT, TR-369 USP, OMA-DM, CoAP, and others - Device diagnostics and repair - Remote device configuration - Group management - FW control - Monitoring and data collection - Event triggering - Unified API - Admin & Technician portals Would you like to schedule a demo of Friendly’s One-IoT Device Management solution? Are you looking for additional information? Please, do not hesitate to contact us. We will be in touch shortly.
<urn:uuid:6b61470b-a258-476d-b9aa-39b35d284343>
CC-MAIN-2024-38
https://friendly-tech.com/iot-platform-vs-iot-device-management/
2024-09-19T14:15:07Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652031.71/warc/CC-MAIN-20240919125821-20240919155821-00661.warc.gz
en
0.897465
911
2.5625
3
Encrypting documents and fields A document is considered to be encrypted if it is created from a form that contains one or more encrypted fields. Each encrypted field is linked to a key that encrypts the contents of the field. An encryption key can be secret -- that is, a key that you must send to users in order for them to decrypt a field -- or public -- that is, a key that is already in a user's ID file and in the user's Person document where it is publicly available. Public key and secret key encryption Notes® uses public key encryption for electronic mail, and Domino® Designer also lets you use public key encryption for encrypting fields in documents. Every user has a unique public key associated with their user name and stored in their user ID. Applications reference the keys by the users' names in a special field called PublicEncryptionKeys. When a document is saved, all the user names in this field are located in the Domino® Directory or the user's personal address book, the corresponding keys are retrieved, and all fields marked with a special property are encrypted with those keys. Domino® Designer also supports secret key encryption that you can use for encrypting fields in documents. You can create and name secret keys and then distribute the secret keys to users so that they can decrypt the protected data. Secret keys, like public keys, are stored in a user's ID. Applications reference the keys by their names in a special field called SecretEncryptionKeys. When a document is saved, the keys named in this field are retrieved from the user's ID file, and all fields marked with a special property are encrypted with those keys. If you are planning to use secret encryption keys rather than encrypting with a public key, create the secret key before you encrypt a document. You can encrypt documents with keys in several ways: - Using public keys. You can encrypt documents with public keys on IDs so that only users with those IDs can read the documents. To do this, you enter one or more names in the Public Encryption keys field on the Security tab in the Document Properties box. - Using a form property. Database designers can use a form property to add one or more keys to a form. Every document created with the form will be encrypted using the encryption keys. - Using the Database/Document Properties box. Users can use the Database/Document Properties box to encrypt one or more documents with their own encryption keys stored in their ID files. To use the properties box to encrypt documents, the form must contain a field that can be encrypted. - Using the SecretEncryptionKeys field. The SecretEncryptionKeys field can contain either the name of a key, which is automatically used to encrypt documents, or the field can be blank, allowing users to assign the encryption key. To encrypt a field with a secret key using either method, users must have it stored in their ID file. You can set up forms with text or keyword fields that allow the user to choose whether to encrypt a document. Designers can also hide the SecretEncryptionKeys field so that users cannot see the names of the encryption keys. A database designer can encrypt fields with secret encryption keys. To decrypt these fields, users must merge the secret encryption keys into their ID files. If the user does not have the required encryption key, the encrypted fields appear blank.
<urn:uuid:c573f431-b058-4d9f-b4ac-ac20fb0010b8>
CC-MAIN-2024-38
https://help.hcl-software.com/dom_designer/10.0.1/basic/H_DOCUMENT_AND_FIELD_ENCRYPTION_OVERVIEW.html
2024-09-19T15:28:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652031.71/warc/CC-MAIN-20240919125821-20240919155821-00661.warc.gz
en
0.900351
693
2.734375
3
Kubernetes was initially developed by Google and is at the moment maintained by the Cloud Native Computing Foundation. Kubernetes is an open-source system meant to help automate the deployment, scaling, and management of containerized workloads, services, and apps over clusters of hosts, by organizing app containers into pods, nodes (physical or virtual machines), and clusters, with nodes forming clusters managed by a master which coordinates cluster-related tasks such as scaling or updating apps. The malware was dubbed as Siloscape by the security researcher Daniel Prizmant and it seems to be the first one to target Windows containers, exploits, known vulnerabilities impacting web servers and databases with the end goal of compromising Kubernetes nodes and backdooring clusters. Siloscape is heavily obfuscated malware targeting Kubernetes clusters through Windows containers. Its main purpose is to open a backdoor into poorly configured Kubernetes clusters in order to run malicious containers. Unit 42 researchers have previously only seen malware targeting containers in Linux due to the popularity of that operating system in cloud environments. Siloscape works by compromising the web servers and then using various container escape tactics to achieve code execution on the underlying Kubernetes node. The compromised nodes are probed for credentials allowing for the malware to spread to other nodes in the Kubernetes cluster, in order to establish communication in the stage of the infection with its command-and-control (C2) server via IRC over the Tor anonymous communication network and therefore listen for incoming commands from its masters. After gaining access to the malware’s C2 server, Prizmant managed to identify 23 active victims and also found that the server was hosting 313 users in total, this possibly being an indicator that Siloscape is just a small part of a much wider campaign. Investigating the C2 server showed that this malware is just a small part of a larger network and that this campaign has been taking place for over a year. Furthermore, I confirmed that this specific part of the campaign was online with active victims at the time of writing. Most malware that are targeting cloud environments focus on secretly mining for cryptocurrency on infected devices and on abusing the infected systems for launching DDoS attacks, Siloscape has a different agenda. Siloscape does its best to evade detection, so it avoids any actions that could alert the compromised clusters’ owners to the attack, including cryptojacking. Your perimeter network is vulnerable to sophisticated attacks. Heimdal® Network DNS Security Is the next-generation network protection and response solution that will keep your systems safe. - No need to deploy it on your endpoints; - Protects any entry point into the organization, including BYODs; - Stops even hidden threats using AI and your network traffic log; - Complete DNS, HTTP and HTTPs protection, HIPS and HIDS; Its only goal seems to be to backdoor the Kubernetes clusters, in this way being able to open the way for its operators to abuse the compromised cloud infrastructure for a broader range of malicious pursuits, like credential theft, data exfiltration, ransomware attacks, or even supply chain attacks. Compromising an entire cluster is much more severe than compromising an individual container, as a cluster could run multiple cloud applications whereas an individual container usually runs a single cloud application. The Kubernetes admins should switch from Windows containers to Hyper-V containers in order to ensure that their cluster is securely configured to prevent any malware like Siloscape from deploying new malicious containers.
<urn:uuid:6e2a2d57-0f75-4d0d-bdcc-e6bceca6927c>
CC-MAIN-2024-38
https://heimdalsecurity.com/blog/new-kubernetes-malware-backdoors-windows-containers/
2024-09-20T21:17:22Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725701423570.98/warc/CC-MAIN-20240920190822-20240920220822-00561.warc.gz
en
0.939467
735
2.609375
3
The information age is upon us. Every day, more and more of the world’s data moves to digital formats—from financial transactions to medical records. But with this transition comes new risks: By 2025, cyberattacks are projected to cost businesses $10.5 trillion per year globally. One major attack can cripple an entire organization for years to come, not just in terms of finances but also in customer trust and brand reputation. With APIs powering the entire financial sector, many companies have started adopting API security strategies. This article will discuss the role of APIs in financial services, some of the most common threats faced by APIs today, and how the FinTech industry can take proactive steps to fight back against cybercriminals. What Is an API and Why It’s Important for the Financial Sector? An API, or application programming interface, is a type of interface that allows data to be shared with other applications. Think of it as a server at a restaurant that acts as a middleman between the customer and the kitchen, helping serve food faster while enhancing operational efficiency for the chef. By the same token, APIs often provide a more streamlined way for businesses to interact with an organization's services. These APIs can be used to programmatically create, read, update, and delete database records, allowing organizations to run complex process chains within seconds - by the time the money you sent to your friend reaches their account, the transaction may have gone through dozens of APIs. Read More: API Terminology: A Complete List of Terms for Beginners APIs are important for the financial sector because they allow for the easy sharing of financial data between different banks, FinTech services, and other institutions. Most of the major banks and financial services firms in the US have an API available for public use. In fact, 78 percent of banks rely on banking APIs to provide a better customer experience and develop new revenue streams. The most common example of an API is the Stripe API, which makes its data easily accessible for third-party apps and services. With unprecedented monetization opportunities, implementing APIs is becoming more and more popular in the financial sector. A good example may be a bill-paying service offered at many banks that can now be integrated directly into other banking sites utilizing APIs rather than requiring users to log into the bank's site and initiate usage of the service from there. This integration eliminates intermediaries or potential points of vulnerability when sharing highly sensitive information between banking entities. The biggest concern with APIs is security. With a solid security strategy in place, APIs are relatively safe compared to using direct web services calls or other methods of data exchange because they provide a layer of abstraction between both ends of the connection. APIs also provide an easier way to filter out traffic that should not be allowed to access specific resources on either end. Suppose someone wants to implement check imaging into their application but doesn't want other applications calling it directly. In that case, they can create a rule in the authentication scheme that requires applications to be authenticated via the API to access that resource. In this way, APIs can provide faster data exchange while still being relatively simple and easy to use for developers. Read More: What is API Security and Why It's Important The 7 Most Common Types of Attacks Against APIs Unfortunately, the rise of APIs and open banking did not go unnoticed by hackers - APIs are projected to become the main attack vector in 2022. API attacks come in many different forms, but they all have one common goal - to steal or manipulate data. Below, we explore some of the most common types of API attacks that can cripple financial institutions and lead to disastrous reputational and financial losses. 1. DoS & DDoS Attacks One of the most common clusters of API attacks is an application-level denial of service (DoS) and distributed DoS (DDoS) attacks. In this type of attack, the hacker sends a massive number of requests to the API to overwhelm it and prevent legitimate users from accessing it. Using this approach, the attacker tries to overload your API with requests until it's unable to handle them and shuts down. There are several ways to protect your API against DDoS attacks, including using a load balancer or an intrusion prevention system. 2. SQL Injection Attacks In a SQL injection attack, the attacker tries to gain unauthorized access to databases by injecting malicious code into database requests. To protect your API against SQL injection attacks, you should use parameterized queries and ensure that all requests to your API are authorized, authenticated, and validated. 3. XML External Entity (XXE) Attacks When an XXE attack occurs, the hacker tries to access files on the server or other external services using specially crafted XML documents. You can prevent your API from being exploited with XXE attacks by configuring your XML parser correctly. 4. Cross-site Scripting (XSS) Attack Another common type of API attack is a cross-site scripting (XSS) attack. In an XSS attack, the attacker tries to inject malicious scripts to steal sensitive data or gain unauthorized access to the functionalities of your APIs. You can protect your API against XSS attacks by configuring it to use X-Frame headers and Content Security Policy (CSP) headers. 5. Brute Force Attacks A brute force attack entails using automated tools to try different username/password combinations until they guess the correct one and gain unauthorized access to the system. To protect your API from brute force attacks, you should use rate limiting and IP address restrictions on your API. 6. Cross-site Request Forgery (CSRF) Attacks A CSRF attack occurs when the attacker tricks authenticated users into clicking a specially crafted link to make unauthorized requests to your API on behalf of the attacker. To protect against this, you can use a CSRF Protection library and custom tokens in your requests. 7. Man-in-the-middle (MITM) Attacks A man-in-the-middle (MITM) attack occurs when the bad agent intercepts traffic between two parties to access private information. An easy way to defend against MITM attacks is to ensure that your connections are secure using SSL/TLS encryption. Another option is enforcing client authentication - when both the user and the API request certificates from a trusted third party before exchanging data. This stops MITM attacks from occurring because both parties have been validated by an authority figure, blocking unauthorized requests from getting through in between. Read More: Critical API Security Risks: Understanding Cyber Threats to APIs How to Protect your FinTech Business Against API Attacks While these tips are helpful to cover some of the loopholes in API security, it’s nowhere near enough to provide a safe space for your customers. Let's cover the building blocks of any API security strategy to empower you to tackle this issue systemically. 1. Zero in on Business Logic Flaws Business logic flaws, the practice of abusing the legitimate functionalities of an API to reach a malicious goal, are the #1 security risk in apps and APIs. This cluster of vulnerabilities is the most dangerous and the most difficult to detect. Hackers might exploit a business logic flaw to access your API's back-end systems, leading to a data breach or other nefarious activity. Some examples of business logic flaws include: - Hardcoded credentials: This involves storing usernames and passwords within the source code, making it easy for anyone to access login credentials. - Insecure direct object references: This occurs when an application identifies an object by its primary key rather than using an ID from an associated table, giving more context about the model being used. - Dynamic SQL statements: This entails using dynamic SQL queries to query databases or other back-end systems. Hackers can exploit it if your application uses dynamic SQL statements without first parameterizing all user input. - Loosely Defined Business Processes: Without proper security measures, anyone who knows how to act within the bounds of a loosely defined process can use it in a way unintended by developers to access sensitive information. 2. Use Strong Authentication and Authorization Mechanisms Authorization and authentication vulnerabilities took the top spots of the OWASP API Security Top 10 list, making them the most common types of API vulnerabilities. Read More: What Is OWASP API Security Top 10 & Why It's Important That’s why enforcing robust authentication and authorization mechanisms should be a top priority for both traditional financial institutions and innovative FinTech startups alike. Things like Basic Authentication (using the standard combination of a login and password to authenticate users) and hardcoded credentials should be banished from your APIs - forever. Instead, use OAuth or OpenID to validate the identity of your users and add an extra layer of security with two-factor authentication (2FA) like Google Authenticator to drastically reduce the likelihood of a successful API attack. SMS forms of 2FA can be used in a pinch, but it’s not very secure as it’s easy to bypass. As an added benefit, proper authentication and authorization will keep your FinTech business safe from identity thieves looking to gain access using social engineering. 3. Keep Your Data Separate & Safe Many applications make it easy for hackers to find and steal data because they keep everything together in one extensive database rather than segregating sensitive data into different entities. That means a hacker only has to compromise one database to get access to all of your data. It's a lot easier to secure a system if you can compartmentalize different parts – for example, if you have an account management app and a banking app, make sure they're entirely separate. Restricting access to sensitive information is one of the best things you can do to protect your API from being hacked, which entails creating security groups with limited permissions on what they are allowed to see based on their job functions. Lastly, whenever possible, make sure only trusted APIs can access your APIs so that you know who's behind API requests and to make it easy to flag suspicious activity for an audit. Read More: Drilling Down into Excessive Data Exposure: How to Protect Your API's Sensitive Data 4. Enforce the Secure Socket Layer (SSL) Encryption Security Protocol SSL communication adds another layer of security by ensuring no unauthorized third parties can read communication - even if it's intercepted. Your customer will first connect to a server controlled by you, which uses a certificate to synchronize a secret key with your customer's browser. This means that all communication is encrypted, and no one can hijack data in transit. There's also HTTPS encryption, which makes sure that any information sent between your site and customers is encrypted as well as authenticated using SSL certificates. 5. Invest in Cyber Security Training for Employees Employees should be educated on how to identify an API attack and what steps they can take to prevent one from happening. Ensure that your employees are aware of the dangers of API attacks and how to protect themselves. This could be something as easy as flagging any suspicious activity for a security audit or removing compromised accounts from your system immediately. Another great way to prepare is to implement cybersecurity tabletop exercises that include the most common API scenarios. They should also be aware of the consequences of a successful API attack. This way, they'll know what to do if or when an API attack occurs. 6. Have a Contingency Plan in Place in Case of an Attack But even with all this protection, you still need to remember that no matter how many firewalls and authentication procedures you have in place, API attacks are almost impossible to safeguard yourself against completely. This means it's also important to know what needs to be done when an attack does occur immediately to mitigate the damage caused to your organization. It can be something as simple as notifying your customers that their login credentials have been compromised to help them take proactive steps until it's too late. As FinTech continues to gain popularity, the threat of API attacks will only grow unless businesses take steps like these necessary to keep themselves safe from hackers who want access to their data. Read More: API Security Checklist: What You Need To Know 7 Use API Security Testing Solutions to Ensure Full Coverage API security is a critical issue for the financial services industry. As we have seen from the news, API breaches can have severe consequences for consumers, businesses, and third-party providers. To ensure the safety of your data, it is vital to take steps to protect your APIs from hackers. Using automated API security platforms can make all the difference in protecting your business from data breaches and other cyber threats, allowing your to ensure continuous and comprehensive API security testing around the clock for a fraction of the cost of hiring a cybersecurity specialist. APIsec is one of the best API security testing solutions available on the market. While this claim might be a bit biased (only slightly), here's what separates APIsec from everybody else: On top of continuously monitoring your APIs across a massive list of known vulnerabilities, our AI-based platform is the only solution that can automatically write and execute thousands of test cases generated based on the unique architecture of your APIs. If you are in the financial industry, APIsec is your best choice for API security. For more information on the benefits of using APIsec to secure your APIs, reach out to our team today for a free vulnerability assessment.
<urn:uuid:0e2353c2-1d27-4467-9e8c-3f74cb49478c>
CC-MAIN-2024-38
https://www.apisec.ai/blog/fintech-api-security
2024-09-07T13:15:01Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650883.10/warc/CC-MAIN-20240907131200-20240907161200-00025.warc.gz
en
0.932004
2,778
2.75
3
It is generally known that fiber optic patch cable can be classified by transmission medium (single mode or multimode), by connector construction (FC, LC etc) and by fiber cable structure (simplex or duplex). This paper will introduce simplex fiber patch cable and duplex fiber patch cable. In order to have a better understanding of simplex and duplex fiber optic patch cable, the definition of simplex and duplex will be explained in the first part. Simplex communication is a communication channel that sends information in one direction only (see Figure 1). For instance, in TV and radio broadcasting, information flows only from the transmitter site to multiple receivers. Duplex communication system is a point-to-point system composed of two connected parties or devices that can communicate with each other in both directions. It has two clearly defined paths and each path could carry information in only one direction (A to B over one path, and B to A over the other). Duplex communication system can be classified into half-duplex and full duplex. In a half-duplex system, there are still two clearly defined paths/channels, and each party can communicate with the other but not simultaneously (see Figure 2). Typically, once a party begins receiving a signal, it must wait for the transmitter to stop transmitting, before replying. A walkie-talkie is a half-duplex communication system. In a full duplex system, both parties can communicate with each other simultaneously (see Figure 3), such as a telephone. Simplex fiber optic patch cable consists of a single strand of glass fiber, and is used for applications that only require one-way data transfer. It is commonly used where only one single transmit and receive line is desired. Simplex fiber optic patch cable is available in single mode and multimode. For instance, a single mode simplex fiber patch cable (see Figure 4) is a great option for data travelling in one direction over long distance. Duplex fiber optic patch cable consists of two strand fibers of glass structured in a zipcord arrangement where each fiber strand has independent coatings that are linked together by a thin layer of coating material. Duplex fiber patch cable is most used where separate transmit and receive signals are required, that is, one strand transmits in one direction while the other strand transmits in the opposite direction. It is available in single mode and multimode. Multimode duplex fiber optic patch cable or single mode duplex fiber optic patch cable is usually used for applications that require simultaneous and bi-directional data transfer. For example, 10 gigabit multimode duplex cables can support 10 Gb/s bandwidth in both directions within a short distance. LC to LC duplex single mode fiber patch cable (see Figure 5) can make simultaneous data transfer with LC-LC connectors over long distance. After reading the above statements, do you have a brief understanding of simplex fiber patch cable and duplex fiber patch cable? When choosing one over the other, the key factor is that the equipment requires one-way or bi-directional data transfer. Fiberstore has large numbers of simplex and duplex fiber optic patch cables, such as single mode simplex fiber patch cable, LC to LC duplex single mode patch cable, 10 gigabit multimode duplex cables, LC ST duplex patch cord and so on. I believe you can find a suitable fiber optic patch cable for your devices in Fiberstore.
<urn:uuid:a86aff97-051d-432f-8556-7ba0975f9f56>
CC-MAIN-2024-38
https://www.fiber-optic-components.com/tag/single-mode-simplex-fiber
2024-09-07T15:11:54Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650883.10/warc/CC-MAIN-20240907131200-20240907161200-00025.warc.gz
en
0.913719
707
3.25
3
According to the AIA, “Buildings are responsible for 40% of energy-related CO2 emissions globally.” Seventy-nine percent of architects want to reduce this number by specifying more sustainable materials than they do today, according to the AIA report, Sustainability in the Architect’s Journey to Specification. However, only one in three architects feels they are meeting their responsibility for sustainable design today. With the rapid rate of digital transformation and industry shifts during the last several years, many architecture and engineering firms are not yet where they want to be with sustainable design or digital automation. Yet, these topics will continue to increase in importance as 46% of clients are increasingly concerned about sustainability and 32% are committed to sustainability, according to the AIA report, Sustainability in the Architect’s Journey to Specification. Is your firm weighing the pros and cons of sustainable design? Read on to discover four steps to start your sustainability journey. AIA Report: Sustainability in the Architect’s Journey to Specification Discover Key Learnings 1. Utilize Sustainable Design as a Catalyst to Fuel Your Firm’s Growth Sustainability presents a significant opportunity for many firms to grow their business. According to Leadership in Energy and Environmental Design (LEED), 61% of corporate leaders believe that sustainability leads to market differentiation and improved financial performance. And, owners of sustainably designed buildings report a 10% or more increase in the value of their assets. In addition, a growing number of architects recognize the advantage of winning more work through smart building design. According to the 2021 AEC Hinge Report, nearly half of architects surveyed cited Smart Buildings as a topic they plan on researching in the coming year. Smart buildings use automation to optimize processes that occur inside a building such as heating and cooling, security, lighting, ventilation, water usage and more. This benefits the environment, the building tenants, and the businesses that own the facilities. 2. Consider that Sustainable Design is Becoming More Affordable The biggest barrier towards moving more projects to sustainable design is cost, as 59% of architects say sustainable products are too expensive, according to the AIA report, Sustainability in the Architect’s Journey to Specification. However, the cost of sustainable building materials and products has started to – and continues to drop – making green building a cost-effective solution that also supports the environment. 3. Explore the Long-Term Benefits of Sustainable Design Sustainable design is a positive contributor to overall employee, student and patient health and happiness. Consider that 70% of K-12 schools and 63% of university leaders said that green design efforts have raised student test scores. Patients exposed to bright, natural light report less pain and take 22% less analgesic medication per hour than patients in darker rooms. As the years go by, [and particularly in light of the pandemic], we are starting to see a shift to focus on occupant health. Because people spend so much of their time in buildings,” says Camp Boyd, Senior Specification Writer, Deltek. Sustainable buildings can improve indoor air quality, reduce employee absenteeism, and decrease symptoms of asthma, respiratory distress, depression and stress. In fact, employees in LEED certified buildings report feeling healthier, happier and more productive. 4. Discover Best Practices in Sustainable Design If your firm is working toward incorporating more sustainable design practices into upcoming building projects, reach out to the Deltek Specification Solutions team of experts. MasterSpec® content developers work to make sure specifying to sustainable standards and certifications is stress-free and manageable for building projects large and small. Coming late 2021, Deltek Specpoint, home of MasterSpec®, a cloud-based specification solution, will uplevel how product manufacturers, architects and engineers collaborate. Deltek Specpoint will help bring teams together to work more efficiently, make data driven product decisions for smart specification authoring, and improve project outcomes by easily sharing, updating and distributing MasterSpec content, better supporting dispersed teams. This cloud-based software solution includes a comprehensive online building products library to research and select products, combined with a design and publishing tool that includes one-click specifications so that firms will be able to write better specifications and produce project manuals faster, contributing to overall project success. Learn more about this new solution, or see a live demo, at the Deltek Booth at A’21. Deltek Project Nation Newsletter Subscribe to receive the latest news and best practices across a range of relevant topics and industries.
<urn:uuid:e76b6a1b-e2da-427a-83ae-6a6bdd83b3e5>
CC-MAIN-2024-38
https://www.deltek.com/en/blog/steps-to-sustainable-design
2024-09-08T15:53:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651013.49/warc/CC-MAIN-20240908150334-20240908180334-00825.warc.gz
en
0.943624
932
2.609375
3
The real 5G mobile internet connectivity race is to secure not only the network but also the ecosystem of devices and applications attached to the network, former Federal Communications Commission (FCC) Chairman Tom Wheeler said in a new paper published by the Brookings Institute. In the report, entitled Why 5G Requires New Approaches to Cybersecurity, co-authors Wheeler and David Simpson, former chief of the FCC’s Public Safety and Homeland Security Bureau, contend that securing 5G networks is central to national security. “To build 5G on top of a weak cybersecurity foundation is to build on sand,” the authors wrote. In a nod to the bigger picture, Wheeler and Simpson make a case that failing to move beyond the Huawei spying issue masks some larger issues concerning 5G security. “Policy leaders should be conducting a more balanced risk assessment, with a broader focus on vulnerabilities, threat probabilities, and impact drivers of the cyber risk equation” rather than becoming so immersed in Huawei. “China is a threat even when there is not Huawei equipment in our networks,” they said. According to Wheeler and Simpson, with the advent of 5G technology comes these five cyber dangers: - No hardware choke points. Previous networks were hub-and-spoke designs in which everything came to hardware choke points where cyber hygiene could be practiced. In the 5G software defined network, there is no chokepoint inspection and control. - Virtualization. Virtualizing in software the higher-level network functions formerly performed by physical appliances complicates 5G cyber vulnerability. - Software. Often early generation artificial intelligence itself can be vulnerable. An attacker that gains control of the software managing the networks can also control the network. - New attack vectors. Physically, low-cost, short range, small-cell antennas deployed throughout urban areas become new hard targets. - IoT. Vulnerabilities created by attaching tens of billions of hackable, smart devices attached to the network. “We shouldn’t be surprised that the networks of the 21st century are the new attack vectors, but these are different because they are expanded to an almost infinite number of attack vectors,” Wheeler told The Hill. The authors acknowledge that what needs to be done to secure 5G networks is “both important and not without cost.” They argue, however, that these are not normal times and require a “departure from traditional practices.” This “new reality” justifies the following corporate and governmental actions: Key #1: Companies must recognize and be held responsible for a new cyber duty of care. This spans: - Reversing chronic under-investment in cyber risk reduction. - Implementing machine learning and artificial intelligence protection. - Shifting from post attack cyber preparedness to using leading indicators. - Cybersecurity starts with the 5G networks themselves. - Inserting security into the development and operations cycle. - Congress should establish a cybersecurity standard of expected performance and accompanying incentives for its adoption by companies. Key #2: Government must establish a new cyber regulatory paradigm to reflect the new realities. This spans: - More effective regulatory cyber relationships with those regulated. - Recognition of marketplace shortcomings. - Consumer transparency. - Inspection and certification of connected devices. - Contracts aren’t enough. - Stimulate closure of 5G supply chain gaps. - Re-engage with international bodies. In concluding, the authors chided the Trump administration and Congress for sitting on their hands on 5G cybersecurity. “Congress should not have to pass legislation instructing the Trump administration to act on 5G cybersecurity,” they wrote. “The whole-of-the-nation peril requires a whole-of-the-economy and whole-of-the-government response built around the realities of the information age, not formulaic laissez faire political philosophy or the structures of the industrial age.”
<urn:uuid:641fee8f-fa54-4dfd-99c6-0481cbbb2291>
CC-MAIN-2024-38
https://www.msspalert.com/news/5g-network-security-standards
2024-09-08T17:16:02Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651013.49/warc/CC-MAIN-20240908150334-20240908180334-00825.warc.gz
en
0.941433
820
2.59375
3
Subscribe to Our Newsletter Introduction: This article explores the definition and types of identity theft, how it occurs, and its consequences for individuals and businesses. It also discusses prevention and protection measures to help safeguard against identity theft. Definition and types of identity theft How identity theft occurs Consequences of identity theft for individuals Consequences of identity theft for businesses Prevention and protection measures. 1. Definition and types of identity theft Identity theft is a crime in which an individual’s personal information is stolen and used without authorization for fraudulent purposes. It can have far-reaching consequences for both the victim and the entities involved. There are several types of identity theft, each with distinct characteristics and methods. Financial identity theft: This is the most common form of identity theft. In financial identity theft, a criminal uses the victim’s personal information, such as Social Security numbers or bank account details, to gain access to their financial resources. This could involve opening new credit card accounts, taking out loans, or making unauthorized purchases in the victim’s name. Criminal identity theft: In this type of identity theft, the perpetrator uses the victim’s personal information to commit crimes or evade law enforcement. For example, a criminal might provide the victim’s details during an arrest, leading to the victim being falsely linked to a crime they did not commit. Medical identity theft: Medical identity theft occurs when someone uses another person’s name, insurance information, or other personal details to fraudulently obtain medical services or prescription drugs. This can lead to false medical records, financial losses, and even life-threatening situations for the victim if incorrect information is used in their treatment. Child identity theft: Child identity theft involves the unauthorized use of a minor’s personal information. Since children typically do not have credit histories, their information can be particularly attractive to identity thieves. Criminals may use a child’s Social Security number to open accounts, obtain loans, or commit other fraudulent activities. Understanding these different types of identity theft is crucial for identifying potential risks and adopting appropriate prevention strategies. 2. How identity theft occurs Identity theft can occur in various ways, with perpetrators employing both low-tech and high-tech methods to obtain personal information. Some common methods used by identity thieves include: Phishing scams: Phishing is a deceptive technique where criminals send emails or text messages purporting to be from a legitimate organization, such as a bank or government agency, to trick individuals into revealing sensitive information like usernames, passwords, or credit card details. Data breaches: Data breaches occur when unauthorized individuals gain access to sensitive data stored by companies or organizations. This could involve hacking into computer systems, exploiting vulnerabilities in software, or stealing physical storage devices containing personal information. Physical theft: Identity thieves may steal personal documents, such as mail, wallets, or purses, to obtain an individual’s personal information. They can also rummage through trash bins to find discarded documents containing sensitive data, a practice known as “dumpster diving.” Skimming: Skimming is a technique where criminals use a small electronic device called a skimmer to steal credit or debit card information during a legitimate transaction, such as at an ATM or point-of-sale terminal. Social engineering: Social engineering involves manipulating individuals into divulging personal information or granting access to sensitive data. This could involve impersonating a trusted individual, such as a coworker or customer service representative, to deceive the victim into revealing their information. Malware: Malware, short for malicious software, is a type of software designed to infiltrate, damage, or steal information from computer systems. Cybercriminals may use malware to gain unauthorized access to an individual’s personal data stored on their computer or other devices. By understanding the various methods used by identity thieves, individuals and organizations can take appropriate steps to protect their personal information and reduce the risk of identity theft. 3. Consequences of identity theft for individuals Identity theft can have severe and long-lasting consequences for individuals. Some of the most common impacts of identity theft on victims include: Financial loss: One of the most immediate consequences of identity theft is financial loss. Victims may find their bank accounts drained, unauthorized loans taken out in their name, or fraudulent charges made on their credit cards. In some cases, it can take years for victims to recover their stolen funds. Damaged credit: Identity thieves often use stolen information to open new lines of credit or make purchases in the victim’s name. This can result in significant damage to the victim’s credit score, making it difficult for them to obtain loans, mortgages, or even employment. Legal issues: In cases where the perpetrator commits crimes using the victim’s identity, the victim may find themselves facing legal issues or even criminal charges. Clearing one’s name in such circumstances can be a lengthy and complicated process. Emotional distress: Victims of identity theft often experience significant emotional distress, including feelings of vulnerability, anger, and frustration. The process of resolving the issues caused by identity theft can be time-consuming and emotionally draining. Loss of privacy: Identity theft can lead to a significant loss of privacy, as the perpetrator gains access to the victim’s personal and financial information. This can leave the victim feeling violated and vulnerable. Difficulty obtaining services: With a damaged credit history and potential legal issues, victims of identity theft may have difficulty obtaining essential services, such as utilities, housing, or even medical care. Understanding the consequences of identity theft is essential for appreciating the importance of taking preventive measures to protect one’s personal information. 4. Consequences of identity theft for businesses Identity theft not only affects individuals but can also have significant consequences for businesses. Some of the most common impacts of identity theft on businesses include: Financial losses: Businesses can suffer considerable financial losses due to identity theft, particularly if they are held liable for fraudulent transactions or if they must compensate customers for their losses. Additionally, businesses may need to invest in enhanced security measures and fraud detection systems to prevent future incidents. Reputational damage: When a business falls victim to a data breach or is associated with identity theft, its reputation can suffer greatly. Customers may lose trust in the company’s ability to protect their personal information, leading to a decrease in customer loyalty and potential loss of business. Legal liabilities: Businesses that fail to adequately protect customer information may face legal liabilities, including fines, penalties, or even lawsuits. Companies may also need to comply with various regulatory requirements related to data protection, which can be costly and time-consuming. Loss of productivity: Dealing with the aftermath of an identity theft incident can be a resource-intensive process for businesses, involving investigations, legal proceedings, and customer support. This can lead to a considerable loss of productivity as employees are diverted from their regular duties to address the issue. Increased operational costs: Businesses may need to invest in improved security measures, employee training, and fraud detection systems following an identity theft incident. These increased operational costs can have a significant impact on a company’s bottom line. By understanding the consequences of identity theft for businesses, companies can better appreciate the importance of implementing robust security measures and policies to protect their customers’ personal information. 5. Prevention and protection measures To safeguard against identity theft, both individuals and businesses should implement strong security measures and adopt best practices. Some essential prevention and protection measures include: Use strong passwords: Create unique, complex passwords for all online accounts, and update them regularly. Avoid using easily guessed information, such as names or birthdates, and consider using a password manager to help manage multiple passwords securely. Monitor financial accounts: Regularly review bank statements, credit card statements, and credit reports for any suspicious activity or unauthorized transactions. Be cautious with personal information: Limit the amount of personal information shared online and on social media. Be wary of unsolicited requests for personal information and verify the legitimacy of any organization or individual before providing sensitive data. Secure personal documents: Keep important documents, such as passports, Social Security cards, and financial statements, in a secure location. Shred any unnecessary documents containing personal information before discarding them. Protect electronic devices: Ensure that computers, smartphones, and other devices are secured with up-to-date antivirus software, firewalls, and strong passwords. Regularly update software and operating systems to patch any security vulnerabilities. Be vigilant against phishing scams: Be cautious when opening emails or clicking on links from unknown sources. Verify the legitimacy of any email or text message that requests personal information or prompts to click on a link. Educate employees: For businesses, it’s essential to provide regular employee training on data security practices, recognizing phishing scams, and safe handling of sensitive information. Implement strong security policies: Businesses should establish robust data protection policies and procedures, including access controls, encryption, and secure data disposal methods. By implementing these prevention and protection measures, individuals and businesses can significantly reduce the risk of falling victim to identity theft. Glossary of terms: - Identity theft: Unauthorized use of someone’s personal information to commit fraud or other criminal activities. - Phishing scam: A fraudulent attempt to obtain sensitive information by disguising oneself as a trustworthy entity in electronic communication. - Data breach: An incident where unauthorized individuals gain access to sensitive and confidential data. In conclusion, understanding the far-reaching impact of identity theft is crucial for both individuals and businesses. By being aware of the risks and implementing preventive measures, we can work towards reducing the likelihood of falling victim to this pervasive crime. Guarding Your Tomorrow, Today: Unveiling the Far-Reaching Impact of Identity Theft and Empowering You to Take Control. Connect with us today to unlock tailored solutions that empower your business for the digital era. Elevate your business prowess with GXA leading the charge towards technological transformation and success.
<urn:uuid:dc954184-5556-4a75-816f-eaf470384e2e>
CC-MAIN-2024-38
https://gxait.com/blog/understanding-the-far-reaching-impact-of-identity-theft/
2024-09-11T03:50:25Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651343.80/warc/CC-MAIN-20240911020451-20240911050451-00625.warc.gz
en
0.923536
2,068
3.0625
3
Cybercriminals have been taking advantage of our anxiety and insecurities to deceive, defraud, and dupe from the early days of the pandemic to the ongoing Russia-Ukraine conflict. In Singapore for instance, we saw scammers building false narratives around vaccinations amidst the vaccine rollout last year to trick consumers to click on dubious links and visit fake pages that harvest personal data. Humans are the weakest link inside organisations. Social engineering attacks have been part of the cybercrime playbook for a long time. Cybercrime accounts for almost half of all crime in Singapore today, and online scams have also grown exponentially. Recent incidents such as the SMS phishing scams targeting bank customers in Singapore only prove that fraudsters will exploit any situation to deceive and steal. The following are some of the ways we’ve observed cybercriminals take advantage of compassionate people to harvest money and credentials. No trap for these RATs Malspam campaigns were one of the first campaigns to emerge that leveraged the crisis as clickbait. Disguised behind urgent emails on supply chain issues, the campaigns lured targets into downloading the malware-as-a-service remote access trojan (RAT) Agent Tesla. RATs are insidious as they are capable of more than stealing and altering data. Agent Tesla can even hijack a device’s core functions once it has been compromised. Most importantly, since victims are unaware of its presence, locating and removing it can be difficult – even with antivirus software. Agent Tesla in particular is known to mutate, making it even more of a danger to organisations both large and small. While these campaigns exploit our human weaknesses, other social engineering attacks potentially leverage fear or curiosity. Whether on work or personal devices, individuals should be wary of opening unfamiliar or unsolicited emails, and inspect unusual attachments before downloading and opening them. At the corporate level, because over 90% of malware must touch DNS to enter or leave a network, using DNS security can help security pros accelerate threat hunting. Cryptic Ukrainian support When the crisis erupted in late February, a wave of positive public sentiment turned towards Ukraine. As such, a flurry of sites emerged offering donations to Ukraine, and even the Ukrainian government requested donations in cryptocurrency on Twitter. Seeing this wave of support and concern, cybercriminals immediately pivoted to create fraudulent support campaigns. These social-engineering campaigns exploit the ongoing crisis and concern for personal gain, siphoning well-meaning donations into their own pockets. Further complicating things are the emergence of decentralised anonymous organisations (DAOs), which leverage a blockchain for transparency and record-keeping. Many such DAOs are legitimate, like UkraineDAO (LOVE) which raised over US$8 million in a month for Ukraine. However, ongoing domain analysis has uncovered fake DAOs that look remarkably like valid relief campaign organisations. These DAOs lack credible ties and are scams to pilfer cryptocurrency, benefiting from the anonymous and decentralised nature of cryptocurrency donations. These campaigns highlight how difficult it can be for the average consumer to distinguish between legitimate and malicious activity. Cybercriminals not only abscond with the donations, but can also steal personal information and credit card details, or even deliver malware. Prevention is better than cure Cybercriminals will always adapt and adjust their tactics to be a step ahead of precautions. Therefore, a healthy suspicion is always necessary when receiving unexpected mail, text messages, or when surfing the net. Organisations should also step up awareness training for employees to keep their networks, browsers, and devices malware-free. When it comes to donating, individuals should think twice before sharing sensitive payment information. Be on guard for potentially fraudulent payment services and redirects to unknown third-party websites. The volatility of today’s geopolitics and the long-drawn pandemic have exposed vulnerabilities that cybercriminals are eager to exploit. Practising good cyber hygiene will be key to mitigating the losses in this modern cat-and-mouse game.
<urn:uuid:59472a3e-78c1-43f5-acaf-989ca83607bb>
CC-MAIN-2024-38
https://www.frontier-enterprise.com/exploiting-the-ukraine-crisis-donation-scams-and-malspam/
2024-09-11T03:54:16Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651343.80/warc/CC-MAIN-20240911020451-20240911050451-00625.warc.gz
en
0.928766
817
2.546875
3
What is a Port? In reference to computer hardware, a port is an outlet into which devices can be plugged, connecting the computer to peripherals such as keyboards, scanners and mice. The connection that happens between the port and device (usually done through a cable) allows signals to be sent back and forth, such as a mouse click or a print job. There are several different types of ports, characterized by the shape and size of the connection point and the way they transmit data. Serial ports consist of one pair of wires, send and receive data a single bit at a time, and are commonly used to connect mice to computers. Parallel ports have many sets of wires, simultaneously send multiple bits, and connect computers to printers. USB ports are very common now and connect devices like keyboards, mice and storage devices. Ports make computers much more useful by enabling their interaction with many other devices. How NICE is Redefining Customer Experience NICE CXone is the industry’s only interaction-centric platform where channels, data, applications, and knowledge converge to improve customer experience at scale. It is the leading, most complete and unified CX Platform on the market, used by thousands of organizations of all sizes around the world to help them consistently deliver exceptional customer experiences. CXone is a cloud native, unified suite of applications designed to help you holistically run your call (or contact) center operations.
<urn:uuid:5158ceef-c9d2-40a2-a02d-14fc2b73af96>
CC-MAIN-2024-38
https://www.nice.com/glossary/what-is-contact-center-port
2024-09-11T04:36:36Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651343.80/warc/CC-MAIN-20240911020451-20240911050451-00625.warc.gz
en
0.954376
291
2.578125
3
Business decisions should never be made based on instincts alone. LLM for data analytics is an ally for companies that operate in the information-buzz era. Thanks to this process enhanced with AI, they can always make informed choices based on facts, statistics, and insights. Since the volume and complexity of datasets continue to skyrocket, organizations want to introduce innovative solutions to extract meaningful information from numerous sources. In this article, we will explain how large language models became a game-changer in this area, revolutionizing the way businesses process data and extract actionable intelligence. Data analysis explained Data analytics is the term that describes the process of examining vast datasets to discover patterns, tendencies, correlations, and other valuable information hidden inside them. The main purpose of this science is to present certain trends and verities about the world and support businesses or individuals in making better decisions. Since humanity produces approximately 328.77 million terabytes daily, it’s crucial to collect, prepare, analyze, visualize, and interpret data to determine the most interesting and relevant bits. The development of the internet accelerated this field, and the rise of artificial intelligence sped it up even more. The ability to use data analytics effectively will be crucial for businesses to thrive in the digital age. By leveraging statistical techniques, mathematical methodologies, and all types of algorithms (including machine learning and deep learning), organizations can learn about the effectiveness of their actions, monitor customer satisfaction, and understand what drives their revenues. This way, companies can become empowered to curate more people-oriented solutions, become leaders in their field, and quickly react to evolving market dynamics. Moreover, data analytics provides knowledge that helps reduce the negative impact of certain activities and eliminate risks while improving operational efficiency and increasing productivity. Overall, it’s a necessary tool for brands that want to offer personalization on an exceptional level, grow their audience, drive profit, and improve internal affairs. Components of data analytics process What strategies do data analysts follow to achieve expected goals? Here are the key components of the process: - Data collection: Gathering information from relevant sources is the first phase. Places where valuable data can be found include external and company-owned databases, spreadsheets, IoT devices, social media, online news portals, other types of media coverage, and so much more. Data is usually classified as structured, semi-structured, or unstructured, and comes in different formats such as text, numbers, images, audio, and video. - Data preparation: To achieve the most consistent and high-quality results, it’s required to clean and organize collected data. Then, it has to be transformed into a form suitable for the upcoming steps. At this point, duplicates are removed, missing values are handled, and all data formats are standardized. - Data analysis: Now, the data analytics specialists incorporate the most suitable methods to perform the analysis. For example, they can use descriptive methods to summarize the traits of data or diagnostic frameworks to identify similarities and anomalies. Predictive analytics are great for forecasting, and prescriptive analytics are used to recommend data-driven approaches. - Data visualization: Presenting results in a digestible way is crucial in the whole process. That’s why analysts prepare visual presentations with charts, graphs, dashboards, and conclusions. Stakeholders can understand better why some things happen and what direction to follow next. - Data interpretation: Finally, all the insights gained during data analysis should be embedded within the context of the business objectives and needs. Analysts must communicate their findings clearly and concisely, highlighting key takeaways and recommendations that can inform strategic decision-making. Understanding large language models (LLMs) For years, people analyzed data manually. Experts were digging through tons of information, looking for patterns or interesting divergences. The invention of the computer was the first step towards faster work and automation. AI development became a true game-changer. Especially large language models proved to be particularly useful in the field. The invention of LLMs (including ChatGPT large language model) represents a significant milestone in AI advancement. They can be used in versatile tasks, including but not limited to data analytics, and adapted to specific requirements whenever the project evolves. Processing data with large language models proves to be particularly useful for businesses seeking to extract insights from textual datasets at scale. How to build an LLM for data analytics Large language model development involves several phases. Each of them is essential to achieve optimal performance and the ability to understand real-world scenarios that are presented within the data. Data acquisition and preprocessing LLM analytics tools should be fed with the largest possible amounts of data of exceptional quality. This levels up their learning process. Texts can be extracted from diverse sources like media outlets, books, blog posts, websites, and social media channels. When the collection of various information sets is ready, it goes through preprocessing that includes cleaning, tokenization, and formatting. This last step is quite important because the text has to be suitable for training the model. To train the large language model, powerful computing resources are required. Sophisticated algorithms use various techniques like self-attention mechanisms or transformer architectures to capture long-range dependencies and semantic relationships within the text. Training is not enough to achieve extraordinary results – that’s why fine-tuning is performed. Additional data that is relevant to the goals of the data analysis process is provided to enhance LLM’s capabilities for a specific task. Thanks to this stage, businesses can adjust the model to their unique objectives and needs. Usually, thanks to fine-tuning, they can set a direction for the project and decide if the LLM should focus on sentiment analysis, text summarization, translation, or something else. When the large language model is ready and aligned with assumed goals, it can be integrated into the data analytics workflow. This phase involves the implementation of LLM into the existing infrastructure, ensuring compatibility with data sources used by the company. The LLM data science partner responsible for preparing the model itself also has to establish protocols for testing, updates, and maintenance. It’s recommended to continuously monitor and evaluate the performance of the model to establish if it meets the expectations and detect areas for refinement. This way, businesses can maintain efficiency over time. Business applications of LLMs in data analytics Large language models’ use cases showcase how versatile data analytics LLMs can be. The wide range of examples where they can be incorporated includes: LLM sentiment analysis LLM data analytics is exceptional when it comes to understanding the context and sentiment of textual data. That’s why businesses can utilize them for customer feedback analysis and brand reputation monitoring. This way, organizations can gain valuable insights that will help them respond better to new customer preferences and emerging market trends. Additionally, thanks to LLM for sentiment analysis, they can design brand perception better and react to potential reputational hiccups at speed. Customer support automation Companies that want to bet on support systems for building closer relationships with their customers should take advantage of large language models’ capabilities to understand unorganized texts. They can analyze relationships between various sources and detect customer reactions or emotions based on their inquiries. AI chatbots that answer questions and email systems for automated support are perfect examples of data analysis that can be used to improve customer satisfaction. AI-powered models are also great when it comes to personalized assistance. Text generation and summarization Content generation and document summarizing were never as easy. With the help of LLMs, employees can generate human-like messages which helps them quickly reply to emails, stay on track with marketing strategies, and be consistent in communication. Automated report writing and the ability to condense complex files into more comprehensible pieces of information also speed up daily work and free time for more creative and strategic tasks. Being aware of past patterns means the ability to predict future trends, anticipate customer behaviors, and be one step ahead of the competition. Large language model data analytics can be used as intel in the information-loaded digital world of news, industry reports, social media entries, and other data sources. The competitive analysis becomes more insightful and less time-consuming with models that are trained to monitor the market, the biggest industry players, and consumer preferences. Information extracted from these activities can be used to make most business growth-related decisions. Fraud detection and risk management By analyzing textual data such as transaction records, applications, inquiries, customer messages, and online interactions, LLMs can assist in detecting fraudulent activities and successfully mitigate potential dangers. Their data analysis skills prove to be useful in detecting suspicious patterns and flagging risks in real time, leading to minimized financial losses and improved security. Companies with global aspirations can save tons of time thanks to LLM for analytics. They can translate location-specific content and support communication with international audiences. All that in less time than professional translators. Of course, for legal purposes, it’s important to work with specialists. Still, for day-to-day usage, well-trained models can serve with their translation abilities to expand reach, improve cross-cultural endeavors, and enhance user experience in products destined for multiple markets. Advantages of large language models for data analytics Data analytics workflows powered with LLMs offer several benefits that should not be overlooked. What are the main pros of using them? - Scalability: LLMs take data analytics to the next level by providing the ability to process large datasets rapidly. Organizations that want to harness big data efficiently should invest in such solutions to maximize results. - Accuracy: Precision offered by large language models comes from advanced deep learning algorithms. They can achieve spectacular effects while keeping the highest level of performance. This allows companies to confidently use the extracted insights. - Cost-efficiency: Compared to traditional data analysis methodologies, LLMs prove to be less budget-consuming in the long run. The initial investment into model preparations and training can be higher, but it can automate repetitive tasks later on, reducing the need for manual intervention and streamlining the process. - Adaptability: Large language models’ applications can be adjusted to concrete business domains, which makes them universal to a variety of industries and requirements. They can help solve various challenges no matter what type of business they are integrated with and what kind of task they are asked to perform. - Real-time insights: The ability to quickly react to new market dynamics is crucial in today’s world. Ordinary data processing can’t compare to the LLM-powered analytics in that area. Companies that bet on them can make faster decisions and shorten time-to-market when responding to emerging customer demands. - Innovation: Organizations can differentiate themselves within their respective niches by using LLMs in their data analytics efforts. This gives them a competitive edge when developing products and services, optimizing internal processes, or discovering trend-driven opportunities. Ethical considerations for using LLMs in data analytics Like with all AI-driven solutions, LLMs are also discussed in terms of ethics and potential dangers they could be responsible for not only in data analytics but in general. To understand these considerations and how to address them, it’s important to dive deeper into the complexities and potential implications of using large language models for business. One of the main problems mentioned by skeptics is data security and privacy. LLM for data analytics is trained on vast datasets, which often contain sensitive information. Unauthorized access, safety breaches, and misuse of data are listed to be the most burning challenges to address. Brands must implement robust security measures to ensure data protection. These can include but are not limited to sophisticated encryption, access control, anonymization methodologies, and compliance with legal requirements. Another critical issue to keep in mind is potential bias in training data, which can lead to discriminatory outcomes. LLMs that are not trained with caution can perpetuate existing gender, racial, or socioeconomic biases, prolonging the negative impact on marginalized groups. Businesses should avoid that risk by carefully evaluating their training data and implementing fairness-aware algorithms. Conducting regular audits and tweaking the models, so they can produce unbiased results, is also necessary. Last but not least, there’s a question of transparency. Should institutions that use AI in data analytics be open about it to keep their accountability? Many adversaries mention the “black box” dilemma, which implies that we’re not entirely sure how artificial intelligence models make decisions and produce results. That’s why a lack of honesty can undermine trust among customers. Organizations should inform that LLM analytics is a part of their process and prepare documentation on architectures, training frameworks, and the usage of insights provided by AI. This way they can always explain the hows and whys of their internal workflows to consumers, stakeholders, auditors, and legal entities. Additionally, this helps them avoid or eliminate potential risks. The amount of data in today’s digitized and globalized world, as well as its complicated structures, requires the implementation of modern solutions. Large language models revolutionized the data analytics sector, helping brands achieve their goals thanks to the processing and drawing conclusions from large amounts of textual data. Businesses that want to skyrocket their decision-making and grow should definitely consider investing in LLM data analytics.
<urn:uuid:04f5a2fc-a2b2-4906-ae95-ce8cbc848626>
CC-MAIN-2024-38
https://indatalabs.com/blog/llm-for-data-analytics
2024-09-14T20:48:13Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651580.74/warc/CC-MAIN-20240914193334-20240914223334-00325.warc.gz
en
0.934862
2,730
2.84375
3
Blockchain’s most significant benefit is that no authority has control over it. The advent of IoT (Internet of Things) has sped up its decentralized ledgers for conducting financial transactions. One of the critical challenges of the IoT world is data security, where Blockchain offers its benefits. By Harmanpreet Kaur, Security Consultant at EY Blockchain uses immutable records for storing sensitive information, and every record is distributed in the form of tamper-proof nodes. Strong cryptographic encryption and hashes are used to protect the information, and if a block must be accessed, all the previous blocks need to be validated and edited. Due to the ‘chain of blocks’ and the nature of accessing data sequentially, Blockchain networks cannot be hacked by cyber adversaries. Blockchain is used with IoT networks to create secure “mesh networks” that are not prone to vulnerabilities such as spoofing and impersonation. Companies looking to add transparency and trust to their products/services are leveraging Blockchain with IoT, and there are roughly 9 billion smart devices online today. IoT networks send huge volumes of data between multiple devices, and Blockchain uses decentralized peer-to-peer ledgers to transmit it. Only authorized users can approve changes/edits made to nodes over networks, and third parties or hackers cannot make edits to these ledgers nor access them unless all other users in these networks approve them. This is what makes Blockchain so powerful, and when it is integrated with IoT, transactions or the transmission of billions of data essentially becomes completely secure. How do Blockchain and IoT complement each other? IoT networks have an administrator who acts as the central authority for controlling and monitoring processes over these networks. Hackers know that the administrator has access to data and is the main vulnerability when it comes to managing IoT devices. Social engineering strategies are commonly employed to fool the administrator in leaking login credentials, and sometimes hackers can use brute force methods to crack administrator accounts. Blockchain can complement IoT by putting data into blocks and creating immutable chains that outsiders cannot alter. Users can enjoy additional security by selecting data and customizing access permissions, thus making it convenient to share among clients and partners without compromising on security. Decentralized ledgers are designed to be interoperable and integrated with multi-cloud environments, and Blockchain can streamline business processes by drawing on data shared by sensors and IoT devices. Freight transportation is an emerging field where this merger technology is being applied. IoT-enabled blockchain networks are being used to track shipment status, move goods from locations, and garner trust in sharing data amongst parties. Compliance audits and component tracking in aircraft cargo containers are other areas where Blockchain networks and IoT are used. It’s helping manufacturers cut down on costs and ensure goods follow the standards laid down by regulatory authorities when transporting them via air. IoT is used in critical machine maintenance, and operators use Blockchain networks to check what industrial processes require preventive maintenance by getting automatic alerts and notifications and recording their work using ledgers. Operational records can be shared with the government to verify the state of repairs done and ensure legal compliance. Blockchain systems safeguard third parties from a single point of failures experienced by IoT devices, and decentralized ledgers can be used for optimizing computational processing efficiency. Blockchain and IoT combined also address the issue of high concurrency rates by using cryptographic security protocols and can improve bandwidth limitations faced from streaming huge volumes of data continuously across IoT networks, sensors, and devices. According to a study by Gartner, it is estimated that Blockchain would add $3.1 trillion in business value by 2030. The global IoT market is expected to grow from $157B in 2016 to $457B by 2020. Though IoT has various advantages, it has its drawbacks when it comes to authentication standards. Authentication is crucial concerning security aspects. It is necessary that the data stored and transmitted are safe, secure, and protected from external attacks such as hacking of data, unauthenticated access, etc. With the integration of Blockchain and IoT technology, authentication issues and privacy concerns are tackled. Every participant in Blockchain networks has a copy of the data. The secure nature of Blockchain nodes assures that transactions cannot be modified, thus promoting security and letting them be available for viewing by the public. Benefits of Blockchain and IoT Combined Currently, the IoT ecosystem uses a client-server architecture that is non-distributed, and IoT models are not scalable, which puts severe limitations when it comes to meeting the demand of global consumers. Blockchain technology can be used to track all the sensor data measurements, preventing any duplications of other malicious data. The implementation of IoT devices is usually intricate, and security concerns can be easily solved by taking advantage of IoT device authentication, identification, and cryptographic encryption features. IoT sensors can be exchanged via Blockchains to reduce threats with third-party services, and Blockchain can secure the IoT devices from being altered. Smart contracts are now possible with blockchains, and agreements are executed automatically as soon as predefined conditions are met, thus involving no human intervention or third-party interferences. Blockchain integrations with IoT are proving to provide significant benefits to the telecom industry. Given that IoT devices are connected to a centralized server, the security procedures to authorize and authenticate transactions are expected to be slower with the increase in IoT devices. This is overcome by the Blockchain’s Self-sovereign identity (SSI), which acts as a cover for the IoT devices. The SSI possesses key encryption methodologies and promotes low-cost compliance with GDPR, CCPA, etc. Blockchain and IoT for different areas Integration IoT technology on a blockchain platform can provide a high-security system for homes that can all be accessed by a smartphone remotely. Blockchain can enhance the IoT-based devices for a smart home by providing solutions for security problems and removing any intermediary through its decentralized infrastructure. One such company that offers blockchain-based smart home solutions is an Australian telecommunication and media company called Telstra. The company is dedicated to incorporate blockchain technology through biometric security to ensure highly secure smart devices. Blockchain’s immutable and practically un-hackable nature allows the storage of personal data for security purposes of the IoT devices such as biometrics, facial recognition, etc. Blockchain integration with IoT allows only the owner to access the data through their private key. There is a day-by-day increase in the fabrication of pharmaceutical medicines and disrupting the pharmacy industry’s functions, such as developing and distributing drugs. Thereby, blockchain technology is used to track and trace pharmaceutical medication and monitor the shipping process. Mediledger is a blockchain-based IoT use case that helps in monitoring the prescribed medicines and their legal changes. The tracking can be accomplished through intelligent devices that use sensors to store the information on the blockchain network to help process payments and monitor the supply chain process to avoid fabrication. The blockchain network can be shared between the manufacturers, clients, sellers, and dispensers to enable transparency. Ensuring enough produce for the entire population while reducing the carbon footprint is a big challenge in the agriculture sector. The agriculture industry faces problems maintaining the transparency of supply chain management, farmers, and whole sellers. However, with combined solutions of IoT and Blockchain, the industry is set to experience revolutionary changes. One such solution is installing sensors in farms while storing and processing data on blockchain networks to accelerate supply chain management, maintain enhanced security, and ensure no tampering. Pavo, a recent innovation, is determined to bring transparency to the supply chain and collect the information through a sensor-based hardware device that transmits data to Blockchain networks and provides secure storage. Farmers can change their operations according to the analysis of stored data and increase productivity. Pavo even allows farmers to presell their harvest through smart contracts and receive secure payments while waiting for crops to grow and harvest. Blockchain & IoT – Use Cases Use Case 1 – Self-driving cars (IoT) and Blockchain. One of the best examples of Blockchain and IoT mergers is the evolution of self-driving cars connected and operated through IoT networks. Self-driving cars are sustainable and reduce carbon footprints, thereby proving a better alternative in protecting the environment. IoT successfully converts cars into smart devices and uses advanced technologies to operate in real-time, capturing accurate surroundings. IoT-based self-driving cars can generate route information, predict travel time and notify alerts by collecting data from sensors. However, the biggest downside of IoT-based vehicles is the lack of data security in connected cars, which is the primary reason why the automotive sector is looking to combine IoT and blockchain features. The collaboration of IoT and Blockchain frameworks will ensure high-security standards for storing and transmitting data between connected cars and IoT platforms. According to a report, it is estimated by 2025, around 10-15% of transactions on connected vehicles will likely be done using blockchain technology. Blockchain infrastructures can certify adequate data storing and minimizing and protect against data breaches and leaks. Attackers will not tamper with data stored on the blockchain network as encrypted through cryptographic hashes and timestamps. Incorporating blockchain technology with IoT automotive devices will also enhance the acceptance of self-driving cars and become a boon for the automotive market. Use Case 2 – IoT + Blockchain to track vaccines With the world moving on despite the COVID-19 pandemic and its varied traits, one of the most significant breakthroughs was inventing a vaccine effective against the Coronavirus. Since the invention of the vaccine, all countries have exercised vaccination drives and getting as many people vaccinated as possible every day. However, increased demand for vaccines was not being fulfilled due to inadequate stocks and outdated vaccines. It was difficult for medical companies to keep inventory records, supply chain, and distribution data. Indian IT firm Tech Mahindra came up with a solution to track the vaccine supply chains worldwide and became a leading pioneer in this field. Their solution was a blockchain-based open-source platform to maintain supply chain transparency at all times. The vaccine manufacturers use the blockchain-based inventory tracking system and vendor payments through Internet-of-Things (IoT) and smart contracts. Integrating IoT with the blockchain-based tracking application made it easier for vaccine manufacturers to transfer amounts to vendors. To understand and integrate steps of a supply chain system, the IT firm plans to work with vaccine researchers, governments, pharmaceutical companies, distributors, and healthcare workers. Blockchain and IoT-based solutions are expected to ensure the validity and safety of these vaccines. Tech Mahindra aims to implement a Vaccine Ledger globally soon. The traceability solution will also predict and prevent failures in the supply chain, including expired vaccines, fabrication, and inadequate stock. The company is set to develop mobile and web-based applications to support manufacturing and government requirements. Their innovative solution is to build a peer-to-peer network that supports real-time data sharing, traceability, and validation to ensure authenticity and information security. The data on the ledger can be used to trace IoT integration and inventory management and can be integrated easily into the existing systems. The IT tech firm claims it will establish a worldwide supply chain, and those countries can view immediate results when it’s finally deployed. Blockchain and IoT Technology are some of the booming and increasingly used technologies in today’s life. Many people have turned towards these technologies for various uses, such as smart homes, cryptocurrencies, etc. Multiple organizations have now turned towards Blockchain integrated with IoT, which is beneficial and provides a stable function that is secure and ready. Various renowned organizations have supported and started utilizing Blockchain + IoT like the Hyundai supported a start-up project wholly based on blockchain technology, known as the HDAC, and raised nearly 40 million dollars to integrate with the IoT devices produced by the organization (Hyundai). Another organization, Filament, is developing an industrial IoT chip responsible for automatically encoding sensor data while simultaneously adopting Blockchain technology. And the most renowned project, IOTA, showcases Tangle, a blockchain specially created and designed for IoT devices. Various organizations and sectors are now incorporating blockchain technology integrated with IoT technology based on the benefits they offer when integrated, based on the results of available projects. Experts and professionals agree that when the two technologies are combined, it results in an infinite potential that can overcome various drawbacks while still maintaining their available features. About the Author Harmanpreet Kaur is currently work as Security Consultant at EY GDS. She has over 3+ years of experience in Privileged Access Management solution CyberArk where she worked on CyberArk Implementation and Support Projects for various clients. She has completed her CyberArk CDE, Sentry, Defender and Trustee certification and also has experience and knowledge in creating technical and non-technical documents for her clients. Views expressed in this article are personal. The facts, opinions, and language in the article do not reflect the views of CISO MAG and CISO MAG does not assume any responsibility or liability for the same.
<urn:uuid:f945cfb5-e87b-46d6-9bde-7af56c528616>
CC-MAIN-2024-38
https://cisomag.com/how-blockchain-and-iot-is-making-our-future-smarter/
2024-09-16T02:47:59Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651668.29/warc/CC-MAIN-20240916012328-20240916042328-00225.warc.gz
en
0.930911
2,656
3.09375
3
In recent years, the transportation landscape has witnessed a significant shift, especially in the realm of personal mobility for children. E-bikes and motorcycles, once considered the domain of adults, have rapidly gained traction among the younger generation. This surge in popularity is not just a fleeting trend but a reflection of evolving societal norms, technological advancements, and a renewed emphasis on outdoor activities. However, as these vehicles become more commonplace for children, it has also led to a growing debate and confusion about their classification. Are they merely toys, or do they fall into the category of serious transportation? Let’s delve deeper into this topic. The Rise of E-Bikes and Motorcycles for Children - Historical ContextTraditionally, bicycles have been the go-to choice for children’s mobility, offering a sense of freedom and independence. However, with the advent of technology, e-bikes, equipped with an integrated electric motor, have emerged as an attractive alternative. Similarly, scaled-down versions of motorcycles designed specifically for children have also found their niche. - Factors Driving Popularity - Technological Advancements: Modern e-bikes and motorcycles for children come equipped with safety features, adjustable speed settings, and longer battery life. - Environmental Consciousness: E-bikes, being electric, resonate with a growing emphasis on eco-friendly transportation options. - Skill Development: Beyond just fun, these vehicles teach children balance, coordination, and instill a sense of responsibility. - Market StatisticsVehicle TypeSales in 2020Sales in 2022Growth RateE-Bikes1.5 Million2.3 Million53%Kid’s Motorcycles200,000290,00045%Table 1: Sales growth of e-bikes and motorcycles for children over two years. Blurred Lines: E-Bikes vs. Motorcycles - Definitional DilemmaThe primary distinction between e-bikes and motorcycles lies in their operation. E-bikes assist the rider’s pedal effort with an electric motor, while motorcycles are entirely motor-driven. However, some e-bikes come with a throttle mode, eliminating the need to pedal, thereby functioning much like a motorcycle. - Regulatory AmbiguityDifferent regions have varying regulations. In some places, e-bikes with a throttle mode might require registration, insurance, and even a license, while in others, they’re treated just like traditional bicycles. - Safety ImplicationsThe overlapping functionalities of e-bikes and motorcycles raise safety concerns. Should children wearing bicycle helmets on e-bikes be mandated to wear more protective motorcycle helmets? The debate continues. Comparison Table: E-Bikes vs. Motorcycles Criteria | E-Bikes | Motorcycles | Power Source | Electric Motor | Gasoline/Electric | Need to Pedal | Yes (Optional in some models) | No | Top Speed | 20-28 mph | 50+ mph | Licensing | Not required (Varies by region) | Often Required | Primary Use | Commuting, Leisure | Transportation, Sport | Table 2: A side-by-side comparison of e-bikes and motorcycles. What are E-Bikes and How Do They Work? E-bikes, or electric bicycles, represent a fusion of traditional cycling with modern technology. At their core, e-bikes are similar to regular bicycles but are enhanced with an electric motor, allowing riders to receive assistance as they pedal or even ride without pedaling at all. This blend of manual and electric power offers a unique riding experience, making e-bikes increasingly popular, especially among children. Let’s explore the intricacies of e-bikes and understand their growing appeal to the younger generation. Definition and Mechanics of E-Bikes - Core Components - Pedal and Drive System: Just like traditional bicycles, e-bikes come with pedals. The rider can choose to pedal manually, use the electric assist, or a combination of both. - Electric Motor: Positioned usually on the hub of the rear or front wheel, or near the bottom bracket, this motor provides the electric assistance. The power of the motor, often measured in watts, determines the bike’s capability to assist riders, especially on challenging terrains like uphill climbs. - Battery: The heart of an e-bike, the battery powers the electric motor. Modern e-bikes typically use lithium-ion batteries due to their lightweight nature and long lifespan. The battery’s capacity, measured in watt-hours (Wh), dictates the range or distance an e-bike can cover on a single charge. - Controller and Display: E-bikes come equipped with a controller that manages the power distribution from the battery to the motor. Many e-bikes also feature displays showing battery life, speed, distance traveled, and the level of pedal assistance. - Working MechanismWhen a rider pedals an e-bike, sensors detect the pedaling effort. Depending on the chosen level of assistance, the motor then provides the necessary boost. Some e-bikes also come with a throttle, allowing riders to receive motor assistance without pedaling. - Types of E-Bikes - Pedal-Assist/Pedelec: The motor provides assistance only when the rider is pedaling. The assistance stops once a certain speed, usually 28 mph, is reached. - Throttle-Based: Riders can control the motor with a throttle, typically located on the handlebar, without the need to pedal. - Hybrid: Combines features of both pedal-assist and throttle-based e-bikes. Popularity Among Children and Reasons for the Same - Empowerment and IndependenceE-bikes empower children, giving them a sense of independence. They can travel longer distances without getting as tired, making it easier to visit friends, parks, or even commute to school. - Skill DevelopmentWhile e-bikes offer motor assistance, they still require balance and coordination, helping children develop these essential motor skills. - Safety FeaturesMany e-bikes designed for children come with adjustable speed settings, allowing parents to set speed limits for safety. - Environmental AwarenessAs the younger generation becomes more environmentally conscious, e-bikes resonate as a green mode of transportation, producing zero emissions. - Fun FactorLet’s face it; riding an e-bike is fun! The thrill of the added speed and the ability to tackle challenging terrains make it an exciting activity for children. Motorcycles for Children Motorcycles, with their roaring engines and the promise of open-road freedom, have always captured the imagination of many. However, when we think of motorcycles, we often picture heavy, powerful machines designed for adults. But just as there are bicycles made for children, there are also motorcycles specifically crafted for the younger generation. These aren’t mere toys but are scaled-down versions of the real thing, designed with safety and the physical stature of children in mind. Let’s delve into the world of motorcycles for children and understand what sets them apart from their adult counterparts. Motorcycles Specifically Designed for Kids - Purpose and DesignMotorcycles for children are primarily designed for learning and recreation. They offer kids an introduction to the world of motorcycling in a controlled and safe manner. These motorcycles are typically lighter, have a lower seat height, and come with power limits to ensure the safety of young riders. - Safety Features - Speed Limiters: Many children’s motorcycles come with adjustable speed settings, allowing parents or guardians to set a maximum speed limit. - Kill Switches: These are safety features that allow for the immediate shutdown of the engine, often attached to the rider via a cord. - Sturdy Build: They are built to withstand the rough and tumble that comes with learning, ensuring durability despite potential falls or mishaps. - Learning PlatformsBeyond just fun, these motorcycles serve as platforms where children can learn the basics of riding, from balancing to throttle control, in a safe environment before transitioning to larger bikes. Differences Between Traditional Motorcycles and Those for Children - Size and Weight - Traditional Motorcycles: Generally larger and heavier, designed to carry adults and sometimes even a passenger. - Children’s Motorcycles: Compact, lightweight, and designed for a single child rider. - Power and Speed - Traditional Motorcycles: Equipped with powerful engines, capable of high speeds and suitable for long-distance travel. - Children’s Motorcycles: Feature smaller engines with limited power, ensuring speeds are kept within a safe range for children. - Purpose and Usage - Traditional Motorcycles: Used for a range of purposes, from daily commuting to touring and racing. - Children’s Motorcycles: Primarily for learning and recreational use, not meant for long-distance travel or high-speed racing. - Safety Features - Traditional Motorcycles: While they come with safety features, the onus is largely on the rider to ensure safety through protective gear and safe riding practices. - Children’s Motorcycles: Packed with additional safety features like speed limiters and kill switches, designed specifically keeping young riders in mind. Safety Concerns: E-Bikes and Motorcycles for Children As e-bikes and motorcycles for children gain popularity, the safety of young riders has become a paramount concern for parents, guardians, and regulators alike. While these vehicles offer a unique blend of fun and mobility, they also introduce a set of risks that are distinct from traditional bicycles. Understanding these risks, backed by injury statistics, and adopting protective measures can go a long way in ensuring a safe riding experience for children. The Risks Associated with E-Bikes and Motorcycles for Children - Speed and Control: The added power from electric motors in e-bikes and the inherent design of motorcycles can lead to higher speeds than traditional bicycles. This can pose challenges for young riders in maintaining control, especially in emergency situations. - Battery-Related Issues: Faulty batteries or improper handling can lead to overheating, fires, or even explosions in rare cases. - Traffic Interactions: Children on e-bikes or motorcycles might be tempted to ride on roads where interactions with larger vehicles pose significant risks. - Inadequate Protective Gear:A false sense of security might lead children (or their guardians) to overlook the importance of protective gear, increasing the risk of injuries during accidents. Comparing Injury Statistics Between E-Bikes and Traditional Bicycles Injury Type | E-Bikes | Traditional Bicycles | Minor Scrapes and Bruises | 40% | 60% | Fractures or Broken Bones | 25% | 15% | Head Injuries | 20% | 10% | Severe Injuries (requiring hospitalization) | 10% | 5% | Fatalities | 5% | 2% | Note: The above statistics are fictional and for illustrative purposes only. From the table, it’s evident that while minor injuries are more common with traditional bicycles, the severity of injuries tends to be higher with e-bikes due to factors like speed and interactions with traffic. Protective Measures and Gear Recommendations - Helmets: A must-have for any rider. For e-bike and motorcycle riders, helmets with a full faceguard are recommended for added protection. - Knee and Elbow Pads: These can significantly reduce the impact during falls, preventing scrapes and more severe injuries. - Gloves: Protect hands from abrasions and offer a better grip on the handlebars. - High-Visibility Clothing: Especially important for evening or low-light conditions, reflective vests or jackets ensure that children are easily spotted by other road users. - Safety Training: Before letting children ride, ensure they undergo basic safety training. This includes understanding traffic rules, emergency braking, and safe riding practices. - Regular Vehicle Maintenance: Regularly inspect the e-bike or motorcycle for any faults, especially the brakes and battery. Ensure tires are properly inflated and all electrical connections are secure. Benefits of E-Bikes and Motorcycles for Kids While safety concerns surrounding e-bikes and motorcycles for children are valid, it’s equally important to recognize the myriad benefits these vehicles offer. Beyond the thrill of the ride, they serve as tools for holistic development, fostering skills and habits that can benefit children in various facets of life. Let’s delve into the advantages of introducing kids to e-bikes and motorcycles. Skill Development and Balance - Physical Coordination:Riding requires a combination of hand-eye coordination and fine motor skills. As children navigate turns, adjust speeds, and maintain balance, they hone these abilities. - Muscle Strengthening:Even with motor assistance, riding engages various muscle groups, especially the core, which is crucial for maintaining balance. - Reflex Development:Quick decision-making is essential while riding, especially in unpredictable situations. This helps in sharpening reflexes and reaction times. Encouraging Outdoor Activities and Reducing Screen Time - Nature Appreciation:Riding outdoors exposes children to nature, fostering an appreciation for their surroundings and the environment. - Physical Health:With the increasing concern of sedentary lifestyles among children, e-bikes and motorcycles offer a fun way to stay active, promoting cardiovascular health and overall fitness. - Mental Well-being:Outdoor activities have been linked to improved mood, reduced anxiety, and better cognitive function. Riding can serve as a therapeutic escape from the digital world. - Combatting Digital Overload:In an age where screens dominate leisure time, riding provides a tangible, exhilarating alternative, pulling kids away from TVs, computers, and smartphones. Building Confidence and Independence - Mastery and Achievement:Learning to ride, especially mastering challenging maneuvers or terrains, instills a sense of achievement in children. Each milestone reached boosts their confidence. - Responsibility:Owning and maintaining a vehicle teaches children responsibility. They learn the importance of regular maintenance, safe riding practices, and respecting traffic rules. - Decision-making:On the road, riders often have to make split-second decisions. This fosters decisive thinking and problem-solving skills. - Independence:Having their mode of transportation, even if it’s just for short distances, gives children a sense of independence. They can visit friends, go to parks, or simply explore their surroundings on their own terms. The Regulatory Gray Area: E-Bikes and Motorcycles for Children The rapid rise in popularity of e-bikes and motorcycles for children has left regulators scrambling to keep up. While these vehicles offer numerous benefits, they also present unique challenges in terms of classification, safety, and public policy. The result is a regulatory gray area, where the rules governing e-bikes and child-specific motorcycles are often ambiguous, inconsistent, or even non-existent. Let’s explore the complexities of this regulatory landscape. Licensing, Registration, and Insurance Requirements (or Lack Thereof) - Licensing: In many regions, e-bikes that adhere to certain power and speed limits do not require a license. However, more powerful e-bikes, especially those with throttle controls, might be classified differently. - Registration: Some countries or states mandate the registration of e-bikes, akin to motor vehicles, while others treat them as regular bicycles. - Insurance: While not universally required, some regions mandate third-party liability insurance for e-bike riders, especially if the e-bike falls under a specific power category. - Motorcycles for Children - Licensing: Typically, any motorized vehicle, including motorcycles for children, requires a license. However, the age at which one can obtain this license varies. - Registration: Almost universally, motorcycles, regardless of their size or intended user age, need to be registered. - Insurance: Given the risks associated with motorcycles, insurance is often mandatory, covering potential damages or injuries. How Different Countries and States are Approaching This Issue - European Union: - E-bikes with a motor output of up to 250 watts and a speed limit of 25 km/h are treated as conventional bicycles. Anything beyond these specifications requires licensing and insurance. - Motorcycles for children are subject to strict regulations, with age restrictions and mandatory safety training. - United States: - Regulations vary significantly from state to state. Some states classify e-bikes into three classes based on their top speed and motor activation method. Licensing and registration requirements differ based on these classifications. - Child-specific motorcycles are generally subject to the same regulations as adult motorcycles, but age restrictions apply for licensing. - E-bikes with a maximum power output of 200 watts do not require licensing or registration. Those with pedal-assist systems and a power output up to 250 watts are also exempted. - Motorcycles for children must be registered, and riders need to obtain a special license, often after undergoing safety training. - Asia (Varies by Country): - In countries like Japan, e-bikes with a power output exceeding 250 watts are considered mopeds and require licensing and insurance. - In India, e-bikes that have a top speed of 25 km/h and a motor output of up to 250 watts are exempt from registration and licensing. Parental Perspectives and Decisions: E-Bikes and Motorcycles for Children The decision to allow a child to ride an e-bike or motorcycle is a significant one for parents. It’s a choice that often stems from a blend of the child’s enthusiasm, the perceived benefits of riding, and the potential risks involved. As e-bikes and child-specific motorcycles become more prevalent, parents find themselves navigating a complex web of information, opinions, and emotions. Let’s delve into the perspectives of parents and the factors that influence their decisions. Real-life Experiences and Testimonials from Parents - Emma, Mother of Two: “My son started with an e-bike when he was 10. It was a great way for him to gain independence, and I felt it was safer than letting him roam the neighborhood on foot. He’s now 13 and has transitioned to a small motorcycle. With the right safety gear and training, I believe it’s a fantastic learning experience.” - Raj, Father of a Daughter: “I was initially against the idea. Motorcycles, even for kids, sounded dangerous. But after seeing the safety features and attending a few training sessions with my daughter, I changed my mind. It’s been a confidence booster for her.” - Linda, Single Mother: “Both my kids ride e-bikes to school. It’s eco-friendly, and they love the sense of freedom. I’ve set speed limits on their bikes, and we’ve established safe routes for their commute.” - Carlos, Parent of Three: “It’s not for every child. My eldest took to it naturally, but my middle child had no interest. It’s essential to ensure it’s the child’s choice and not parental pressure.” Factors Parents Consider Before Allowing Their Children to Ride - Safety: The paramount concern for most parents. They often research the safety features of the vehicle, invest in protective gear, and ensure their child receives proper training. - Child’s Maturity and Responsibility: Not all children are ready to handle the responsibility that comes with riding. Parents often assess their child’s maturity level, decision-making skills, and general awareness before giving the green light. - Purpose of the Vehicle: Is it for recreation, commuting to school, or both? The intended use can influence the decision, with some parents more comfortable with the idea of recreational use in controlled environments. - Legal and Regulatory Considerations: Parents need to be aware of the local regulations regarding licensing, registration, and age restrictions. Compliance with the law is a significant factor in the decision-making process. - Peer Influence: Sometimes, the decision is influenced by peer pressure or the child’s friends engaging in similar activities. Parents often weigh the pros and cons of allowing their child to join in versus feeling left out. - Cost: E-bikes and motorcycles for children can be expensive. The financial aspect, including the cost of the vehicle, maintenance, safety gear, and training, can influence the decision. The Future of E-Bikes and Motorcycles for Children The realm of personal mobility is undergoing a transformative phase, with e-bikes and motorcycles for children at the forefront of this evolution. As technology continues to advance at an unprecedented rate, the next decade promises exciting developments in this sector. From enhanced safety features to innovative designs, the future of e-bikes and motorcycles for children is bright. Let’s explore the potential trajectory of these vehicles in the coming years. Technological Advancements and Safety Improvements - Battery Technology: As research in battery technology progresses, we can expect e-bikes to have longer ranges, reduced charging times, and increased overall efficiency. This will make e-bikes even more appealing for daily commuting and recreational activities. - Smart Features: Integration of IoT (Internet of Things) in e-bikes and motorcycles will lead to features like real-time tracking, geofencing (to set boundaries for where the bike can go), and even parental controls to monitor and limit speed remotely. - Advanced Safety Systems: Future e-bikes and motorcycles for children might come equipped with advanced braking systems, collision avoidance sensors, and even AI-driven features that can predict potential hazards and take preventive actions. - Ergonomic Designs: As understanding of child physiology and ergonomics improves, we can expect designs that are more tailored to children’s body structures, ensuring comfortable and safer rides. - Eco-friendly Materials: With a growing emphasis on sustainability, future e-bikes and motorcycles will likely incorporate more eco-friendly materials, both in the vehicle’s construction and the battery components. Predictions for the Next Decade - Rise in Popularity: As urban areas become more congested and there’s a collective shift towards eco-friendly transportation, e-bikes, in particular, will see a significant rise in popularity among children and teenagers. - Regulatory Clarity: Given the increasing adoption of these vehicles, governments worldwide will likely introduce clearer regulations, focusing on safety standards, age restrictions, and usage guidelines. - Customization: Future e-bikes and motorcycles for children might offer extensive customization options, from design aesthetics to performance features, allowing for a more personalized riding experience. - Integration with Smart Cities: As cities become smarter, e-bikes and motorcycles could seamlessly integrate with urban infrastructure, communicating with traffic lights, parking spots, and even other vehicles to ensure safer and more efficient rides. - Affordability: As technology becomes more mainstream and production scales up, prices of e-bikes and motorcycles for children are likely to become more affordable, making them accessible to a broader audience. Navigating the world of e-bikes and motorcycles for children presents both exciting opportunities and genuine concerns for parents. These vehicles offer children a unique blend of skill development, outdoor engagement, and a sense of independence. However, they also come with potential safety risks, regulatory ambiguities, and notable financial considerations. For parents contemplating this decision, thorough research, prioritizing safety training, staying updated on local regulations, and maintaining open communication with their child are crucial. By carefully weighing the benefits against the challenges and ensuring proper precautions, parents can make informed decisions that enrich their child’s experiences while prioritizing their safety and well-being.
<urn:uuid:4fbfd036-9252-4389-9c73-45c324340e75>
CC-MAIN-2024-38
https://hextechsecurity.com/e-bikes-vs-motorcycles-for-children-a-comprehensive-guide/
2024-09-17T09:25:11Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651750.27/warc/CC-MAIN-20240917072424-20240917102424-00125.warc.gz
en
0.939208
4,944
2.8125
3
Microlearning, also referred to as performance support, focuses on delivering knowledge at the point in time that it is needed to accomplish a task. Microlearning refers to the delivery of knowledge in small chunks at the time of need. Microlearning is ideal for knowledge intensive roles where employees may need certain information prior to or while performing a task. Nearly every job can benefit from Microlearning, especially Sales and Service professionals and critical operations staff in vertical markets, such as Aerospace, Defense, Healthcare, and Manufacturing Microlearning is a growing trend and a key feature of modern learning platforms. Sharing short videos and voice memos within social media platforms is a common practice, and modern knowledge workers will easily adapt to microlearning within the enterprise. Also called the anti-LMS, microlearning focuses on content and outcomes versus tracking.
<urn:uuid:f434f106-7cdd-4c58-9073-c163fa142c2e>
CC-MAIN-2024-38
https://aragonresearch.com/glossary-microlearning/
2024-09-07T18:27:55Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650898.24/warc/CC-MAIN-20240907162417-20240907192417-00125.warc.gz
en
0.942753
164
2.875
3
Organizations experience the appearance of numerous threats that may have a direct impact on the organization’s functioning, its image, and legal compliance. The two of them are the Network and Information Systems Directive 2 (NIS2) and the Digital Operational Resilience Act (DORA). Therefore, one could quite logically conclude that the efficiency of such regulations is inconceivable without efficient risk management programs. Risk management is thus defined as a strategic management process that enables the identification, evaluation and minimization of risks in order to improve on the sustainability of the organization. However, risk management is even more important in the case of NIS2 and DORA since these regulations are designed to provide protection of the assets, IT systems, and financial services on the organisational level. The knowledge of risks’ management is the understanding that you may be prepared for the shift in the legislation concerning your organization, to keep on operating, and to safeguard the property of the stakeholders. Risk Management Strategy: Main Components This is especially important concerning the provisions of NIS2 and DORA, and consequently, it is suggested to start the process of creating the risk management strategy. This strategy should encompass the following key elements: - Risk Identification and Assessment: Assess the future threats that your organisation may experience which may be in the form of cyberrisks, operations risks and noncompliance risks. - Risk Mitigation and Control: All the described risks have to be managed and reduced using security solutions, incidents handling plans and other actions that will make the employees of the organization understand the risks. - Continuous Monitoring and Evaluation: The risk profile is fixed and has to be updated frequently, the efficiency of the risks management controls, which are in use at the moment, has to be determined and adjusted according to the changes in the risks and the new legislation. - Governance and Accountability: For the purpose of performing the mentioned roles, responsibilities and decision-making matrix of enhancing the management of risks and governance of the organization. - Integrated Approach: Ensure that you have a proper risk management plan that corresponds to your business plan and that it is harmonious with your organization’s goals, activities, and tools. If the following aspects are taken into consideration, the requirements of NIS2 and DORA can be met, and your company will be able to create a stable and flexible risk management system. Risk Assessment and Risk Identification Regarding NIS2 and DORA Compliance The first key step in risk management process is risk identification or risk appraisal. This assessment should therefore ought to concentrate on the particular risks that your organisation is exposed to in regards to NIS2 and DORA compliance. - Regulatory Compliance Risks: Analyse the degree of conformity of your organization with the requirements defined by NIS2 and DORA, and the potential non-conformity. - Cybersecurity Risks: Determine the threats and vulnerabilities for your organizations information systems like data loss, virus and intrusion. - Operational Risks: Identify risks that may impact the execution of your organisation’s strategic activities; these could be system failures, supply chain interruptions, and human errors. - Reputational Risks: In this case, it is necessary to identify the possible outcomes of compliance noncompliance or security breaches in relation to your company’s image and reputation. - Financial Risks: Look at the risks which are likely to be associated with non-compliance which may include fines, legal expenses and the impact that such fines may have on the stability of your organization. Via risk assessment, one is in a position to determine all the risks that an organization is prone to so that he or she can take measures to deal with the risks in question. Risk Management and Risk Mitigation in organizations The next process after the risk analysis and the risk evaluation is the management of the risks that are prevailing in your organization. This may include: - Implementing Robust Security Controls: Use firewalls, IDSs and IPSs, encryption methods, access restrictions, etc., at least at two levels. - Developing Incident Response and Business Continuity Plans: Ensure that your organization has incident and business continuity management plans that are well-coordinated and well thought out in the event of disruptive incidents in order to be prepared and be able to respond to them effectively. - Enhancing Employee Awareness and Training: Develop a risk management culture by providing training and risk awareness programs to ensure that your employees understand the risks and where to report them. - Establishing Effective Governance and Oversight: This will enable one to be able to assign clear and well defined roles and responsibilities in relation to risk management and also to enable proper decision making on the risk management. - Collaborating with Regulatory Authorities and Industry Peers: Meet the regulatory agencies and other players in the industry to make sure is in touch with current changes in the regulatory factors and to know how it can share and develop strategies of dealing with risks. When implementing these risk management practices, you will be able to significantly reduce the risks that your organization faces and ensure the necessary conditions for its success in the context of compliance with NIS2 and DORA. Risk Management Processes and Controls The management of risks therefore has to be a well coordinated process with laid down procedures and measures that can be implemented, controlled and even checked. Some key steps in this process include: - Risk Identification and Assessment: Risk registers, risk matrices, risk heat maps should be employed to establish a way to permanently assess and perform recurring risk reviews that your organisation is facing. - Risk Mitigation Strategies: Some of the competencies include: Assess and measure risks and design and deploy particular risk management strategies including risk avoidance, risk mitigation, risk transfer or risk acceptance based on the organization’s risk tolerance. - Monitoring and Reporting: Present methods to assess the impact of the risk management programme and new risks and how to report the findings to the stakeholders on a regular basis. - Continuous Improvement: Thus, it is vital to assess the relevance and efficiency of the risk management policies and controls periodically in relation to the emerging risks and the new legislation. - Integration with Business Processes: In essence, risk management should be a part of the systems and controls that are inherent in the firms’ business operations. Hence, you can use the above risk management processes and controls to make the risk management activity in your organisation systematic and performant, according to the NIS2 and DORA standards. Best Practices for Successful Risk Management in the Context of NIS2 and DORA Compliance To achieve successful risk management in the context of NIS2 and DORA compliance, consider the following best practices: - Establish a Risk Management Framework: An organization should make sure that it has the appropriate risk management framework that complies with the international standards like ISO 31000 or NIST SP 800-37 to contain the process of risk management. - Ensure Top-Down Commitment: To do this, you have to win the approval of the organizations leadership to champion on the risk management process. - Foster a Risk-Aware Culture: Foster risk responsibility in your organization by making sure all the employees in the organization understand the risks that surround them and how to report them. - Leverage Technology and Automation: Use Implant Risk management software and automatic monitoring equipments in the risk management process to improve their flow. - Maintain Regulatory Awareness: It is important that you frequently monitor and update your organization’s implemented risk management measures with the new and changing regulations, rules and recommendations of NIS2 and DORA. - Collaborate with Stakeholders: This is important in order to learn from the other officials, organisations and companies and also to know how they have implemented the measures of reducing the risk and the outcomes they have had. - Continuously Evaluate and Improve: The risk management ideas, procedures, and controls should be subject to regular updates for the new risks and the regulations. A strong and flexible system of risk management will aid your organisation in meeting the NIS2 and DORA requirements, as well as enhance its effectiveness. As a result of this constant change, risk management becomes the competency that can assist organizations on the requirements that have been set by NIS2 and DORA. Consequently, with the aid of a great risk management program, one is equipped with the capability to identify and weigh the risks that an organization is threatened with and therefore, be prepared for the suitable enforcement of the requirements of such crucial regulations.
<urn:uuid:54fa0de9-6a4d-417a-a5d5-37b7d9f66975>
CC-MAIN-2024-38
https://blog.avatier.com/mastering-risk-management-a-key-strategy-for-nis2-and-dora-compliance/
2024-09-08T22:55:22Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651035.2/warc/CC-MAIN-20240908213138-20240909003138-00025.warc.gz
en
0.947831
1,772
2.546875
3
John the Ripper is a free password cracking software tool. Initially developed for the Unix operating system, it now runs on fifteen different platforms (eleven of which are architecture-specific versions of Unix, DOS, Win32, BeOS, and OpenVMS). It is one of the most popular password testing and breaking programs as it combines a number of password crackers into one package, autodetects password hash types, and includes a customizable cracker. It can be run against various encrypted password formats including several crypt password hash types most commonly found on various Unix versions (based on DES, MD5, or Blowfish), Kerberos AFS, and Windows NT/2000/XP/2003 LM hash. Additional modules have extended its ability to include MD4-based password hashes and passwords stored in LDAP, MySQL, and others. Cracking password in Kali Linux using John the Ripper is very straight forward. In this post, I will demonstrate that. John the Ripper is different from tools like Hydra. Hydra does blind brute-forcing by trying username/password combinations on a service daemon like ftp server or telnet server. John however needs the hash first. So the greater challenge for a hacker is to first get the hash that is to be cracked. Now a days hashes are more easily crackable using free rainbow tables available online. Just go to one of the sites, submit the hash and if the hash is made of a common word, then the site would show the word almost instantly. Rainbow tables basically store common words and their hashes in a large database. Larger the database, more the words covered. One of the modes John the Ripper can use is the dictionary attack. It takes text string samples (usually from a file, called a wordlist, containing words found in a dictionary or real passwords cracked before), encrypting it in the same format as the password being examined (including both the encryption algorithm and key), and comparing the output to the encrypted string. It can also perform a variety of alterations to the dictionary words and try these. Many of these alterations are also used in John’s single attack mode, which modifies an associated plaintext (such as a username with an encrypted password) and checks the variations against the hashes. John also offers a brute force mode. In this type of attack, the program goes through all the possible plaintexts, hashing each one and then comparing it to the input hash. John uses character frequency tables to try plaintexts containing more frequently used characters first. This method is useful for cracking passwords which do not appear in dictionary wordlists, but it takes a long time to run. John the Ripper uses a 2 step process to cracking a password. First it will use the passwd and shadow file to create an output file. Next, you then actually use dictionary attack against that file to crack it. In short, John the Ripper will use the following two files: Cracking password using John the Ripper I will also add john to sudo group, assign /bin/bash as his shell. There’s a nice article I posted last year which explains user creating in Linux in great details. It’s a good read if you are interested to know and understand the flags and this same structure can be used to almost any Linux/Unix/Solaris operating system. Also, when you create a user, you need their home directories created, so yes, go through creating user in Linux post if you have any doubts. Now, that’s enough mambo jumbo, let’s get to business. First let’s create a user named john and assign password as his password. (very secured..yeah!) root@kali:~# useradd -m john -G sudo -s /bin/bash root@kali:~# passwd john Enter new UNIX password: <password> Retype new UNIX password: <password> passwd: password updated successfully root@kali:~# Now that we have created our victim, let’s start with unshadow commands. The unshadow command will combine the extries of /etc/passwd and /etc/shadow to create 1 file with username and password details. When you just type in unshadow, it shows you the usage anyway. root@kali:~# unshadow Usage: unshadow PASSWORD-FILE SHADOW-FILE root@kali:~# unshadow /etc/passwd /etc/shadow > /root/johns_passwd I’ve redirected the output to /root/johns_passwd file because I got the ticks for organizing things. Do what you feel like here. Cracking process with John the Ripper At this point we just need a dictionary file and get on with cracking. John comes with it’s own small password file and it can be located in /usr/share/john/password.lst. I’ve showed the size of that file using the following command. root@kali:~# ls -ltrah /usr/share/john/password.lst You can use your own password lists too or download a large one from Internet (there’s lots of dictionary file in terabyte size). root@kali:~# john --wordlist=/usr/share/john/password.lst /root/johns_passwd Created directory: /root/.john Warning: detected hash type "sha512crypt", but the string is also recognized as "crypt" Use the "--format=crypt" option to force loading these as that type instead Using default input encoding: UTF-8 Loaded 2 password hashes with 2 different salts (sha512crypt, crypt(3) $6$ [SHA512 128/128 SSE2 2x]) Will run 2 OpenMP threads Press 'q' or Ctrl-C to abort, almost any other key for status password (john) 1g 0:00:00:06 DONE (2015-11-06 13:30) 0.1610g/s 571.0p/s 735.9c/s 735.9C/s modem..sss Use the "--show" option to display all of the cracked passwords reliably Session completed root@kali:~# Looks like it worked. So we can now use john –show option to list cracked passwords. Note that it’s a simple password that existed in the dictionary so it worked. If it wasn’t a simple password, then you would need a much bigger dictionary and lot longer to to crack it. root@kali:~# john --show /root/johns_passwd john:password:1000:1001::/home/john:/bin/bash 1 password hash cracked, 1 left root@kali:~# John the Ripper advanced commands: Now that we have completed the basics of John the Ripper and cracked a password using it, it’s possibly time to move on to bigger and more complex things. For that you should check the documentation on cracking MODES and examples of John the Ripper usage. John the Ripper's cracking modes - Click to expand John the Ripper’s cracking modes. Mode descriptions here are short and only cover the basic things. Check other documentation files for information on customizing the modes. This is the simplest cracking mode supported by John. All you need to do is specify a wordlist (a text file containing one word per line) and some password files. You can enable word mangling rules (which are used to modify or “mangle” words producing other likely passwords). If enabled, all of the rules will be applied to every line in the wordlist file producing multiple candidate passwords from each source word. The wordlist should not contain duplicate lines. John does not sort entries in the wordlist since that would consume a lot of resources and would prevent you from making John try the candidate passwords in the order that you define (with more likely candidate passwords listed first). However, if you don’t list your candidate passwords in a reasonable order, it’d be better if you sort the wordlist alphabetically: with some hash types, John runs a bit faster if each candidate password it tries only differs from the previous one by a few characters. Most wordlists that you may find on the Net are already sorted anyway. On the other hand, if your wordlist is sorted alphabetically, you do not need to bother about some wordlist entries being longer than the maximum supported password length for the hash type you’re cracking. To give an example, for traditional DES-based crypt(3) hashes only the first 8 characters of passwords are significant. This means that if there are two or more candidate passwords in the wordlist whose first 8 characters are exactly the same, they’re effectively the same 8 character long candidate password which only needs to be tried once. As long as the wordlist is sorted alphabetically, John is smart enough to handle this special case right. In fact, it is recommended that you do not truncate candidate passwords in your wordlist file since the rest of the characters (beyond the length limit of your target hash type) are likely still needed and make a difference if you enable word mangling rules. The recommended way to sort a wordlist for use with default wordlist rule set is: tr A-Z a-z < SOURCE | sort -u > TARGET See RULES for information on writing your own wordlist rules. “Single crack” mode. This is the mode you should start cracking with. It will use the login names, “GECOS” / “Full Name” fields, and users’ home directory names as candidate passwords, also with a large set of mangling rules applied. Since the information is only used against passwords for the accounts it was taken from (and against password hashes which happened to be assigned the same salt), “single crack” mode is much faster than wordlist mode. This permits for the use of a much larger set of word mangling rules with “single crack”, and their use is always enabled with this mode. Successfully guessed passwords are also tried against all loaded password hashes just in case more users have the same password. Note that running this mode on many password files simultaneously may sometimes get more passwords cracked than it would if you ran it on the individual password files separately. This is the most powerful cracking mode, it can try all possible character combinations as passwords. However, it is assumed that cracking with this mode will never terminate because of the number of combinations being too large (actually, it will terminate if you set a low password length limit or make it use a small charset), and you’ll have to interrupt it earlier. That’s one reason why this mode deals with trigraph frequencies, separately for each character position and for each password length, to crack as many passwords as possible within a limited time. To use the mode you need a specific definition for the mode’s parameters, including password length limits and the charset to use. These parameters are defined in the configuration file sections called [Incremental:MODE], where MODE is any name that you assign to the mode (it’s the name that you will need to specify on John’s command line). You can either use a pre-defined incremental mode definition or define a custom one. As of version 1.8.0, pre-defined incremental modes are “ASCII” (all 95 printable ASCII characters), “LM_ASCII” (for use on LM hashes), “Alnum” (all 62 alphanumeric characters), “Alpha” (all 52 letters), “LowerNum” (lowercase letters plus digits, for 36 total), “UpperNum” (uppercase letters plus digits, for 36 total), “LowerSpace” (lowercase letters plus space, for 27 total), “Lower” (lowercase letters), “Upper” (uppercase letters), and “Digits” (digits only). The supplied .chr files include data for lengths up to 13 for all of these modes except for “LM_ASCII” (where password portions input to the LM hash halves are assumed to be truncated at length 7) and “Digits” (where the supplied .chr file and pre-defined incremental mode work for lengths up to 20). Some of the many .chr files needed by these pre-defined incremental modes might not be bundled with every version of John the Ripper, being available as a separate download. See CONFIG and EXAMPLES for information on defining custom modes. You can define an external cracking mode for use with John. This is done with the configuration file sections called [List.External:MODE], where MODE is any name that you assign to the mode. The section should contain program code of some functions that John will use to generate the candidate passwords it tries. The functions are coded in a subset of C and are compiled by John at startup when you request the particular external mode on John’s command line. See EXTERNAL. What modes should I use? See EXAMPLES for a reasonable order of cracking modes to use. John the Ripper - Usage Examples - Click to expand John the Ripper usage examples. These examples are to give you some tips on what John’s features can be used for. 1. First, you need to get a copy of your password file. If your system uses shadow passwords, you may use John’s “unshadow” utility to obtain the traditional Unix password file, as root: umask 077 unshadow /etc/passwd /etc/shadow > mypasswd (You may need to replace the filenames as needed.) Then make “mypasswd” available to your non-root user account that you will run John under. No further commands will need to be run as root. If your system is ancient enough that it keeps passwords right in the world-readable /etc/passwd, simply make a copy of that file. If you’re going to be cracking Kerberos AFS passwords, use John’s “unafs” utility to obtain a passwd-like file. Similarly, if you’re going to be cracking Windows passwords, use any of the many utilities that dump Windows password hashes (LM and/or NTLM) in Jeremy Allison’s PWDUMP output format. Some of these utilities may be obtained here: 2. Now, let’s assume you’ve got a password file, “mypasswd”, and want to crack it. The simplest way is to let John use its default order of cracking modes: This will try “single crack” mode first, then use a wordlist with rules, and finally go for “incremental” mode. Please refer to MODES for more information on these modes. It is highly recommended that you obtain a larger wordlist than John’s default password.lst and edit the “Wordlist = …” line in the configuration file (see CONFIG) before running John. Some wordlists may be obtained here: Of those available in the collection at the URL above, all.lst (downloadable as all.gz) and huge.lst (only available on the CD) are good candidates for the “Wordlist = …” setting. 3. If you’ve got some passwords cracked, they are stored in $JOHN/john.pot. The john.pot file is not meant to be human-friendly. You should be using John itself to display the contents of its “pot file” in a convenient format: john --show mypasswd If the account list gets large and doesn’t fit on the screen, you should, of course, use your shell’s output redirection. You might notice that many accounts have a disabled shell. You can make John skip those in the report. Assuming that the disabled shell is called “/etc/expired”, the command would be: john --show --shells=-/etc/expired mypasswd or shorter, but will also match “/any/path/expired”: john --show --shells=-expired mypasswd or if you also want to ignore some other shell, say “/etc/newuser”: john --show --shells=-expired,newuser mypasswd To check if any root (UID 0) accounts got cracked: john --show --users=0 mypasswd or to check for cracked root (UID 0) accounts in multiple files: john --show --users=0 *passwd* *.pwd To display the root (username “root”) account only: john --show --users=root mypasswd And finally, to check for privileged groups: john --show --groups=0,1 mypasswd 4. You might prefer to manage the cracking modes manually. It is wise to start with “single crack” mode: john --single mypasswd or since the GNU-style double dashes are optional and since option names can be abbreviated for as long as they remain unambiguous: john -si mypasswd You should not abbreviate options in scripts which you would want to work with future versions of John since what is unambiguous now might become ambiguous with the addition of more options. If you have more files to crack, it is preferable to load them at the same time: john --single passwd1 passwd2 john --single *passwd* *.pwd This way, John will run faster and might even crack more passwords than it would if you ran it on each password file separately. 5. To catch weak passwords not derived from readily available users’ personal information, you should proceed with cracking modes demanding more processor time. First, let’s try a tiny wordlist with word mangling rules enabled: john --wordlist=password.lst --rules mypasswd or abbreviating the options: john -w=password.lst -ru mypasswd Then proceed with a larger wordlist, also applying the mangling rules: john --wordlist=all.lst --rules mypasswd If you’ve got a lot of spare disk space to trade for performance and the hash type of your password files is relatively slow, you may use John’s “unique” utility to eliminate any duplicate candidate passwords: john --wordlist=all.lst --rules --stdout | unique mangled.lst john --wordlist=mangled.lst mypasswd If you know that your target hash type truncates passwords at a given length, you may optimize this even further: john --wordlist=all.lst --rules --stdout=8 | unique mangled8.lst john --wordlist=mangled8.lst mypasswd Alternatively, you may simply use huge.lst available on Openwall wordlist collection CDs. It has word mangling rules pre-applied for the most common languages and it has any duplicates purged. Depending on target hash type, the number of different salts (if applicable), the size of your wordlist, rules, and processor performance, wordlist-based cracking may take anywhere from under a second to many days. You do not have to leave John running on a (pseudo-)terminal. If running John on a Unix-like system, you can simply disconnect from the server, close your xterm, etc. John will catch the SIGHUP (“hangup” signal) and continue running. Alternatively, you may prefer to start it in the background right away: john --wordlist=all.lst --rules mypasswd & Obviously, the “&” is specific to Unix shells and will not work on most other platforms. You may further enhance this by specifying a session name: john --session=allrules --wordlist=all.lst --rules mypasswd & This ensures that you won’t accidentally interfere with the instance of John running in the background if you proceed to start other sessions. To view the status of a running session, use: for the default session or: for any other session. This works for both interrupted and running sessions. To obtain the most up-to-date information from a running session on a Unix-like system, send a SIGHUP to the appropriate “john” process. Any interrupted sessions may be continued with: Finally, to make John have less impact on other processes, you should set the option “Idle = Y” in the configuration file (see CONFIG). The default may vary depending on the version and build of JtR. To only crack accounts with a “good” shell (in general, the shell, user, and group filters described above work for all cracking modes as well): john --wordlist=all.lst --rules --shells=sh,csh,tcsh,bash mypasswd Like with all other cracking modes, it is faster to crack all the files you need cracked simultaneously: john --wordlist=all.lst --rules passwd1 passwd2 You can crack some passwords only. This will try cracking all root (UID 0) accounts in all the password files: john --wordlist=all.lst --rules --users=0 *passwd* Alternatively, you may wish to not waste time cracking your very own passwords, if you’re sure they’re uncrackable: john --wordlist=all.lst --rules --users=-root,solar *passwd* Sometimes it is useful to split your password hashes into two sets which you crack separately, like: john --wordlist=all.lst --rules --salts=2 *passwd* john --wordlist=all.lst --rules --salts=-2 *passwd* This will make John try salts used on two or more password hashes first and then try the rest. Total cracking time will be almost the same, but you will get some passwords cracked earlier, which is useful, for example, for penetration testing and demonstrations to management. Similarly, you may check all password hashes with a small wordlist, but only those that you can check faster (with “–salts=2”) with a larger one. With large numbers of password hashes and/or with a highly non-uniform distribution of salts, it may be appropriate to use a threshold larger than 2 with “–salts” (sometimes even values as high as 1000 will do). Note that the default wordlist rules include “:” (a no-op – try words as they are in the list) on the first line. If you already ran through a wordlist without using rules, and then decided to also try the same wordlist with rules, you’d better comment this line out. 6. The most powerful cracking mode in John is called “incremental” (not a proper name, but kept for historical reasons). You can simply run: john --incremental mypasswd john -i mypasswd This will use the default “incremental” mode parameters, which are defined in the configuration file’s section named either [Incremental:ASCII] (for most hash types) or [Incremental:LM_ASCII] (for Windows LM hashes). By default, the [Incremental:ASCII] parameters are set to use the full printable ASCII character set (95 characters) and to try all possible password lengths from 0 to 13 (if the current hash type has a lower maximum password length, incremental mode’s length limit is reduced accordingly). [Incremental:LM_ASCII] is similar, except that it takes advantage of LM hashes being case-insensitive and of their halves being limited to 7 characters each. Don’t expect “incremental” mode sessions to terminate in a reasonable time (unless all the passwords are weak and get cracked), read MODES for an explanation of this. In some cases it is faster to use some other pre-defined incremental mode parameters and only crack simpler passwords, from a limited character set. The following command will try 10 different characters only, passwords from “0” to “99999999999999999999” (in an optimal order): john -i=digits mypasswd Of course, you can use most of the additional features demonstrated above for wordlist mode with “incremental” mode as well. For example, on a large-scale penetration test, you may have John crack only root (UID 0) accounts in a set of password files: john -i -u=0 *.pwd 7. If you’ve got a password file for which you already have a lot of passwords cracked or obtained by other means, and the passwords are unusual, then you may want to generate a new charset file, based on character frequencies from that password file only: john --make-charset=custom.chr mypasswd Then use that new file with “incremental” mode. If you’ve got many password files from a particular country, organization, etc., it might be useful to use all of them for the charset file that you then use to crack even more passwords from these files or from some other password files from the same place: john --make-charset=custom.chr passwd1 passwd2 [ Configure your custom "incremental" mode now. See below. ] john -i=custom passwd3 You can use some pre-defined or custom word filters when generating the charset file to have John consider some simpler passwords only: john --make-charset=my_alpha.chr --external=filter_alpha mypasswd If your “pot file” got large enough (or if you don’t have any charset files at all), you might want to use it to generate a new set of main charset files: where “makechr” is a script that invokes “john –make-charset=…” with varying filenames, for all of the external mode word filters defined in the configuration file. In this example, John will overwrite the charset files with new ones that are based on your entire $JOHN/john.pot (John uses the entire “pot file” if you don’t specify any password files). 8. Finally, you might want to e-mail all users with weak passwords to tell them to change their passwords. (This is not always a good idea, though, since lots of people do not check their e-mail or ignore such messages, and the messages can be a hint for crackers.) Edit the “mailer” script supplied with John: the message it sends and possibly the mail command (especially if the password file is from a different machine). Then run: Please refer to CONFIG for general information on the configuration file and its possible locations. 1. Let’s assume that you notice that in some password file a lot of users have their passwords set to login names with “?!” appended. Then you just make a new “single crack” mode rule (seeRULES for information on the syntax) and place it somewhere near the beginning: Hint: if you want to temporarily disable all of the default rules, you can simply rename the section to something John doesn’t use and define a new one with the section’s old name, but be sure to leave the “List.” prefix of the name intact to maintain correct configuration file syntax. All the same applies to wordlist mode rules as well. 2. If you generate a custom charset file (described above) you will also need to define a configuration file section with the “incremental” mode parameters. In the simplest case it will be like this (where “Custom” can be replaced with any name you like): [Incremental:Custom] File = custom.chr This way, John will only use characters from passwords used to generate the charset file only. To make John try some more characters, add: Extra = !@#$% These extra characters will then be added, but still considered the least probable. If you want to make sure that, with your extra characters, John will try 95 different characters, you can add: CharCount = 95 This will make John print a warning if it only has fewer than 95 characters in its charset. You can also use CharCount to limit the number of different characters that John tries, even if the charset file has more: CharCount = 20 If you didn’t use any filters when generating the charset file, setting CharCount this low will make John never attempt rare characters and character combinations, not even for really short passwords, spending the time on simple longer candidate passwords instead. However, the default length switching is usually smart enough so that you shouldn’t need this trick. To make John try passwords of certain lengths only, use the following lines: MinLen = 6 MaxLen = 8 Setting “MinLen” high, as in the example above, is reasonable if shorter passwords weren’t allowed to set on the machine you got the password file from (however, note that root can usually set any password for any user and there are often loopholes in operating systems’ password policy enforcement capabilities). On the contrary, you may want to set “MaxLen” low if you think there are a lot of short passwords. 3. Another example: a lot of users at some site use short duplicated words as their passwords, such as “fredfred”. As the number of such potential passwords is fairly low, it makes sense to code a new external cracking mode that tries them all, up to some length. You can find the actual implementation of such a cracking mode with lots of comments in the default configuration file supplied with John. Please refer to EXTERNAL for information on the programming language used.
<urn:uuid:0a1273d1-dc6a-41f4-b3a9-e79f5e0981b6>
CC-MAIN-2024-38
https://www.blackmoreops.com/2015/11/10/cracking-password-in-kali-linux-using-john-the-ripper/
2024-09-09T23:54:03Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651164.37/warc/CC-MAIN-20240909233606-20240910023606-00825.warc.gz
en
0.879331
6,457
2.65625
3
By Matt Bogan You’re aware of the need for data encryption. However, you may not have a clear picture of how it works. Cryptography—the study of secure communications—is a complex field of science, intersecting a wide range of disciplines and constantly evolving. Cryptography predates computers, with roots reaching back to ancient Greece and Egypt. It’s believed that Julius Caesar sent secure messages to his most trusted staff by replacing each letter of the alphabet with the letter three places after it: A became D, B became E, and so forth. This method of data encryption is sometimes called a Caesar cipher. By the ninth century, mathematicians applied statistical analysis to defeat simple substitution ciphers such as the Caesar cipher. Thus we have perhaps the first historical instance of an encryption algorithm being cracked. As long as we’ve needed to encode information, there have been others wanting to decode it for nefarious reasons. In WWII, Allied cryptanalysts working to crack encrypted Axis communications are widely credited with shortening the war by months or even years. Today, computer scientists called “white-hat hackers” work to defeat existing encryption protocols to identify weaknesses before they can be exploited. The complexity of encryption algorithms and the protocols that implement them have increased dramatically. Despite that, having an uncrackable encryption remains hypothetical. Instead, data security aims for a moving target: difficult enough to crack with present technology that it’s astronomically unlikely anyone would be able to. Yet available computing power continues to increase, turning unlikely into likely. A protocol that today would be considered too complex for even a supercomputer to crack in any reasonable amount of time may be trivial to defeat on consumer-grade hardware in a handful of years. DES (Data Encryption Standard), a protocol developed by IBM and implemented by government agencies and militaries around the world as recently as the early 1990s, was cracked in less than a day in 1999 by a cluster of thousands of computers working together. By 2016, it was cracked by a single computer using off-the-shelf components, and today, using advanced cryptographic attack methods, DES can be defeated in under a minute with hardware you can pick up at your favorite big-box store. Like their ninth-century counterparts using mathematics to peer behind the curtain of simple substitution ciphers, modern cryptanalysts apply the latest technological advances to accomplish what was previously impossible. If everyone’s motive in defeating encryption was altruism, this discussion might be academic, but this is not the case. Far from the image of loosely affiliated teenage malcontents portrayed in popular media, hacking in the twenty-first century is big business. With backing from organized crime syndicates and foreign governments, the goals are a lot more sinister than causing a little digital mischief. Your data is a battlefield, and encryption is the arms race. The cryptographically secure lifespan of a common algorithm known as MD5 was roughly a decade, and the SHA-1 algorithm fared only slightly better. Concerns over vulnerability have contributed to tech giants like Apple, Google, and Microsoft deprecating protocols based on them. The 1.0 and 1.1 versions of TLS (Transport Layer Security) have recently joined Caesar ciphers and DES in the boneyard of obsolete cryptography. These methods once seemed functionally impenetrable, but with subsequent technological advancements, they now offer only moderate inconvenience to a well-resourced attacker. With the successors already being subjected to the scrutiny of security professionals on both sides of the battle, we can be sure more encryption methods will eventually join them. With the target of “safe enough” advancing, as businesspeople, solution providers, and individual consumers, it’s critical to ensure we keep pace. Ask your vendors what they’re doing to stay current with the latest data encryption advancements. Remember that not all encryption is equal. Older products relying on encryption methods that were state-of-the-art when they were originally developed are unlikely to offer much protection against an attack today. When assessing products to sell to your clients or use in your own business, keep in mind that marketing terms such as secure and compliant are only as meaningful as the person or organization making that claim. Savvy consumers should look for products that have been audited by independent security experts. And once you have purchased a solution, make sure you keep it current with vendor-recommended updates. It’s the only way to be certain your data will be as safe tomorrow as it is today. Matt Bogan is the product manager for Startel, a leading provider of best-in-class contact center solutions. He has been involved in the contact center industry for over fifteen years. Startel’s upcoming CMC 16.0 release incorporates the latest in encryption technology.
<urn:uuid:908511e2-5603-4b46-8898-e01358312e48>
CC-MAIN-2024-38
https://connectionsmagazine.com/article/data-encryption/
2024-09-11T05:40:49Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651344.44/warc/CC-MAIN-20240911052223-20240911082223-00725.warc.gz
en
0.964417
991
3.59375
4
Android and iPhone ambient light sensors can be turned into cameras, allowing malicious actors to secretly film unsuspecting victims and their surroundings. Smartphone ambient light sensors are typically used to automatically adjust screen brightness. However, researchers at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) claim the sensors can be used to craft a photo of the user’s surroundings. A group of researchers proposed a computational imaging algorithm that allowed the recovery of an image of the environment from the screen’s perspective. All it took was single-point light intensity changes in these sensors. “Ambient light sensors are tiny devices deployed in almost all portable devices and screens that surround us in our daily lives. As such, the authors highlight a privacy threat that affects a comprehensive class of devices and has been overlooked so far,” Princeton University professor Felix Heide, who was not involved with the paper, said. The new study suggests that ambient light sensors could intercept various user gestures, such as swiping and sliding, and capture how users interact with their phones while watching videos. The critical point of the study was to dispel a conviction that ambient light sensors can’t reveal any meaningful private information to attackers, so apps should be able to freely request access to them. According to Yang Liu, a PhD at the MIT Electrical Engineering & Computer Science Department (EECS) and CSAIL, ambient light sensors capture what we’re doing without permission, and combined with a display screen, these sensors can pose privacy risks to users. Researchers suggest that operating software makers tighten up permissions and reduce the precision and speed of the sensors. One way to combat the security issue would be to allow users similar control over app permissions to use ambient light sensors as users have with camera usage. Additional measures could encompass future devices to have ambient light sensors facing away from the user, for example, to the side of the device. The more researchers dig for innovative ways to spy on users, the deeper the hole gets. For example, a 2023 report suggested that attackers could intercept what users are typing by intercepting the sound of keystrokes via a remote conversation.
<urn:uuid:20e36daf-6f17-4259-a11c-dea8b0753324>
CC-MAIN-2024-38
https://cybernews.com/privacy/iphone-android-ambient-light-sensor-spying/
2024-09-11T05:42:46Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651344.44/warc/CC-MAIN-20240911052223-20240911082223-00725.warc.gz
en
0.926937
442
3.09375
3
How can Digital Forensics help your investigation? Today, almost every crime involves a digital device. By forensically acquiring, decoding, and analysing the data, experts can provide an overview of what’s on the device, how it got there, and who put it there. Any device with data storage has the potential to be analysed. Data can be acquired and examined from various devices, including laptops and mobile phones, CCTV, drones, game consoles and other devices. Alongside these digital forensic services, cell site specialists can provide an additional layer of analysis by plotting the potential location of a device – often adding value to the evidence and assisting with either proving or disproving a case. Talk to our Digital Forensics team today Contact UsJoin our Digital Forensics experts on Thursday 13th July at our upcoming webinar: Drone Analysis in Digital Forensic Investigations. […] Read MoreAs the use of computer forensics in criminal investigations is now commonplace, news reports have made ordinary computer users aware […]
<urn:uuid:4b152bdf-2ed2-43a6-b3fe-9743b77e9f96>
CC-MAIN-2024-38
https://www.intaforensics.com/digital-forensics/
2024-09-11T07:14:16Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651344.44/warc/CC-MAIN-20240911052223-20240911082223-00725.warc.gz
en
0.920752
204
2.53125
3
Don’t be Intimidated by HTML or CSS: An Easy Explainer on Code For those who aren’t tech-savvy, or even technologically inclined, code may seem much too complicated and frightening. However, there’s no need to be intimidated by HTML or CSS, as it’s much simpler to understand and to grasp than you might think. Code is the very fabric and foundation of every kind of technology, from your television to your smartphone to applications and websites. Once you have a basic knowledge of what it is, you’ll become familiar with how this unique tech language works. For a beginner, it’s important to know the two main elements, which are HTML, and CSS. What is HTML? HTML stands for hypertext markup language. It’s basically the framework or layout upon which a website is built, much like the framework that goes up during construction of a home. It gives a website its shape and structure. You use tags to mark elements within the website, which then allows the website to be read on browsers. Once you have a grasp of how to use HTML, you can use it to build websites, among other things. It comes in very handy for anyone who needs even a basic knowledge in order to work (or play) in the digital world. What is CSS? CSS stands for cascading style sheets. What CSS does is it allows you to create and correctly format the appearance and the layout of webpages. With a basic knowledge of CSS, you can tweak things to achieve a desired layout, as well as to change things like font size, colors, etc. CSS is what allows a website developer to set the look, the tone, and the feel of a website. It makes it easy to do quick changes and updates, as well. If you want to change the size of the font on your main page, you can do that just by tweaking the CSS, which is much easier than having to go through and re-write the html. The Importance of Grammar When it comes to writing code, each character, and the placement of each character, matters. If one element is out of place, it can throw everything off. Since codes are essentially languages, they have specific grammar rules, so to speak. Each type of code, or language, has its own set of rules that needs to be followed in order for the code to be understood. The computer will not be able to read it properly if just one seemingly minute detail is out of place. A computer goes through reading each character very quickly, and it can get thrown off by a grammar mistake. While it may sound scary, it should not be intimidating, because once you get those characters in the right spot, everything falls in to place. Once you have mastered the rules for a specific language, they will likely stick. How Code Benefits You Once you learn even just a little bit about coding, you’ll feel as though you’ve learned a new secret language. Things that once seemed complicated and mysterious, such as your smartphone, will now be just a little clearer to you. It will likely feel less foreign and mysterious. In addition, having knowledge of code can be beneficial in both your personal and professional life. Whether you’re seeking to branch out in the workplace, launch a new small business, or just want to start up a new website, you’ll be able to do so once you have learned the basics of how to code. To wrap things up: What is code? It’s the language of technology, and learning it is highly beneficial.
<urn:uuid:21ec47c0-109b-4041-9da7-0085274016e4>
CC-MAIN-2024-38
https://www.cebodtelecom.com/dont-be-intimidated-by-html-or-css-an-easy-explainer-on-code
2024-09-12T12:11:28Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651457.35/warc/CC-MAIN-20240912110742-20240912140742-00625.warc.gz
en
0.95829
749
3.046875
3
What is a virtual machine (VM)? A virtual machine (VM) is an operating system or application environment that is installed on software which imitates dedicated hardware. The end user has the same experience on a virtual machine as they would have on dedicated hardware. It is a fundamental part of many modern cloud infrastructures, the other component is called the Hypervisor, which is a piece of software that creates and runs the virtual machines. Why use a virtual machine (VM)? Using a virtual machine can reduce your hardware costs and enable you to be more flexible and scalable in your infrastructure. Latest Virtual Machine Insights When you choose between on-premises and cloud you have to first assess which one is more fitting to your strategy – not the other way around. Read about the three fundamental questions you need to consider for it. Two key areas are to be considered for choosing ‘the right tech’ for your business: Real-time data analysis for your large quantities of data and keeping flexible with cloud, on-premises and hybrid strategies. Run Exasol on Microsoft Azure cloud and leverage the flexibility, support for local storage, and durable, highly available Azure Storage. Interested in learning more? Whether you’re looking for more information about our fast, in-memory database, or to discover our latest insights, case studies, video content and blogs and to help guide you into the future of data.
<urn:uuid:36b7ecff-23a5-4f4d-8ea4-4cfb1e895ea3>
CC-MAIN-2024-38
https://www.exasol.com/glossary-term/virtual-machine/
2024-09-12T11:12:41Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651457.35/warc/CC-MAIN-20240912110742-20240912140742-00625.warc.gz
en
0.928539
295
2.875
3
Object-oriented programming languages came into their commercial heyday in the 1990s, and one of the fringe benefits of object-oriented programming research was the development of a standard way to diagram the internal structure of programs. A standard managed by the Object Management Group, UML (Unified Modeling Language) is officially described as “the industry-standard language for specifying, visualizing, constructing and documenting the artifacts of software systems” (go to www.eweek.com/links to access the UML Resource Center Web site). Grady Booch, Ivar Jacobson and Jim Rumbaugh, three leading software design experts who ended up working together at Rational Software Corp., initially designed UML. There are seven basic UML diagram types, ranging from the high-level use case diagram, showing the people and tasks in a business process, to the low-level class diagram, showing class, method and variable names used in a particular program. High-level diagram types such as use case, activity or sequence are effective ways for users (through an interview with someone who knows UML) to specify how applications should work. Class diagrams, on the other hand, are programmers tools. UML-based modeling tools take completed class diagrams and use them to directly generate code. Modeling tools can also reverse-engineer existing code back to class diagrams to visually show how a programs components fit together. One area that UML does not currently address is database modeling, now based on entity-relationship model diagrams. There is some exploratory work going on to see how UML could be used to do database modeling. Other future directions for UML include models for designing real-time systems, more ways to describe the run-time behavior of software, and distributed computing modeling. Using UML to describe software plans in a clear, visual way will help analysts ensure that their software plans meet business needs and help developers supply finished software that does the same.
<urn:uuid:6de70715-0262-4221-bd5c-e2bb12143779>
CC-MAIN-2024-38
https://www.eweek.com/development/uml-setting-the-code-diagramming-standard/
2024-09-13T17:56:22Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651535.66/warc/CC-MAIN-20240913165920-20240913195920-00525.warc.gz
en
0.925181
405
3.609375
4
I’m all about being educated and understanding the risks involved every time you connect to the internet or access your company’s network or handle sensitive data. But can being educated actually work against cybersecurity efforts? A study conducted by online IT training company CBT Nuggets found that overconfidence in our abilities, and in our devices, can end up resulting in identity theft. According to the survey of more than 2,000 respondents, 18 percent of those who admit to being tech savvy are more likely to become victims of identity theft. Also, more education increases your risk, with 24 percent of PhDs stating that they have been a victim of identity theft at some time, as opposed to 11 percent of high school graduates and 14 percent of those with a bachelor’s degree. Now, there could be a number of reasons for the disparity in the numbers, such as the amount of time someone with only a high school degree spends on a computer versus that of someone with an advanced degree. Also, higher-level staff are specifically targeted through spearphishing and whaling attacks. One of the statistics I want to highlight focuses on the types of devices and operating systems we use. Are you surprised that Apple users (22 percent) are more likely to become an identity theft victim over Windows users, but that Android phone users are more likely to be hit than iPhone users? The phone issue is likely due to stricter app rules and access in Apple devices than in Android, but the Apple versus Windows statistics show that users still carry a misconception that Macs are free from cybersecurity threats. The study also broke down how these different populations addressed security practices. For instance, PhDs again are at a higher risk of password theft because, as a group, they have the lowest percentage of unique passwords compared with those with lower degree levels. Perhaps it is not a mystery why they are also at the highest risks of identity theft. (Of course, we’ve also learned that IT staff aren’t very good about password management, either, which affirms the results of this study.) At the same time, these so-called tech savvy folks don’t seem to care much about using good security practices, as a Naked Security blog explained in discussing this study: When asked why they didn’t follow basic security recommendations, 40 percent of respondents said they were too lazy, found it to be too inconvenient, or they didn’t really care. In fact, only a smidgen of respondents – 3.7 percent – said they follow all of the basic security requirements…Which is weird, given that 65.9 percent said that having their personal information compromised is a medium or huge risk. Knowing about cybersecurity, understanding the risks, yet thinking that your tech smarts are enough is a recipe for a security disaster. Education and training are absolutely necessary in today’s digital landscape, but they mean nothing if we don’t practice what we’ve learned. Sue Marquette Poremba has been writing about network security since 2008. In addition to her coverage of security issues for IT Business Edge, her security articles have been published at various sites such as Forbes, Midsize Insider and Tom’s Guide. You can reach Sue via Twitter: @sueporemba
<urn:uuid:25e03df2-dfca-41fa-a65b-e9735c6c27de>
CC-MAIN-2024-38
https://www.itbusinessedge.com/security/does-education-and-tech-savviness-put-you-at-greater-risk-for-identity-theft/
2024-09-20T01:37:10Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652073.91/warc/CC-MAIN-20240919230146-20240920020146-00025.warc.gz
en
0.967339
674
2.59375
3
Introduction to Data Centers Data centers are a one-stop centralized facility that accommodates computer hardware, networking devices, storage systems, and other infrastructure, all of which involve storing, processing, and handling data. From small server rooms to high-tech, large-scale facilities, there has been an evolution for data centers as they have grown in capacity and effectiveness. This evolution also ensures an increase in their handling capacity as far as processing requirements were concerned. Data centers are needed to save and retrieve data to facilitate cloud computing, online applications, streaming media services, and data analytics. Components Of Data Centers - Servers and Hardware: These are the key computers operating at data centers and undertake crucial tasks such as processing data, storing information, and hosting applications, among others. Other hardware components include racks, power supplies, cooling systems, and security devices. - Networking Equipment: Networking equipment like routers, switches, and cables link servers together with gadgets inside these facilities. They also allow the exchange of information from external networks into them. - Storage Systems: Users have the flexibility to store data in various formats, including files, databases, and multimedia content. They can choose the storage solution that best fits their needs, whether that’s hard disk drives (HDDs), solid-state drives (SSDs), or storage area networks (SANs). - Cooling Systems & Power Systems: Cooling systems help regulate temperature and maintain optimal conditions, preventing equipment overheating, while power systems ensure a continuous supply of electricity through generators, UPS units, batteries, etc. Types Of Data Centers - Enterprise Data Centers: Enterprise data centers are the largest single group of enterprise-owned IT sites in terms of number worldwide. These sites are owned by individual companies that provide services such as application hosting database storage network connections. - Colocation Data Centers: Colocation data centers rent space in their racks to various clients, who then manage their respective IT operations. When outsourcing IT services through colocation data centers, businesses retain this hardware and application control. - Cloud Data Centers: Data centers that provide cloud services over the Internet through a pay-per-use pricing model are referred to as cloud data centers. These centers support virtualized environments and scalable infrastructure, among other things, that enable on-demand access to applications and data. - Edge Data Centers: Edge data centers are designed to place computing resources close to end users. This proximity helps reduce latency and network congestion that can arise from relying on centralized cloud infrastructure or other remote service providers. They focus on edge computing, content delivery, and the highly distributed nature of the Internet of Things (IoT). Instead of relying on a central location or the cloud for processing, these data centers support processing IoT data locally, improving efficiency and responsiveness. Aspect of Data Center Operations - Data Security: To safeguard information stored within edge data centers from cyber threats and data breaches, they employ firewalls, encryption, access controls, and intrusion detection systems. These security measures help prevent unauthorized access and ensure data remains protected. - Redundancy & Fault Tolerance: To ensure reliability, fault-tolerant designs in the data centers incorporate redundancy to minimize downtime in the event of hardware failures. Backup systems further enhance the availability and reliability of data center services, ensuring consistent performance even during unexpected disruptions. - Energy Efficiency & Sustainability: They aim to reduce energy consumption by using efficient cooling mechanisms and renewable resources for power. They follow practices to help data centers become more sustainable and environmentally friendly. - Disaster Recovery: They have disaster recovery plans that help sustain business activities and restore lost information during emergencies, natural calamities, or system crashes. - Modular Data Centers: Modular (containerized) refers to any design that involves the construction of prefabricated modules, allowing fast deployment and scalability adjustability according to IT’s changing requirements. - Hyperscale Data Centers: Hyperscale data centers can be thought of as a set of data center nodes linked together within a single facility. They’re well suited for cloud providers, internet companies, and large enterprises that need extensive computing storage and networking capabilities. - Green Data Centers: It is vital to note that green data centers significantly reduce their carbon footprint by employing energy-efficient designs, using renewable energy sources, and using sustainable practices. - Edge Computing and Micro Data Centers data-contrast=”auto”>Edge computing pushes processing and storage capabilities closer to end-users and IoT devices, making real-time data processing possible for low-latency applications. Micro data centers are used for edge computing deployments in distributed environments. - Artificial Intelligence: AI technologies that encompass machine learning, predictive analytics, and automation will improve data center management, optimization, and resource allocation. - 5G and IoT Impact: The rollout of 5G networks, combined with the rapid growth of Internet of Things (IoT) devices, is increasing the demand for edge computing capabilities. This need arises due to the requirement for high-speed connectivity and real-time data processing at the network edge. - Data Center Automation: Data center operation can be streamlined by automation tools such as AI-driven systems together with software-defined infrastructure, leading to efficiency improvement as well as manual task reduction. - Hybrid and Multi-Cloud Environments: Companies manage hybrid environments by implementing strategies that combine on-premises services with public cloud and private cloud integration. Data centers play a crucial role in this setup, especially as firms increasingly adopt a multi-cloud approach to enhance flexibility and optimize resources. In modern IT infrastructure, data centers are valued for supporting digital services, application delivery platforms, and storage needs for businesses and individuals. Understanding their components, types, and operations is essential for navigating emerging technology trends and global digital transformation. Share this glossary
<urn:uuid:eba49a70-d13f-4c94-b375-a2fba5cbde94>
CC-MAIN-2024-38
https://kanerika.com/glossary/data-centers/
2024-09-07T19:52:53Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650920.0/warc/CC-MAIN-20240907193650-20240907223650-00225.warc.gz
en
0.912116
1,194
3.375
3
As Spring approaches, not only do the flowers begin to blossom and the winter clouds disperse, the tech industry begins to understand what the rest of 2017 may bring. In the cybersecurity world, the dangers lurking in the Smart Home because of glaring holes in device security have been widely publicised. The CES show in Vegas showcased any number of products that will be released later this year to deal with the glaring problem. A succession of attacks on US academia through so-called Smart light bulbs and even Smart vending machines have only highlighted the deficiencies of insecure networks. It appears that nobody is safe and even internet-connected teddy bears and plushes have been gateways for hackers. That would be to exaggerate the problem, even at a time when cybersecurity is of the utmost importance. While anything new, such as IoT will always be of interest, sometimes it is the old ways that remain just as important. Cybersecurity issues are nothing new. The internet had its 25th birthday last fall and hacking was born on the same day. In those disconnected times, viruses had to be loaded by an external floppy disk, a notion now that seems almost romantic and naive. The spread of the internet then provided a whole new global delivery system to mess with individuals and businesses Some companies missed the boat completely. Microsoft incorrectly believed everything would stay on the desktop, leaving a hole for Google to step in and carve out the internet themselves. Apple’s operating system managed to insulate itself from the viruses that began to proliferate over email connections. The hackers, however, persisted. Consumer security has changed extraordinarily over those 25 years. The acceleration of distribution and the chance for intelligent hackers, be they benign or malicious, to exploit an under-thought-out system has been revolutionary. From floppy disks to internet-connected teddy bears covers a wide expanse. Those companies that invested in anti-virus solutions have moved with the times to protect consumers now have more than two decades of experience of constantly innovating and updating a (relatively cheap) product, something that IoT security companies can only dream of. Moreover, anti-virus products are as easy to use as pointing a mouse on a pad. Recently, there has been contention about whether the anti-virus industry is coming to an end, whether the products themselves are a problem. This, however, misses the point. Anti-virus is still important, because every consumer has different levels of digital and computer literacy. BullGuard is a consumer security company based in Silicon Valley that has been proactive in preparing products for the upcoming IoT battle, but is a business based on anti-virus products. CEO Paul Lipman has his eye on the crystal ball, but also an eye on the anti-virus business that his business is based on. “The Smart Home is clearly a Dumb One with cheaply made devices and minimum password protection mixing with little interoperability, but I don’t understand why the consumer appears to be blamed for the situation. “I’ve worked in this industry for ten years and seen a lot of changes and there will be more to come. But, don’t forget how anti-virus products managed to keep the internet going during the years when it was most vulnerable. I really think it still has a huge role to play in protecting the consumer,” he said. Conversely, new experimental research from security company Kaspersky Labs suggests that consumers don’t even care about their content. A report released today showed that consumers would give away all their digital content for just over $11, even if 49% of these very same people regarded their family photos as the most precious data they owned. It seems to be a mad, mad world where consumers are blamed for not protecting their Smart Homes, consumers would sell their content for peanuts and the 25-year-old history of anti-virus products is sidelined by future issues concerning IoT.
<urn:uuid:29fa4abe-83ea-48c6-a922-7dc4054f8a75>
CC-MAIN-2024-38
https://www.cybersecurity-review.com/anti-virus-still-has-an-important-role-to-play-in-cybersecurity/
2024-09-09T02:12:56Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651053.52/warc/CC-MAIN-20240909004517-20240909034517-00125.warc.gz
en
0.970236
808
2.625
3
The earthquake that hit the island of Haiti in January 2010, and the humanitarian disaster that followed, triggered an unprecedented outpouring of aid. Individuals, governments and businesses donated billions of dollars. Volunteers and humanitarian agencies descended on the island to help Haiti’s stricken population and to repair the country’s obliterated infrastructure. This spontaneous wave of people and resources called for careful coordination, and a number of information technologies were put to work. The Haiti relief effort provided a showcase for these technologies, many of which were used again in response to subsequent disasters. As a recently published report from the United Nations found, the Haiti relief effort demonstrated the power of web-based technologies to support complex operations in real-time, even in the most challenging of environments. But it also highlighted the risk of information overload at a time when focus and strong leadership are of the essence, and underlined the importance of telecommunications infrastructure, without which these tools would be of little use. One of the tools used in Haiti was Ushahidi, an open source tool that allows users to report events via text message, and for those reports to be located on an online map. The tool was originally developed in Kenya as a method of reporting incidents of electoral corruption. In Haiti, eyewitnesses used Ushahidi to report events such as sightings of survivors who were trapped under collapsed buildings. Using the GPS coordinates of the phone that the text messages were sent from, these sightings were mapped online, allowing volunteers to send relief resources to where they were needed most. Ushahidi played a similar role in coordinating relief following the earthquake in Chile a month later. What is particularly remarkable about Ushahidi, and many of the other disaster relief tools that have recently emerged, is that it was developed by volunteers – programmers who give their free time to build applications that can save lives. These tools are the work of loosely coupled organisations such as CrisisCamp, a community of volunteer developers that organises networking events around the world. Its members collaborate on prototype methods for aggregating and understanding crisis data and build applications to help organise disaster response teams. The tools that CrisisCamp members donated to the Haiti relief effort ranged from a mobile application that translates text messages from Haitian Creole into English to an entire online social network for relief workers. In January 2011, a year after the earthquake, the United Nations Foundation, along with the Harvard Humanitarian Initiative and the non-profit Vodafone Foundation, published the preliminary findings of an investigation into how volunteer-built technologies shaped relief efforts in Haiti. “Some of these efforts made a visible difference in the way that the response efforts unfolded,” writes the report’s author, John Crowley, a research fellow at the Harvard Humanitarian Initiative. “Others proved that non-traditional actors – technologists who work outside humanitarian operations and volunteer their time – could build applications that approached problems in novel ways, even when those tools could not be integrated into the operations in Haiti.” The report highlights the particular importance of mobile telecommunications, especially SMS, during emergency situations. In Haiti, the ubiquity of mobile services provided an easily accessible channel for affected citizens to communicate with aid workers. But while the abundance of information technologies available supported vital communications, it also produced a massive torrent of data. Upon arriving in Haiti in January 2010, aid workers were presented with a “perfect storm of information management challenges”, the UN Foundation report notes. Much of the baseline data relating to the country and its citizens – in paper and digital formats – had been destroyed as a result of the earthquake and its aftershocks. Humanitarian workers and volunteers first of all faced the task of re-establishing some of the data that was already lost. The study implies that the added responsibility of managing and reacting to data streams from volunteer-developed platforms “created an overwhelming sense of information overload” for aid workers, ultimately complicating their work. Another challenge identified by the UN report was the public nature of these web-based relief management tools. The information presented on systems such as Ushahidi is available not only to aid workers but any interested Internet user. Crowley argues that visualising the data in such an open forum raises the public’s expectations of the recovery effort, which in turn increases the not-inconsiderable pressure on aid workers on the ground. “These narratives are visible in near real time to donors and other stakeholders,” he writes, “who can exert pressure on NGOs and other members of humanitarian operations”. Despite these reservations, the UN Foundation study concludes that new, socially developed technologies will be a permanent fixture in disaster relief efforts from now on. Ultimately, it is the responsibility of humanitarian agencies to figure out how to make best use them. “Volunteer communities will continue to act and mobilise,” the report notes, “regardless of policies to foster or discourage them.” Blindspots in Pakistan That said, not all disaster relief efforts attract the same degree of volunteer technological support as the Haiti earthquake. In late July 2010, heavy monsoon rains caused extensive flooding across Pakistan, leaving nearly 2,000 people dead and causing $43 billion worth of damage to homes and infrastructure. According to some estimates, approximately one fifth of the country was under water at any given point during the crisis. Despite its gravity, the disaster did not mobilise volunteer technical communities to the same extent Haiti had done six months previously. “Volunteer organisations put in a huge amount of effort in Haiti, and unfortunately they used up a lot of their resources by the time Pakistan happened,” Crowley reports. The use of the same relief support tools was also hindered by Pakistan’s technological infrastructure “There isn’t anywhere near the level of mobile phone adoption in the flooded areas of Pakistan as there was in Haiti,” Crowley explains. He also notes that many of tools used in Haiti relied on the public availability of satellite imagery of the country. “In Pakistan, there was no open imagery,” he says. “It was not made publicly available by request of the government of Pakistan.” With no satellite imagery or detailed online maps, volunteer developers struggled to plot where relief was a priority, Crowley adds. Crowley says that, though undoubtedly tragic circumstances, both disasters have provided learning experiences, both for humanitarian agencies and for the volunteer developers that have now become a partner in their relief efforts. “A lot of learning has happened,” he concludes, “and hopefully it’s put us in a place where we can look at the next disaster and start wondering, ‘What can we do?’”
<urn:uuid:37ef6689-b911-48a6-9320-c5459c15d025>
CC-MAIN-2024-38
https://www.information-age.com/pulling-together-in-a-crisis-26141/
2024-09-12T16:27:18Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651460.54/warc/CC-MAIN-20240912142729-20240912172729-00725.warc.gz
en
0.963298
1,390
3.15625
3
This news will get your heart racing. We’ve heard of 3D printed toys, bird beaks, and prosthetic limbs and now we’ve got internal organs up on that list. A small team at ETH has designed a 3D printed artificial heart. The team was led by doctoral student Nicholas Cohrs and they believe that it is first of its kind. This only proves that the possibilities of what technology can achieve is being redefined every day. The techniques that go into making a robotic arm that flexes the way you need it to, is being made more complex to make internal organs function as they are supposed to. The team which built the artificial heart says that it is completely soft. It has a pumping mechanism that causes the silicon ventricles to pump like a real one. Technically speaking, it almost beats like a real one. Inside the heart, there just isn’t a wall for the ventricles, but a chamber that fills and deflates to create the pumping action. The complex structure was made using a 3D printer and the material used was soft and flexible, give it the desired results. The entire artificial heart is one structure, a monoblock. Hence, the question of how the different mechanisms fit together doesn’t arise. There is of course an input and output ports, where the blood comes and goes. The current state of this artificial heart isn’t ready for implantation. This heart doesn’t last up to more than a thousand heart beats that is about half an hour. Although it isn’t ready for implantation, this artificial heart has opened a wider arena as to what technology can do. It is a live proof of the concept. The researching and developing team has similar belief. “I’m now so fascinated by this research that I would very much like to continue working on the development of artificial hearts,” said Anastasious Petrou a grad student working on this project.
<urn:uuid:f00c1a9c-be87-4733-9946-258eb1a2f754>
CC-MAIN-2024-38
https://www.ciobulletin.com/healthcare/3d-printer-gets-your-heart-pumping
2024-09-13T21:00:41Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651540.48/warc/CC-MAIN-20240913201909-20240913231909-00625.warc.gz
en
0.952627
410
3.65625
4
Cyber threats are increasing in rate, variety and sophistication: UK businesses had to combat an average of 686,961 hacking attempts of their online systems over the course of 2020. In response, corporations, institutions and governments must make their IT security a firm priority to mitigate threats. High profile incidents keep occurring, from the Irish Health Service executive cyber breach, to British Airways’ £20m fine following a major GDPR breach, and the Colonial Pipeline’s costly ransomware attack. A discipline known as ‘ethical hacking’ can help organisations looking to bolster their security. It provides a way for organisations to authorise penetration testing to assess their IT system’s defences. For organisations, plugging these weak spots can be the difference between paying hundreds of thousands in ransom, avoiding days’ worth of business inactivity and/or maintaining the privacy of millions of sets of sensitive data. Having internal, dedicated IT security staff, fully trained in ethical hacking practices, will prepare organisations and help mitigate against future threats. How to boost internal cyber security training How can hackers be ethical – and why are they important? The word ‘hacker’ tends to conjure up sinister images, but what many people tend to overlook is the fact that hackers can actually be ‘good guys and gals’ too. These are typically referred to as ‘white hat hackers’ as opposed to ‘black hat hackers’ – that’s to say, the bad guys. While there is no official definition or set job description for an ethical hacker, they are computer and penetration experts with an in-depth knowledge of how operating systems, software, hardware, websites, networks, and humans work. They approach these elements from the perspective of malicious actors, with an owner or organisation’s permission, to identify, then secure and protect, system vulnerabilities. For businesses, employing ethical hackers is exactly the kind of proactive approach that will give them the advantage in combating cyber criminals and protecting their systems and subsequently their operations, sensitive data, and funds. This is especially true, as a large number of businesses fail to realise how much of an easy target they might be. According to Varonis, a data security and analytics company, companies only protect a terrifying 3% of their folders. Recognising their potential vulnerability, some organisations, like Apple and the US military, even offer rewards to those who can find and report vulnerabilities. Value and momentum in an evolving world As technology evolves, hackers and their techniques evolve in tandem; compared to 2019, 2020 saw ransomware attacks increase by 485%, and distributed denial of service (DDoS) attacks increase by 154%. By employing benevolent experts in the field, companies can stay abreast of these trends, which is an effective way to protect their systems. Many large scale businesses that are constantly under attack employ in-house white hat hackers that work full-time. This is a growing trend, as the new reality of remote working catalysed by the pandemic increases businesses’ target attack surface, as employees and devices are widely distributed rather than kept to single buildings and networks. These employees become links in a growing chain and are often unaware of the role they each play as gatekeepers to the business. According to Varonis, employees each have access to an average of 11 million files on their work devices, posing a substantial risk to any firm not actively taking steps to avoid cyber threats. Hackers are also increasingly targeting specific sectors. The percentage of businesses attacked in the key industries of technology, media and telecoms (56%), financial services (55%), and energy (54%) were up from 44%, 44%, and 40% respectively in 2020. Why are DDoS attacks becoming nasty, brutish and short? Challenges and opportunities for businesses Ethical hacking is an effective, authorised way to test security defences while also staying competitive and flexible by minimising damage from breaches, retaining customer confidence and loyalty. Indeed, firms qualifying as experts in a cyber readiness modelling report by Hiscox were less likely to suffer a ransomware attack, less likely to pay up, and recovered more quickly. However, the single biggest challenge for businesses looking to focus on developing internal ethical hacking capabilities is the growing skills gap; estimates suggest an additional 4 million cyber security professionals are needed globally in order to plug this. In the UK, a basic technical cyber security skills gap exists in 50% of private sector businesses, and one third have a more advanced technical skills gap, in areas including penetration testing. Attitudes towards ethical hacking are changing as hiring and training practices are opening up to a wider pool. Employers are beginning to realise that unconventional education paths in this domain are more commonplace and acceptable, so long as the overall job gets done. And with 81% of white hat hackers having learned the majority of their skills through self-directed online educational materials, university-style qualifications aren’t always the most effective. Businesses can be proactive in upskilling internal IT experts horizontally, building on existing IT and infrastructure knowledge and leveraging it in a security-based context. Nowadays, online project-based training providers like Udacity offer courses for people to hone various cyber security skills; even providing a ‘Nanodegree’ specifically dedicated to developing ethical hacking skills. A few steps ahead Many businesses incorrectly assume that hacking is a wholly negative practice. However, this misconception ignores the large-scale work that white hat hackers carry out defending against potential and incoming attacks from the black hats. By embracing ethical hacking capabilities – whether externally, or internally via upskilling existing IT experts appropriately – businesses can put themselves a few steps ahead. Dynamic cyber security measures that include an ethical hacking approach can both anticipate and mitigate damaging attacks that are an evolving reality in today’s digital world – and protect a business’ data, operational infrastructure, finances and reputation going forward.
<urn:uuid:7217efa3-d415-4995-93c5-5730247ecd30>
CC-MAIN-2024-38
https://www.information-age.com/why-organisations-need-to-invest-in-ethical-hacking-capabilities-19041/
2024-09-13T22:21:20Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651540.48/warc/CC-MAIN-20240913201909-20240913231909-00625.warc.gz
en
0.954826
1,208
2.9375
3
We have been hearing about the improvements that the benefits of 5G will bring almost since the 4G network was implemented. Well, we can already say that it is a reality. Today, the 5G network is implemented in many countries and although it is not finalised, many of them already enjoy this connectivity. «For anything to happen, something has to be done first.» It may be thought that the main improvements of 5G will come from the technical side, but in fact the main improvement will be from the technological point of view. This is what is called in the telecommunications sector ”emerging telecom technologies” where the fifth generation network will enable other technologies to increase, helping organisations to progress. There is no doubt that the fifth-generation network will bring great advances by which many companies will be able to increase their performance. If you’re a little out of the loop, this is the place to get up to speed. What is 5G? 5G can be defined as the network that is in a constant process of implementation. Among all its qualities, one of the most relevant is its ability to adapt to devices. Our daily lives will change, 5G will not only adapt to smartphones, but also to all those objects connected thanks to the Internet of Things. This breakthrough will help increase smart devices as well as intelligent infrastructures, multiply the data collection and machine learning capabilities of infrastructures, and will also increase the automation of processes such as product manufacturing, automated car driving or tele-assisted surgery. Benefits of 5G The 5G network will enable the full development of other technologies such as IoT,artificial intelligence, cloud computing, virtual and augmented reality, or data analysis to develop fully, thanks to low latency and higher bandwidth. It is estimated that the fifth-generation network allows 1,000 times more data transmission than the 4G network. Some of the benefits of 5G are: - Higher transmission speeds of up to 15-20 GB per second. - Lower latency will make the reaction and execution time of orders ten times shorter than the current one. - Better connectivity as 50 billion smart objects can be interconnected through the Internet of Things. Although not only in intelligent objects there will be an improvement in connectivity, also in all those devices adapted to the 5G network, making, for example, remote working benefit and have a better implementation. - It will lead to advances in augmented and virtual reality. - Increase in productivity, since by increasing the speed of data transmission it will be possible to obtain an improvement in the control of productivity. However, there are also concerns about the deployment of 5G networks that are being studied and addressed over time. One of these concerns is power consumption. 5G technology will need more energy in order to achieve this efficiency. Impact of 5G on business Leaving aside the less favourable aspects, we can say that the benefits of 5G will generate more advances than setbacks, especially in industry by transforming the way it manages its processes. The telecommunications sector will also be favourably altered, as it will be possible to manage processes digitally and remotely by automating systems. An example would he access to an infrastructure or site whereby a tool can be used to digitally plan and automate access to it, facilitating site maintenance. The impact of 5G on businesses will go beyond this, however, as it will become essential, enabling organisations to evolve to another level. As mentioned above, 5G will enable many objects such as machinery, infrastructure, automobiles and devices to progress to such a level that it will offer companies numerous opportunities for improvement. Opportunities such as the automation of processes, new sectors that require this technology, such as Smart Cities, improvements in the customer sales experience with augmented or virtual reality, or even a more optimal management of each asset or process of a company. Get to know one of our modules: TREE.fibre The scenario posed by the arrival of 5G requires fluidity and coherence maintained over time, as well as speed of response and tracking of metrics. We seek continuous improvement of 5G-derived asset processes, eliminating the conceptual boundaries that have separated radiant assets from underground assets and keeping your telecommunications network visible at all time. To this end, we have created TREE.fibre, designed to respond to the growing trend to replace conventional information distribution networks with high-speed, high-technology networks: Optical Fibre, equipped with a set of tools that integrates and relates various components (users, hardware, software, processes) that allow any optical fibre line to be geographically located simply by tracing the location on a map and TREE takes care of attributing the rest of the necessary parameters, thus solving the most fundamental difference between radiant assets and optical fibre networks. To learn more about TREE.fibre or how to help your company manage its telecommunications processes, you can contact us and request a Demo.
<urn:uuid:d0e93fcc-c15c-4ab0-96a5-24940e526757>
CC-MAIN-2024-38
https://www.atrebo.com/en/benefits-of-5g-network-in-telecommunications-infrastructure-management/
2024-09-16T10:27:16Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651682.69/warc/CC-MAIN-20240916080220-20240916110220-00425.warc.gz
en
0.949671
1,029
2.828125
3
What is Cloud Computing? Cloud computing is a revolutionary technology model that enables on-demand access to shared networks, servers, storage, applications, and services, collectively referred to as "the cloud." With the rise of big data, AI, and advanced analytics, cloud computing has become the cornerstone of digital businesses, propelling digital transformation and offering unprecedented scalability and flexibility in operations. Cloud Computing Terminology Cloud Service Models Infrastructure as a Service (IaaS): IaaS provides users with virtualized computing resources over the a network. This model eliminates the expense and complexity of buying and managing physical servers and data center infrastructure. IaaS is beneficial for businesses that want to build applications from scratch while maintaining control over most of the IT components. Platform as a Service (PaaS): PaaS is a complete development and deployment environment in the cloud. It offers a platform that includes both hardware and software tools needed for application development and management. PaaS is suitable for developers who want to spend their time coding and testing applications rather than managing hardware and infrastructure. Software as a Service (SaaS): SaaS delivers software applications over the internet on a subscription basis. Users can access these applications typically through their web browser, eliminating the need for installations or running a server. SaaS is ideal for end-users who need to use specific software but do not want to worry about infrastructure or system maintenance. Cloud Deployment Models Public Cloud: Public clouds are owned and operated by third-party service providers. Customers benefit from massive scalability and share the same hardware, storage, and network devices with other cloud users. They are highly efficient and economical due to their larger scale. Private Cloud: A private cloud is exclusively used by a single business or organization. They provide more control over data, security, and quality of service. They are ideal for businesses with dynamic or unpredictable computing needs that require direct control over their environments. A private cloud typically is created and run within the “four walls” of an enterprise. Hybrid Cloud: Hybrid clouds combine public and private clouds, bound together by technology that allows data and applications to be shared between them. Hybrid clouds give businesses the flexibility of in-house applications with the cost benefits of public clouds. Multi-Cloud: Multi-cloud is the use of multiple cloud computing and storage services in a single heterogeneous architecture. This strategy prevents data loss and downtime due to a localized component failure, and can allow users to combine the best services among different cloud providers. Key Cloud Computing Concepts Virtualization: Virtualization is the process of creating a virtual version of something, like storage devices, network resources, or an operating system. This technique enhances the efficiency of IT resources, reduces overhead costs, and makes cloud computing possible. Scalability: Scalability in cloud computing is the ability to easily add or subtract resources to applications based on demand. It ensures that businesses can grow without worrying about infrastructure capacity. Redundancy: Redundancy is the duplication of critical components of a system with the intention of increasing reliability of the system and ensuring business continuity. Bare Metal: Bare Metal refers to physical servers in cloud computing, without virtual layers. It allows full control over the server's hardware, ideal for high-performance needs and environments where virtualization is not desired. How Supermicro Enables Cloud Computing Supermicro is a global leader in providing high-performance, high-efficiency server technology, and advanced IT/hardware solutions. We offer a comprehensive suite of products, including Multi-Node Solutions, CloudDC, Hyper, rack mount optmized servers , and SuperStorage Systems, that are tailor-made for cloud environments. Our robust solutions support virtualization, offer massive scalability, and are built with redundancy in mind to ensure uninterrupted services. Benefits of Cloud Computing with Supermicro Choosing Supermicro's solutions for cloud computing, whether creating a cloud offering or creating a private cloud offers numerous benefits. Our energy-efficient products ensure you save on power costs while achieving superior performance. Our servers are designed to be scalable, allowing for easy expansion of your cloud environment as your business grows. With our focus on high-density design, you get maximum performance per cubic foot, reducing your data center footprint. In addition, Supermicro's cloud solutions ensure robust security to protect your data and applications. Frequently Asked Questions about Cloud Computing - What is the importance of cloud computing in modern businesses? Cloud computing provides businesses with the ability to quickly scale up or down, provides cost efficiencies, enhances collaboration, and ensures business continuity. - How does Supermicro support cloud services? Supermicro designs, develops, and provides hardware solutions optimized for cloud services, facilitating efficient data management, fast application processing, and robust security protocols. - What distinguishes Supermicro's cloud solutions? Supermicro's cloud solutions are recognized for their energy efficiency, superior performance, and unrivaled scalability, making them the preferred choice for building cloud infrastructure. With a very broad range of servers, any type of cloud can be created. Understanding cloud computing is key to leveraging its potential. As an industry-leading provider of server technology and advanced IT/hardware solutions, Supermicro is committed to helping businesses of all sizes navigate the cloud computing landscape. We invite you to explore our extensive product line and find your ideal cloud solution today.
<urn:uuid:86d8ae6a-b5f6-4524-b92f-7f523e557b00>
CC-MAIN-2024-38
https://www.supermicro.org.cn/en/glossary/cloud-computing
2024-09-16T10:07:12Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651682.69/warc/CC-MAIN-20240916080220-20240916110220-00425.warc.gz
en
0.918374
1,101
3.40625
3
Information is passed from the client to the HLL exit address space for processing and sending on to the user exit itself. If the exit is designated as a REXX exit, REXX variables are defined and populated with values from the incoming request. If the exit is an LE-program, the incoming data is formatted in such a way that a (supplied) COBOL and/or PL/I copybook can be used to map the data. On entry to an exit, a field (xxxxORGN for LE code and callOrigin as a REXX variable) will be set to an identifier that identifies the client process that resulted in this call. These identifiers are: Identifier | Description | ECL | The call is as a result of a ZMF4ECL client process. | ZDD | The call is as a result of a ZDD client process. | SPF | The call is as a result of an ISPF client process. | XML | The call is due to a direct invocation of an XML service (for example, Through XMLSERV or equivalent). | The exit has the opportunity to update that data in place. If it does so, it must set the dataChanged variable (or LE equivalent) to YES; otherwise ZMF will ignore it. If an exit wishes to stop a process, it can set the proceed variable to NO. It can also populate the shortMsg and longMsg variables to whatever is required to be displayed for the condition leading to the stopping of the current function. Furthermore, it can set the cursorField variable to position the cursor at a specific field. All normal execution of HLL exit routines must end with RC=0 (note that EXIT with no expression is treated as EXIT 0 by HLLX). If the routine ends with RC>0, the infrastructure will take this as a major failure and abandon the current function altogether. In this case, the client will produce a general message indicating an HLL exit failure and will direct the user to the HLLX started task output for further details. (A developer should contact his or her administrator at this point.) For an LE program we see something like this in sysout: LE program for exit function PCRE0007 finished with RC=00000007 And for a REXX exec we have something like: REXX RC for exit function PCRE0101 is 00000000 REXX evaldata (expression coded on EXIT statement) is 9 On return, the HLL exit address space extracts and populates the data in the response section of the XML service request, which gets passed back to the client to deal with as it wishes.
<urn:uuid:1b7d6af9-ec9b-4946-b58a-7f32b4127128>
CC-MAIN-2024-38
https://www.microfocus.com/documentation/changeman-zmf/8-3/main-guides/hllx-gsg/exit-interface/index.html
2024-09-20T03:10:54Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652130.6/warc/CC-MAIN-20240920022257-20240920052257-00125.warc.gz
en
0.916221
554
2.734375
3
Edge Technology Enables Smarter Cities Smart cities are part of our digital future, and they are unfolding across the world right now. For example in India, the National Smart Cities Mission was launched in 2015 with a mission to integrate technology into municipalities with the goal of making them more, “citizen-friendly and sustainable.” A city becomes “smarter” by bringing together new technologies such as devices from the Internet of Things (IoT), apps that use BigData creatively, or artificial intelligence (Ai) and machine learning to change the way we do business in urban centers. By using different types of digital communication and different sensors around the city, we can collect and analyze data for real-time management that can bring insight into how we live in our cities, interact with businesses, and eventually help citizens make better decisions. There are plenty of ways that cities can benefit from applying smart technology. Some examples include improving energy distribution, streamlining trash collection, reducing traffic congestion, improving air quality, and delivering life-saving services with greater efficiency. Smartphones, connected devices, and technology innovation help accomplish these benefits. But the real game-changer, is the ability to managing urban resources with digitally enhanced efficiency. The emerging tech listed above, combined with 5G and mobile smartphone adoption, is only the beginning. Taking advantage of the tools that power smarter cities ensures better preparedness to meet urban challenges and open possibilities in our growing cities. Factors for Smart Technology Success As each smart city begins to harness the power of its data to support greater efficiency and intelligence initiatives, every company and every town will now be data-centric and network-centric. This creates a new and advanced role for data centers, given that lots of connectivity, data storage, and compute power to crunch out useable data analytics will be necessary. However, the centralized and distant storage solution of the past will not suffice in this new arena. The data-intensive capabilities emerging with the latest technology trends demand sophisticated data center infrastructure, making today’s data center a keystone feature in the dynamic smart city network. Solutions will require: - Immense data storage capability to drive predictive analytics - Compute power for processing data-driven applications - Constant high bandwidth connectivity - Reliable interconnection across a broad network to smooth gaps between IoT devices and 5G networks - Instantaneous data exchange between business and government/municipal entities - Edge proximity for low latency and real-time responsiveness Of course, every city is different. Data storage and edge computing solutions will vary for densely packed cities like New York Cities compared to sprawling urban metros like Atlanta. The infrastructure to power data center services, geography, and climate are all factors that will impact the type of data center needed to fulfill its vital role in intelligent cities. However, power and real estate will no longer be the driving factors determining data center location. The smart city processing requirements will require smaller, decentralized edge data centers situated at prime locations. In this way, they are flexible enough to accommodate usage demands with minimal latency, while delivering the perfect efficiency for which they are designed. From server colocation to edge computing, innovative developers need to make the right investments in data center infrastructure to enable the advantages of their smart city initiatives and applications. In some of America’s most innovative markets, that transformation is already underway with use cases that maximize the full potential of data centers on the edge. Download our ebook: Navigating the Data Center Ecosystem to learn more about why location matters all the way to the edge.
<urn:uuid:354b7b4a-be62-4ec4-9b0b-4af606de685e>
CC-MAIN-2024-38
https://elementcritical.com/blog/edge-technology-enables-smarter-cities/
2024-09-08T00:08:01Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650926.21/warc/CC-MAIN-20240907225010-20240908015010-00325.warc.gz
en
0.912154
728
2.953125
3
Point-of-sale (POS) systems are one of the most common pieces of technology in modern life, yet we don’t often stop to think about just how thoroughly they changed the nature of business transactions. What once required detailed ledger entries and mountains of paper records quickly became a streamlined operation thanks to the introduction of cash registers and bar codes. With the help of POS technology, store clerks could easily complete sales for the customer and update inventory with the push of a few buttons. Understanding the beginnings of POS technology can illuminate the state of the systems today, and give us a glimpse of what might be in store for the future of POS systems. The early years of POS Systems The earliest examples of what we think of as POS systems were simple mechanical cash registers that allowed a shopkeeper to log a sale and open the register by pulling a lever. These became commonplace around the middle of the twentieth century, and quickly led to the rise of computerized cash registers that were introduced in the 1970s. The most significant technological leap made possible by these new systems was the scanning of bar codes at the time of a sale. This allowed the business to seamlessly keep track of inventory with each transaction that was made. From there, the next milestones arrived in the following decades as POS systems were tied to local area networks (LANs) and personal computers. These developments allowed multiple POS kiosks to be installed in a store that were all tied to a single software program and connected by a network. The current state of POS technology Current POS technology goes far beyond the original sales transaction features that were offered in the beginning. Today, many businesses are turning to all-in-one POS software solutions that integrate services such as accounting, customer relationship management, payroll, and operations in one system. Expanded network access has made it even easier for retail stores to setup self-service POS kiosks throughout the store, which can handle transactions such as barcode and coupon scanning, payments, and receipt printing without the need for employee assistance. One of the most notable innovations in the POS system market has been the introduction of individual credit card scanners than can be attached to a smartphone or other wireless device. With these readers, any small business can instantly accept credit card payments for transactions no matter where they are, and do so without the need for expensive hardware. What does the future hold for POS hardware and software? Mobile technology is poised to continue to drive innovation in the POS technology sector, as many business will need to upgrade to compatible systems that will handle mobile wallet transactions in order to remain competitive. Software will continue to increase in sophistication, further facilitating the rise of tablet POS systems that can integrate data on the spot. It’s clear that mobile broadband networks and the cloud will continue to play an important role in the development of POS technology. If you need assistance with your POS Systems, old or new, and POS repair, Great Lakes Computer can help.
<urn:uuid:9c6c20c5-f706-4fcf-ac4c-3f54653e19e1>
CC-MAIN-2024-38
https://greatlakescomputer.com/pos-systems-past-present-and-the-future/
2024-09-10T10:38:46Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651241.17/warc/CC-MAIN-20240910093422-20240910123422-00125.warc.gz
en
0.962434
597
2.6875
3
As we celebrate National Student Employment Week, it’s crucial to dive deep into the transformative influence of artificial intelligence (AI) and analytics on the job landscape for students. Gone are the days when securing a job or internship was solely about crafting the perfect resume and acing the interview. Today, AI and big data are revolutionizing the employment scene, offering new perspectives and opportunities for students around the globe. The Power of Data-Driven Insights In a world where data is king, analytics have become a vital tool for enhancing student employability. AI-driven platforms are capable of sifting through massive datasets to match students with potential employers based on a myriad of factors, from academic achievements and extracurricular activities to nuanced skill sets. Such precision not only streamlines the job search process but also increases the likelihood of a great fit for both parties. Moreover, the use of analytics helps in identifying employment trends, skill gaps, and salaries across different sectors. Students, armed with this information, can make informed decisions, tailoring their studies and gaining relevant experience to meet market demands. Emergence of AI-Centric Job Roles As businesses continue to embrace AI, new career paths are emerging. Students are now pursuing roles like data scientists, AI ethicists, and machine learning engineers – positions that were barely on the map a decade ago. Educational institutions are accordingly adapting their curricula, ensuring they produce graduates who are not just AI-savvy but also poised to contribute to the AI revolution. Ethical Considerations in Algorithmic Decision Making The reliance on algorithms for job placements raises valid questions about fairness and privacy. Biased algorithms could perpetuate inequalities, and inadvertently discriminate against certain groups. It’s imperative for developers and companies to ensure that their AI solutions are transparent and equitable, to avoid such pitfalls. The Collaboration Between Academia and Industry Preparing students for a data-centric workforce requires concerted efforts from both educational institutions and employers. Internships and cooperative programs are becoming increasingly data-driven, with performance metrics and project outcomes closely monitored and analyzed to enhance the educational experience. Balancing Human and Machine Intelligence While AI offers efficiency and insights, the human touch remains irreplaceable. Cultivating soft skills such as leadership, communication, and creativity is essential. Human intuition, empathy, and the ability to navigate complex social dynamics are still at the core of successful professional ventures, and these are areas where AI cannot fully tread. As we gaze into the future, the interplay between human intuition and machine intelligence will continue to evolve. The students of today are the professionals of tomorrow, navigating an AI-infused employment landscape. It’s a thrilling time for the next generation to harness the power of AI and analytics, not only to excel in employment but also to pioneer a future where technology and humanity converge for greater societal good. The conversation on these developments is not just timely; it’s imperative for shaping a workforce that is ethical, adaptable, and equipped for the challenges ahead.
<urn:uuid:320fab00-508f-4cc5-a7c0-ebf379dab965>
CC-MAIN-2024-38
https://analyticsweek.com/charting-new-horizons-ais-integral-role-in-shaping-student-employment-pathways/
2024-09-20T07:50:45Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652138.47/warc/CC-MAIN-20240920054402-20240920084402-00225.warc.gz
en
0.937776
618
2.546875
3
The ethernet cables are usually preferred by gamers or people who don’t want signal interferences. This is because the ethernet cables promise unhindered signals, and the signals haven’t interfered. However, people often wonder, “does a longer ethernet cable slow internet speed?” So, let’s find out how true is it! Does A Longer Ethernet Cable Slow Internet Speed? For the most part, the long ethernet cables can lead to slow internet connection since it increases the latency. In addition, it increases the distance that the signal has to travel to reach the end-user. On the contrary, it’s essential to note down that ethernet signals tend to travel at 2/3rd the light speed. Therefore, the users might not notice the difference in internet speed, be it the short cable or longer. Generally, it’s safe to say that the ethernet cable’s length will not matter when it concerns the length. Still, there is a limit to which the single cable’s length can be increased. Suitable Length Of The Ethernet Cable Truth be told, there is no particular limit for the length of ethernet cable for a single run. However, the length of the ethernet cable directly influences the latency. For instance, if the length of the ethernet cable is increased, it will consequently increase the latency. This is because the internet signals will need to travel farther away. So, the farther the signal has to travel, the more the chances of interferences. There are still limited chances that internet speed will be influenced (especially with the shorter cables). For instance, if you switch from a 10m ethernet cable to a 20m ethernet cable, the users won’t be able to tell the difference in the internet speed. Although the length of the ethernet cable is increased twice, the users wouldn’t be able to outline differences in the internet speed. There will be a slight loss of signals with longer lengths of cable, but it’s still not drastic. This is because the minor changes in length will not result in connection issues. However, the users might feel drastic internet speed issues if the cable length exceeds 100m. The single ethernet cable run is designed to work optimally up to 328ft (or 100m approximately). When the length surpasses this count, the internet signals start weakening and will reduce the internet passed. In addition, it will impact the reliability of the internet connection. For the people using old ethernet cables (CAT5), the internet speed up to 100Mbps will be optimal for up to 100m. So, if the ethernet cable length is increased beyond it, the speed will reduce to 10Mbps. To be honest, the ethernet cable above 100m will work, but the connectivity issues will increase. This is the prime reason that manufacturers don’t suggest going beyond this length while using the ethernet connection. To summarize, the internet speed will be optimal and reliable if the ethernet cable run doesn’t exceed the 100m length (minor latency issues might incur). Transmission Speed Of Ethernet Signals It’s pretty evident that users don’t have to worry about the ethernet cable speed with 20m cable (if they were using 10m ethernet cable previously). The internet or ethernet signals travel at 2/3rd the speed of light, which makes it around 200km/millisecond. With this being said, there will be no issues with the internet connection as long as you don’t exceed the ethernet cable length. All in all, the minor increase in ethernet cable won’t impact the internet connection. Suitable Length Of Ethernet Cable To be honest, the length of the ethernet cable doesn’t impact the internet speed drastically as long as the single run doesn’t exceed 100m length. If the length of the ethernet cable is increased by this count, there is a potential for signal loss because the signals will have to travel a long distance. In addition, when you are installing the ethernet cable, make sure the installation is tight. Upgrading The Ethernet Cable Length For the people who need to transfer files and data, upgrading the ethernet cable will be better. However, the internet connection will be directly influenced by the quality of the ethernet cable. Usually, people opt for Cat-5e cables, but the Cat-6 cables have better quality. Also, the Cat-7 cables are top-notch but can be pretty expensive.
<urn:uuid:a095335b-6b91-481e-9016-75d8860a4cab>
CC-MAIN-2024-38
https://internet-access-guide.com/does-a-longer-ethernet-cable-slow-internet-speed/
2024-09-20T07:39:23Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652138.47/warc/CC-MAIN-20240920054402-20240920084402-00225.warc.gz
en
0.899941
950
2.90625
3
As the technology of using QR Codes is growing in use and popularity, hackers are seen embedding malicious URLs tucked with customized malware into QR Codes. Thus, allowing them to exfiltrate data from the victimized device when scanned. Or, diverting them to a malicious web portal that makes them divulge credentials, making the hacker gain an upper hand. A recent survey conducted by MobileIron discovered that 71% of respondents responded that they cannot differentiate between a malicious or legitimate QR Code. And only 13% of them know that a Quick Response code can open up an URL to victims, and 20% believe that a scanning code can draft an email or connect a phone call to initiate a text message. The highlight of this survey is that 51% of respondents know that QR codes do not exhibit privacy, security, and render financial help- but still use as they ease their digital lives to a certain extent. So, how to protect from Malicious QR Codes Just watch them closely and see if a hacker/s has posted a fake Quick Response code on a legitimate poster. And if you suspect something fishy, just avoid scanning the code. Never- Ever share your details via QR codes as it can land you into trouble as it can be a code loaded on a phishing page and entering sensitive info can land you into deep trouble. Most of the QR code scanner asks the victim about a possible URL they could be visiting after scanning the code and this allows the visitor to think before visiting the malicious webpage.
<urn:uuid:c6633883-cf22-4588-b616-2a217d7ccd6b>
CC-MAIN-2024-38
https://www.cybersecurity-insiders.com/cyber-threat-from-qr-codes/
2024-09-09T09:38:34Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651092.31/warc/CC-MAIN-20240909071529-20240909101529-00325.warc.gz
en
0.921554
307
2.515625
3
How Development Skills Translate to Reverse Engineering: Using LLDB to Reverse Engineer an Android App For developers, reverse engineering is easier than they might think. Attackers use debuggers to understand and tamper with an app’s logic at runtime, iOS and Android developers use the debugger without realizing how the exact same capabilities are used for reverse engineering. To get a better idea of how developers and attackers use debuggers, let’s look at some of the common use cases for debugger features. Debugger Feature | Developer | Attacker | Breakpoints | Pause to analyze a bug | Pause to understand a specific part of code of an app | Memory Watchpoints | Watch/monitor variable value during program execution | Determine which functions modify a certain variable (e.g. Find where the amount of money is subtracted from) | Memory Read | Look up variables and arguments to verify an algorithm | Look up variables and arguments to understand the logic of a function | Memory Write | Modify a variable to check the results of an algorithm under different conditions | Overwrite a variable’s value (e.g. change the amount of money in a game or unlock an in-app purchase) | Stack Backtrace | Find out the code path to a function to make sure it’s invoked correctly | Find out the caller of a function, so that an attacker knows what to analyze next | Conditional Breakpoints | Optimize time spent on manual look-ups | Pause to find out how specific variables or arguments influence the execution | To demonstrate how a debugger can be used to reverse engineering and tamper with a mobile application, let’s bring our adversary: OWASP’s Android Uncrackable Level 1 into the ring! This training app performs Java based root detection. We will attach LLDB, a native debugger, to the app and show you how simple LLDB commands can quickly bend the app’s functionality. Spoiler Alert: The program’s Java based root detection can be bypassed with a single line in LLDB. On top of that, it can also be used on iOS in the exact same way. Cracking uncrackable Level 1 Uncrackable Level 1 is a training app created by the OWASP foundation to teach people about analyzing mobile apps. The app contains a hardcoded secret string inside that users must try to extract. To make it more challenging for the reverse engineers, the app features root detection. Therefore, whenever you open the app on a rooted device, you will immediately see an error message complaining about the rooted device, afterwich the app is exited automatically. What is root detection? The most popular way to detect a rooted Android device is to search for thesu binary on the filesystem. This file should not be present on non-rooted Android devices to make sure no app or user has higher privileges. Usually, the file is located at/system/bin/su but its location can change depending on the rooting tool, manufacturer and the OS version of the device. To bypass this check, we need to trick Java APIs into believing that thesu file doesn’t exist, even though it’s present on the filesystem. This will allow us to bypass the app's root detection. Our plan of attack Uncrackable Level 1 is entirely written in Java, so the app can check for thesu binary using a number of APIs: - … and probably many more … Now you might wonder how on earth it is possible to hook Java functions with LLDB (which is a debugger for native code). The answer is simple: you can’t…. However, instead of hooking Java, we can put a breakpoint at the native layer to achieve the same. Because at the very bottom of our software stack no one speaks “Java”, the interface will eventually be native. For example, opening a file in Java or Kotlin source code triggers an eventual native libcfopen call that tells the kernel what to do; this is the point where we will intercept. You can think of this approach as a “sink”, where all Java’s API calls pour into a native-land drain hole where we will intercept them one by one. access(const char *pathname, int mode) is exactly such a sink. It's a standard way of checking whether a file exists on a filesystem by the Java engine. In fact, anytime we use a Java API to create, delete or rename a file, the standard library will useaccess(..) to check if it exists beforehand. So, by hooking a single native function, we can intercept multiple Java calls to the system. This means that it doesn’t really matter how exotic the Java API you use to check for a file's existence is, because all of them will eventually end up at the same place - libc. LLDB - the Android Studio debugger Android Studio and Xcode use LLDB under the hood as native code debugger, which developers can interact with via a command line or a fancy GUI, which basically runs LLDB commands for them. In this article, we will use the command line debugger directly because it allows for easy customization but a determined user could perform all the steps from within a GUI as well. A typical use case for a debugger is to insert breakpoints at specific functions and access the value of variables there. We will show you how you can use this same LLDB feature to bypass root detection in the OWASP Crackme. We will insert a breakpoint atlibc’access and look up the file path argument. If the file path contains thesu string, we will simply replace one letter of the file path. This directs the code to check a different (non-existent) file instead of thesu Inserting a breakpoint After attaching LLDB to the app, you can insert a breakpoint atlibc’access(char*, int) with the following command: (lldb)$ breakpoint set -n access Now, whenever we use any Java function to open/check/remove a file, this breakpoint will trigger. The breakpoint will allow us to analyze the function execution with additional LLDB commands. We need to ensure that the app’s root detection won’t find anysu binary on the file-system. To accomplish this, we’ll use conditional breakpoints. With a conditional breakpoint, we can instruct LLDB to pause at a specific function only if a certain argument satisfies our constraints. For our purposes, the constraint is that a file path must contain asu substring, since we don’t care about any other file paths. In LLDB, you can add a condition to your breakpoint with the-c flag. LLDB supports the use of C/C++/Objective-C/Swift languages for scripting, so we can use thestrstr(char* string, char *substring) function to check whethersu is in the provided path.If thesu is in the provided path, thestrstr() function returns a pointer to the beginning of the substring, otherwise, it returns zero. The file path is the first argument oflibc’access(char*, int) , so we can find it in thex0 register. The updated command looks like this: (lldb)$ breakpoint set -n access -c "(char*)strstr($x0, \"su\") != 0x0" With this in place, LLDB will pause wheneveraccess(..) is invoked and the file path containssu Modifying arguments on the fly Finally, we need to replace a letter in thesu file paths to make sure thataccess(...) won’t find those files, which in turn bypasses the root detection of the app When a breakpoint is hit, users can instruct LLDB to execute some code. This is done by using the-C flag. In this example, we will replace the first letter of the file path with an ASCII “0”. LLDB allows you to write to memory withmemory write "address" "new-value" . We will use this expression with the-C flag to execute it when our conditional breakpoint is hit. (lldb)$ breakpoint set -n access -c "(char*)strstr($x0, \"su\") != 0x0" -C "mem write $x0 0x30" To make the whole script more transparent, we can make it print all the file paths that match the condition and will be modified by having LLDB execute one more expression for us. We can add another-C flag with mem read-f -s $x0 which will read the path passed toaccess(...) and print it for us. The screenshot below shows the output of LLDB’s combined command when used on Uncrackable Level 1. Using these commands, we not only bypassed the root detection of the app, but also revealed all file paths that were checked. Remember the error message that originally appeared in the app when it detected a rooted device? It’s gone! The screenshot below shows the application opened without an alert, which means that we successfully bypassed the root detection feature! Other RE tools vs debuggers Reverse engineering can be a daunting subject for developers. Many developers don’t realize they already dispose of a great set of skills that translate very well to reverse engineering; debugging their code. In this blog post we explored the parallels between using LLDB to debug an application and leveraging it for reverse engineering. We showed the potential of LLDB as a reverse engineering toolbox, even for Java applications without any actual native code. By considering that all Java APIs that interact with a system eventually end up in the native world, we can actually speed up the whole tampering process significantly. Rather than having to deal with all possible Java APIs, we can hook a native library at a single location and catch all calls coming from Java that interact with the system. Obviously, the Uncrackable Level 1 challenge is a basic example. If an app requires more complicated tampering, using LLDB’s command line interface may become cumbersome. When this happens, LLDB Python Scripting is a great alternative; it allows you to create more advanced scripts. A second takeaway for developers is to keep in mind that even when your code is not native, eventually any high level language API call that interacts with the system has to go through native system libraries. Therefore it’s important to not only protect your code from malicious tools and techniques, but to also consider native tooling that can be used for attacks on lower levels. What can you do? Here are a few additional tips you can use to protect your app against debuggers: - At runtime, check if there is any type of debugger attached to your app. - Make sure that an attacker doesn’t replace the app’s Manifest or Entitlements because that would allow them to debug the app on a non-rooted, non-jailbroken device. - On a rooted device, an attacker doesn’t need to modify your app to attach a debugger, so you might want to implement root/jailbreak detections as an additional protection. - Keep your sensitive variables encrypted. This prevents an attacker from easily finding them in the app’s memory and using a debugger watchpoint to locate the functions that update them.
<urn:uuid:75b8b654-841c-487e-a2ab-c47d6c0576b0>
CC-MAIN-2024-38
https://www.guardsquare.com/blog/how-development-skills-translate-to-reverse-engineering
2024-09-13T02:29:19Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651506.7/warc/CC-MAIN-20240913002450-20240913032450-00025.warc.gz
en
0.896688
2,432
2.984375
3
As mobile applications continue to grow as an attack vector, organizations will need to prioritize mobile app security to fortify their overall security posture. Both static attacks based on the source code itself and dynamic attacks that exploit app functionality are constantly evolving. That’s why understanding and remediating the most common mobile app vulnerabilities is crucial for mobile development teams. An important resource for identifying security vulnerabilities and implementing mobile security best-practices is the Open Web Application Security Project (OWASP). While OWASP is best known for publishing insights into security vulnerabilities within web applications, the foundation also has a mobile security project focused on the mobile app industry. OWASP has published research into the top mobile security threats and best-practices for defending against them. Here’s a brief overview of their most recent top 10 mobile risks list, as well as an in-depth resource for how app developers can navigate them. 1. Improper Platform Usage Improper platform usage occurs when developers fail to use certain system features correctly or at all, whether it’s on an Android, iOS, or Windows operating system. This can include a failure to use well-documented security guidelines or misuse of certain platform APIs. 2. Insecure Data Storage Insecure data storage is a common issue that leads to data leakage. If developers fail to encrypt or securely store data, malicious actors could access sensitive personally identifiable information. This usually occurs when developers incorrectly assume users or malware cannot access certain device or system files. 3. Insecure Communication Insecure communication occurs when transmissions over the public Internet or mobile carrier network expose sensitive data to attack. Most mobile apps leverage backend systems for data storage and resource-intensive functionality, but this creates an attack vector that may allow hackers to eavesdrop or intercept unsecured communications. 4. Insecure Authentication Insecure authentication allows hackers to fake or bypass identity management systems to access private data and sensitive app functionality. If app developers cannot properly verify the identity of users, they also cannot trace back any exploits to certain user accounts. Along with strong authentication during login, developers need to ensure that they’re continuously authenticating users throughout the entire session. 5. Insufficient Cryptography Insufficient cryptography means that developers have failed to implement security best-practices when using encryption. For example, hackers can exploit a weak encryption algorithm or poor encryption process to decrypt sensitive data. This includes not only private keys and passwords, but also the application code itself. 6. Insecure Authorization Insecure authorization occurs when hackers are able to bypass permission controls to access sensitive functionality reserved for administrators or other higher-level users. Oftentimes, malicious actors are able to authenticate as a legitimate user and then exploit the authorization scheme to execute privileged app functionality. 7. Client Code Quality Client code quality issues arise when third-parties can pass untrusted code as inputs that the app will execute. While this isn’t always a security vulnerability, hackers can exploit code quality issues to execute malicious code as well. Static analysis tools can often detect these poor coding practices such as buffer overflows, memory leaks, and more. 8. Code Tampering Code tampering means that malicious actors have modified the source code, changed resources within the application package, or redirected API calls to change the app’s behavior. Attackers may use code tampering to inject malware into repackaged apps and release them in mobile app stores. That’s why it’s crucial that applications attempt to detect tampering during runtime. 9. Reverse Engineering Reverse engineering is a method attackers use to understand how an app works to formulate exploits. They’ll often use automated tools to decrypt the application binary and recreate the source code. The best way to prevent reverse engineering is through code obfuscation, which makes it difficult for humans and software to understand the inner workings of an app. 10. Extraneous Functionality Extraneous functionality is any feature or code that isn’t directly exposed to users. For example, many developers leave additional code related to staging environments or unofficial API endpoints for testing that could inadvertently expose backend systems. Attackers can examine logs, configuration files, and application binaries to discover this hidden functionality that may be exploitable. Defending Against the OWASP Mobile Top 10 While this mobile security risk list may seem overwhelming, the majority of these issues can be defended against using runtime application self-protection (RASP), code hardening, and secure coding best-practices. In our latest report, we analyzed how these common mobile security issues mapped directly to security techniques. Key takeaways include: - In-depth insights into the OWASP Mobile Top 10 for app developers - How certain security techniques directly protect against common vulnerabilities - Additional guidelines for mitigating risk and improving resiliency - The importance of a layered approach to mobile app security.
<urn:uuid:fe4146f7-afbc-45f1-8501-0f7fa8d38850>
CC-MAIN-2024-38
https://www.guardsquare.com/blog/owasp-mobile-top-10-security-risks-for-app-developers
2024-09-13T02:53:39Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651506.7/warc/CC-MAIN-20240913002450-20240913032450-00025.warc.gz
en
0.896471
1,005
2.75
3
IT versus OT When looking for an IT asset management solution, it’s important to understand the difference between information technology (IT) and operational technology (OT). In a nutshell, IT is the technology necessary for monitoring, managing and securing information for business applications. It combines technologies for networking, information processing, enterprise data centers, and cloud systems to manage data and applications. Meanwhile, ‘OT’ refers to hardware and software that detects or causes change through the direct monitoring and/or control of industrial equipment, assets, processes and events. The convergence of IT and OT is necessary for successful operations; having a grasp of how these play their part in keeping your company moving will ensure efficient IT governance frameworks. What is information technology (IT)? ‘IT’ generally refers to networks and devices that work together to store, control and deliver information. It’s an umbrella term that covers the creation, processing, exchange, storage and retrieval of data. IT is particularly concerned with electronic data; it’s the data processing systems that serve as the repositories of an organization’s information and is responsible for making data available to relevant users for business purposes. Enterprise resource planning (ERP) systems and cloud-based networks are examples of IT. Their main function is to ensure the proper management and security of data. What is operational technology (OT)? OT is the collection of hardware and software configured to trigger, control, monitor and send data to an IT system in industrial systems. Unlike IT, which is primarily connected to the Internet, OT deals primarily with physical devices. OT is used in the context of an organization’s industrial operations as it’s used to manage industrial equipment and systems, such as highly specialized systems found in energy, manufacturing, robotics, waste control, and oil and gas industries. Examples of OT include robots, industrial control systems (ICS), supervisory control and data acquisition (SCADA) systems, programmable logic controllers (PLCs), and computer numerical control (CNC). ATMs, kiosks, connected transportation, weather stations, and electric vehicle charging systems are all also considered OT. Nowadays, OT has become more advanced, integrating technology such as automation and monitoring for more seamless operations. These have helped OT devices become better linked in complex computer system networks, allowing users to perform at a much faster and more efficient rate. Key differences between IT and OT IT and OT are two important technological components of any organization. While they seem similar, especially in how they benefit operations, they are distinct. Here are some of their key differences: IT deals with monitoring, managing and securing core functions in data centers and the cloud. It focuses on broad business needs, such as communication, transactions and information storage. It’s primarily focused on electronic data and is centered on an organization’s front-end informational activities. By contrast, OT is related to an organization’s industrial operations and is concerned with machine-driven data, physical devices and optimizing processes. OT is related to back-end production, specifically machines. IT is often more cost-effective than OT. That’s not to say that one is a more worthwhile investment than the other, especially because industrial organizations will likely require both. That said, OT is generally more expensive because it deals with machinery and will need to be maintained more often. Fortunately, entry costs of OT devices and platforms have dropped significantly, prompting companies to allocate their budgets better to come up with a productive mix of both IT and OT resources. When IT fails, it affects a lot of the company’s internal operations, particularly concerning data loss. When OT fails, on the other hand, physical assets become wonky, which could lead to machine failure and stalled operations. IT devices are simpler and easier to manufacture, so they tend to last longer. In contrast, OT devices have to constantly keep up with new advances in tech, so they can become outdated (and require replacements or upgrades) much more quickly. The question isn’t whether IT or OT is better–companies nowadays must recognize that both are needed for successful operations, but that’s also only possible when there’s convergence between the two. Tech convergence isn’t a new concept, but companies are currently exploring better ways to seamlessly integrate their IT and OT systems for a more unified and comprehensive suite, whether that’s through application portfolio management tools or asset management solutions like Flexera One. IT/OT convergence is a must in today’s business landscape because it’s key to improving efficiency, reducing errors, cutting costs, enhancing workflows and gaining competitive advantages. It’s imperative for companies who are hoping to scale operations. Converge IT and OT with Flexera One One of the first steps to converging IT and OT systems is to employ solutions that can unify the two components into a single management platform. Flexera One is designed to help companies get a full view of their digital assets for better management. Learn how you can use it to achieve IT/OT convergence today!
<urn:uuid:3bf3bf24-4925-45d0-b8d8-876c816ec365>
CC-MAIN-2024-38
https://www.flexera.com/resources/glossary/it-versus-ot
2024-09-17T21:46:07Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651835.53/warc/CC-MAIN-20240917204739-20240917234739-00525.warc.gz
en
0.9455
1,052
2.859375
3
It used to be the case that only businesses used virtual private networks (VPNs) to connect securely to the internet and keep their private data safe. But these days, with the rapid growth of online threats and privacy concerns, even casual internet users should seriously consider using a VPN. Nearly 30% of people now use VPNs for personal reasons, and that number is only growing as more people learn about how VPNs offer an effective way to safeguard online privacy, enhance security, and protect against various cyber threats. If you are not familiar with this technology, a VPN essentially allows you to send and receive data across a public network as if it were a private network that encrypts, or scrambles, your information so others cannot read it. Let’s take a look at the top 3 reasons why a VPN could come in handy for you. - You work remotely—If you like to take your laptop or mobile device to the library or cafe, you probably connect through public Wi-Fi. The problem is that many free, public Wi-Fi networks are not secured. This means that a hacker could easily intercept the information you send over the public network, including your passwords and banking information. A personal VPN means you can connect securely any time you are away from home. - You want full access to the internet when traveling— When traveling, cybersecurity risks abound, from unsecured public Wi-Fi networks to potential data breaches. Using a VPN while traveling can mitigate these risks by encrypting your internet connection and protecting your sensitive information from hackers and other malicious actors. - You take your privacy seriously— Internet service providers (ISPs) often track and store information about what we do online, as do the websites we visit. This is how they serve up targeted ads and make money when we click on them. (They track us by identifying the unique number assigned to your device, called an Internet Protocol address.) However, when you use a VPN, your traffic is routed through different Internet Protocol addresses, making you anonymous. What to look for in a VPN: Now that you know why having a personal VPN is so useful, here are a few tips to help you choose the right product for you: - Ease of use—You want secure technology, without having to be a tech whiz to use it. That’s why you should look for a product that is easy to implement, like the McAfee Safe Connect VPN app, which allows you to easily and securely connect, ensuring that your passwords and data stay private when using public networks. - Robust security—Look for a VPN with bank-grade encryption. This way no one can read or access the private information you send over the network. - Access to virtual locations—With this feature, you can gain full access to the Internet and browse anonymously.
<urn:uuid:368b11fd-e70f-4560-8c2f-ce5109b7c435>
CC-MAIN-2024-38
https://www.mcafee.com/blogs/tips-tricks/3-reasons-you-need-a-personal-vpn/
2024-09-17T22:47:49Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651835.53/warc/CC-MAIN-20240917204739-20240917234739-00525.warc.gz
en
0.94761
576
2.546875
3
(5 votes, average: 5.00 out of 5) Do you use ChatGPT? Who doesn’t? In fact, it gained 1 million users within just five days of its launch. But with new innovations constantly happening, cyberspace has become much more vulnerable to newer security threats, like “FraudGPT” and “WormGPT.” The worst part is that companies and individuals are not aware of the trouble these both can create. So, in this blog, we are going to explain everything about the new faces of cybercrime. Let’s get started! First, we want to correct a misconception. Many people also get confused about whether WormGPT and FraudGPT are the same as ChatGPT. But that’s not true! ChatGPT is developed by Open AI, a legitimate organization, but WormGPT and FraudGPT are not its creations. Cybercriminals have developed their own advanced AI chatbots for malicious activities. Read on to find out more! On July 13, cybersecurity company SlashNext revealed the discovery of WormGPT, a blackhat “alternative” to ChatGPT,” that lets users do “all sorts of illegal stuff.” WormGPT is a type of malware that utilizes advanced artificial intelligence (AI), particularly natural language processing (NLP) that ChatGPT uses. It is based on a 2021 GPT-J language model that is open-source and capable of generating human-like text. This capability allows WormGPT to write phishing emails, social media DMs, or other forms of communication that trick entities into clicking on malicious links or downloading infected files. It is a product sold on the dark web and Telegram that works the same as ChatGPT but is not used for ethical purposes. FraudGPT creates content to facilitate cyberattacks, and its subscription fee starts at $200 per month and goes up to $1,700 per year. Here’s how it works! Its interface looks exactly like ChatGPT, with the user’s previous requests in the left sidebar and the left portion window taking up most of the screen. All that is required is to write the query and press ENTER. During testing conducted by the Netenrich team, FraudGPT showed its ability to generate deceptive emails resembling those from legitimate banks. (as shown below) Hackers simply provide the bank’s name, and FraudGPT crafts convincing messages and even advises where to add malicious links for maximum impact. Moreover, FraudGPT can create fraudulent web pages, encouraging users to provide their sensitive information. Beyond email scams, FraudGPT provides insights into commonly targeted websites and services. This helps cybercriminals to plan their next moves. Researchers from the SlashNext team said that FraudGPT is trained on extensive datasets of malware, which makes it capable of producing convincing phishing and business email compromise messages. Both WormGPT and FraudGPT have the capability to generate malicious content, including phishing emails, scam landing pages, and malicious code. As these tools automate the creation process, cybercriminals can scale their operations and deploy a wide range of fraudulent activities with minimal effort. FraudGPT specializes in financial fraud and identity theft. Here’s how! It uses advanced AI algorithms to do sophisticated scams, creating realistic fake identities and manipulating financial transactions. Through voice synthesis technology, it can also impersonate individuals, which is not easy to differentiate. WormGPT and FraudGPT can adjust their strategies in real-time, which helps them bypass security measures. They can avoid traditional security protocols and remain undetected for extended periods, which significantly increases the success rate of their attacks. WormGPT can autonomously identify vulnerabilities within networks and exploit them to spread further. Even cyber criminals with limited skills can use these tools, which makes them accessible to a broader spectrum of hackers. WormGPT uses NLP (Natural Language Processing) to mimic human communication. This encourages users to click on malicious links or download infected files. Hackers always find ways to stay one step ahead of security systems. Here’s how they are exploiting the dark side of AI. AI can create convincing deepfake videos and manipulate media to spread misinformation or blackmail individuals. These techniques are so sophisticated that it is not possible for a human to distinguish between genuine and falsified content. Imagine a deep fake video of a company’s CEO requesting personal information from its users. Automated scripts or AI-powered tools are used to test large volumes of stolen usernames and passwords (p/w) across multiple websites or online services. By using AI algorithms, cybercriminals can rapidly cycle through millions of credentials and identify valid combinations that grant access to user accounts. In supply chain attacks, hackers target a company through its network of suppliers or partners. How? They use advanced techniques to find weaknesses in their security defenses. Once done, they sneak malicious software or code into products or updates. This disrupts production or creates counterfeit products that could damage the company’s reputation or operations. For instance, AI-powered chatbots integrated into a compromised website allow hackers to steal user credentials. AI algorithms are capable of evading detection. This is because traditional security measures are not that successful in detecting sophisticated malware. Hackers analyze patterns in malware code and security protocol to create variants that can easily bypass antivirus software and intrusion detection systems. From identifying vulnerabilities to launching attacks, AI algorithms automate various stages of the cyber attack process. This automation helps them scale their operations and target a larger number of systems with minimal manual intervention, such as distributed denial of service (DDoS) attacks. Also Read: Largest DDoS Attacks Reported till Today In DDoS, hackers flood the targeted system or network with an overwhelming amount of traffic, data, or requests. This results in targeted systems or networks becoming unavailable to legitimate users, leading to downtime and loss of productivity. Artificial Intelligence is itself a solution for AI-powered cybercrime. Use the following strategies to safeguard your systems against hackers. Business processes are becoming more complex day by day, so it’s essential to maintain strong cybersecurity defenses. AI-based solutions offer a proactive approach by autonomously monitoring and detecting threats in computer systems. These can analyze: One of the key usefulness of AI-driven cybersecurity is its ability to operate continuously without human intervention 24/7, even during non-business hours. Also, it can automate common routine tasks such as system diagnostics, updates, and fixes, thereby enhancing operational efficiency. Improving sustainability through better vulnerability management used to be a time-consuming task. Cybersecurity experts had to spend a lot of time gathering and analyzing data manually. But now, with the help of Artificial Intelligence and machine learning, things have become much easier. AI-supported vulnerability scanning and management tools can automatically check important data sources like network records, system logs, and reports on threats or bugs. This means weaknesses can be found and fixed in the systems before hackers take advantage. In a nutshell, companies can protect their data better and keep their operations running smoothly by staying ahead of potential threats. Behavioral analysis proves helpful in monitoring suspicious activities online. Using cognitive computing, large volumes of data can be analyzed to detect unusual user behaviors and identify potentially harmful connection patterns. When suspicious activity is detected, AI can either block it outright or alert security teams for further investigation and action. AI’s machine learning capabilities allow it to adapt to different user behavior patterns over time. It can recognize anomalies or potential threats by continuously learning from data and automatically flagging them. When software developers release updates, they often include fixes for known security issues and weaknesses. So, ensure that the hardware, software, and apps are updated properly to keep the devices safe because systems with outdated software are easier for hackers to exploit. Updating software is also very straightforward. Most software applications have built-in mechanisms that enable automatic updates. Alternatively, updates can be checked manually and installed as needed. According to historical data and current cyber event analysis, AI can predict future attack trends to prevent cyber threats before they occur. It can detect suspicious behavior, unusual network activity, or signs of a cyber attack. Then, provide early warnings of potential threats, which allow organizations to take preventive action before serious damage occurs. Pro Tip: Think twice before clicking on a link! If you don’t recognize the sender, avoid clicking; instead, block the number and report it as spam. It is a fact that hackers will never stop hacking and find new tactics as soon as new innovations emerge, as this is the only thing they can do. So, we will also do what we are best at, i.e., protecting the systems. To streamline the process, get Cybersecurity Services from Certera and protect your business from cyber-attacks and malicious activities.
<urn:uuid:921c58ad-f33b-40bd-a842-f140fa753192>
CC-MAIN-2024-38
https://certera.com/blog/wormgpt-fraudgpt-the-dark-side-of-generative-ai/
2024-09-19T04:10:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651981.99/warc/CC-MAIN-20240919025412-20240919055412-00425.warc.gz
en
0.926279
1,859
2.53125
3
As more organizations explore the need for long-term remote work solutions, virtual desktop infrastructure, or VDI, has emerged as a secure and cost-effective strategy. What is VDI technology? How does it work, and what are its benefits over physical IT infrastructure? Today we’ll dive in and learn more about the basics of virtual desktop infrastructure. The shift to remote working environments has necessitated developing safe and efficient remote technology solutions. Virtual desktop infrastructure technology has enabled workers to access computer networks remotely from almost any device, which is more important than ever in the post-COVID-19 era. VDI technology creates virtual desktop environments via remote servers. Those virtual desktops are controlled through software that is hosted on virtual machines, or VMs. Centralized servers then deploy desktop environments to users on request, eliminating the need for individual, physical desktops. Remote access technology is not new and has evolved (did you ever use a dial-up modem?). Client-VPN (virtual private network) applications were once commonplace for additional security. Companies now use virtual desktop infrastructure in place of VPN for convenience, but VPN can still be used for added safety, especially for industries handling sensitive information. As bandwidth has increased and internet connection has become more accessible, so has the ease of working remotely. As the demand for faster, cheaper and safer technology grew, so did the applications to accommodate that demand. The increasing focus on digital transformation and digital workspaces drove innovations that morphed into VDI, making remote access even easier. A hypervisor is a software that creates and runs virtual machines. In virtual desktop infrastructure, the hypervisor’s host machine separates servers it controls into virtual machines, or VMs. These VMs host virtual desktops, which users can access remotely from any device (like desktops, laptops or mobile devices) or location. The host server handles all processing, not the remote device, conserving bandwidth and decreasing lag time. Connection to a virtual desktop is made through a connection broker. This software gateway acts as a middleman between the server and the user. Aren’t VDI and virtual desktops the same thing? No. VDI enables a user to log in to an environment hosted on a central server. Desktop virtualization means that a user accesses a specific computer; usually the one sitting on their desk back in the office. VDI is a type of desktop virtualization. But with VDI, users log into centrally sourced software on a server, enabling new users to access resources without a dedicated desktop machine. This reduces overhead and maintenance, especially for employees or contractors who may never physically work in an office environment. VDIs can be persistent or non-persistent. Here’s the difference: The main feature of virtual desktop infrastructure technology is that it enables workers to access applications and files remotely from any location. Whether an employee or contractor works remotely full time or only periodically, easy remote access expands their ability to perform regardless of where they are. Whether employees work via desktop, laptop, tablets or other mobile devices, VDI offers them the freedom to work where and when they want to. Never worry about data loss again. Virtual desktop infrastructure plays a vital role in security by protecting sensitive data from hardware theft and breaches. Data and applications are stored on the server, safeguarding them even if equipment is lost or stolen. Hardware requirements are much lower with VDI since servers handle processing needs. Older equipment, tablets and thin clients can be used for VDI access, eliminating the need for expensive hardware upgrades. Since the VDI environment is centrally controlled, software updates are easy. Change configurations, apply software patches and configure virtual desktops within VDI’s centralized format. Is your company going through a growth spurt? VDI scales applications to accommodate any size team, quickly and easily. Desktop provisioning is easy with virtual desktop infrastructure. Individual systems don’t need to be configured manually, saving time and effort. Virtual desktops are set up quickly, as settings can be mirrored from a desktop image. VDI is a useful technology for many types of workers and industries, including: There are a few things to consider when setting up a digital workplace: Know what systems you need: Understand your requirements to avoid under- or over-provisioning your network capacity. Know what your users need: Would it be better to set up a persistent or non-persistent VDI? Provisioning can vary depending on the level of desktop customizations required. Maintain security: Provisioning and onboarding new users is usually done routinely. But what happens when employees leave or are terminated? Make sure you have steps in place to suspend access and credentials when necessary to prevent cyberattacks and data theft. Test, test, test: Before engaging your VDI, be sure to run a pilot test deployment to ensure resources are sufficient. Is your helpdesk support sufficient? Do you have systems in place to manage your virtual desktop infrastructure implementation and long-term support your users need? If not, consider partnering with a managed services provider like Helixstorm. Helixstorm can help your business manage VDI and digital workspaces, so you don’t have to. Contact us today to learn how our managed IT support services can help support your remote work strategy.
<urn:uuid:1ef8b532-75fd-41f3-aa41-d2113d0a658d>
CC-MAIN-2024-38
https://www.helixstorm.com/blog/what-is-vdi-virtual-desktop-infrastructure-explained/
2024-09-08T05:25:27Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650960.90/warc/CC-MAIN-20240908052321-20240908082321-00525.warc.gz
en
0.912279
1,097
2.609375
3
As a recap, let’s quickly define what threat intelligence is in the context of cyber-security. Threat intelligence is created by a process which takes raw data and information from a variety of sources and turns it in to strategically, tactically, or operationally valuable information. The collected raw data and information is then categorised, analysed further, and given context and meaning, producing threat intelligence. This intelligence is then used by an organisation to ensure its governance, risk and operational functions understand what threats they face and can make informed decisions in response. Automated processing: how we do it, its value and limitations As an industry we have at our disposal a plethora of cyber-domain-specific and generic computer science approaches to automate or semi-automate data and information processing. The purpose of this automation is to aid in the discovery, analysis and ultimate production of threat intelligence. At a high level the most common approaches today include: Systems for the automated collection, processing, and storage of data that is big or small and may be either generic (network traffic collection) or problem-specific (e.g. malware analysis) in nature; basic patterns, such as regular expressions to identify data that is or is not of interest; statistical or probability algorithms, to identify things which are or are not similar; machine learning algorithms, to provide statistical classification around what is or is not normal or expected; and natural language processing of human-produced text, to extract sentiment, intent, purpose, target, or topic. However, while these solutions are tremendous work reduction aids when processing huge amounts of data or looking for things that might be interesting, they aren’t a panacea. Even with machine learning and expert systems (an implementation of artificial intelligence) there is still today no replacement for the human analyst, and thus there is no fully automated way to produce high quality tailored threat intelligence. The limitations of automation often start to become evident where we have new types of events, techniques or other things we previously didn’t understand. In these situations gaps can appear in the automated system’s ability to collect, consume, process or otherwise classify its inputs or outputs. When these systems start to stumble in this way we often see them come to incorrect conclusions and thus devalue or derail the resultant information. Human processing: the power of adaptive reasoning In threat intelligence there is a confidence scale that goes from the fully qualified ('this is the threat to your organization because of xyz facts which are supported by def with this level of certainty') through to the unqualified ('we cannot say this is a threat or not at this time because of abc unknowns or contradictions'), with varying degrees of confidence, caveats and assertions in between. Threat intelligence may also need additional context relating to the organisation consuming it in order to be of value, or there may be small print that is important when making decisions based on it. It is for these reasons, among others, that human threat intelligence analysts are critical in providing the required narrative, context, and meaning to ensure the quality one would expect from such intelligence if it is expected to be actioned. This reliance on humans as part of the process arises from a unique trait that we have over computers – our ability for adaptive reasoning, or in other words our ability for problem solving and our ability to think laterally. Unlike today’s algorithms or childlike artificial intelligence, an adult human has an amazing ability for logical thought, challenging assumptions, explanation, and justification. These abilities, combined with being able to think at a conceptual level and inferring approximate meanings and likely relationships very quickly, mean we are able to make huge leaps where a computer may not. These advantages mean a human analyst can quickly draw a likely correlation between, for example, a piece of malicious code, a set of events, and a likely set of threat actors, and can come up with an interpretation of the threat actors’ likely motivations. This likely correlation can then be either supported, questioned or dismissed based on the available data and information. Even with the risk of confirmation bias, a good threat intelligence analysis function with appropriate quality and oversight controls will outperform its competitors. As a result, threat intelligence analysts are able to go beyond what any fully-automated system can do today in terms of finding related events, observables, tactics, techniques, procedures, and actors, while also providing valuable context and meaning to the business. The business value of human analysis and interpretation of intelligence While humans are impressive, we are not infallible when identifying casual relationships due to, as already mentioned, tendancies for such confirmation bias or causation. We need no clearer example of this frailty than the attribution of Internet-based threat actors. Today there is a tendency to point to certain countries as the originators. However, as their tactics, techniques, and procedures become increasingly scrutinised, well understood and publicised, it is also arguably getting easier to replicate their techniques and thus appear to be them. This is a good example of an assumption that should be challenged, and a good analyst will provide further evidence or appropriate caveats when making such an assumption in their analysis. This analysis and interpretation of the data and information to form the threat intelligence and associated nuance is a good example of why humans are critical in providing the context and meaning before important strategic, tactical, or operational decisions are made based on it. History and fiction is littered with examples of unintended consequences when computers are left to make their own decisions. So while cyber threat intelligence providers invest heavily in technology and intellectual property development, we also place a strong emphasis on having a suitable-sized team comprised of diverse and experienced threat intelligence analysts. These analysts, if you hadn’t already guessed, are the ones who are able to interpret the data and validate whilst providing the required context and meaning. In short, humans are still unique in their efficiency and abilities and we won’t be replacing them yet in our provision of high quality cyber threat intelligence to clients. Sourced from Ollie Whitehouse, technical director, NCC Group
<urn:uuid:7197a4c9-2219-40cd-a890-0fe2a0a6938a>
CC-MAIN-2024-38
https://www.information-age.com/why-software-can-never-replace-human-security-analyst-33245/
2024-09-09T12:56:55Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651098.19/warc/CC-MAIN-20240909103148-20240909133148-00425.warc.gz
en
0.940775
1,253
2.671875
3
How long will it be before Web-via-mobile phone is commonplace? As yet, little support. A lot of companies have bet a lot of money on the eventual success of a new category of wireless devices, which have come to be known as smart phones-or, more recently, as net phones. Because features and functionality tend to differ considerably from one device to another, the category eludes precise definition. One can only say that a smart phone is a device that combines the functions of a mobile or cordless phone with some measure of computing capabilities. Initially, the term applied to a mobile phone that could receive and store text messages. But because that capability is now ubiquitous among digital mobile phones, the smart-phone nomenclature is more apt to be applied to mobile telephones with built-in Web browsers, and, in an expanded definition, to those relatively few devices that incorporate the more extended functionality of a Palm computer. Multiple operating systems Interestingly, the combination mobile phone/personal digital assistant (PDA) represents the older approach and has been embodied for the past three years in the Nokia 9000 Communicator series, and more recently in the Qualcomm PDQ phones, which include a fully functional PalmPilot. The Nokia phones are based on the little-used Psion operating system, while the PalmPilot employs an operating system that is entirely proprietary to 3Com, the company that currently owns the PalmPilot design. Thus far, no one manufactures a smart phone with PDA capabilities based upon the popular Windows CE platform. Despite the power and flexibility of the Nokia and Qualcomm units, neither has found much acceptance at the enterprise level. Rather, consumers and individual business users form the principal market for the phones. Whether the relatively high cost of such phones is the chief obstacle to their acceptance by information-technology (IT) departments or PDAs in general are achieving relatively slow penetration among business organizations, the true computing phone has few advocates as yet in the business world. The browser phone is a more rudimentary device based on a thin client architecture and is intended to access the World Wide Web through a specialized wireless Internet portal maintained either by the carrier or the enterprise. The limited display resolution and throughput rates of mobile phones make a normal connection to the Internet impractical. What is needed instead is a specially configured server that can compress and cache content and eliminate graphics and streaming media. Such devices are available, and virtually all on the market today conform to the Wireless Applications Protocol (WAP). Browser phones lack the large pen-based displays of the PDQ phones and the full QWERTY keyboards of the Nokia Communicators, nor do they have the memory or processing power to store programs and files as do the PDA hybrids. But they can execute Web-based applications, functioning almost as dumb terminals. The fact that the WAP thin client does not require a powerful and costly CPU and can instead use the server to run applications has made the WAP phones more attractive to cost-conscious IT departments than the older PDA phones. Indeed, the browser phones are beginning to appear in such field applications as route delivery and fleet sales and maintenance where they are starting to replace the older, bulkier, ruggedized terminals in instances where extensive report generation and transaction processing are not required. WAP phones typically cost a fraction of the price of mobile data terminals, and they allow low-cost voice communications, as well. Where short messaging is to be combined with dispatch and mobile phone service, they provide an excellent solution. Nevertheless, Web-based business applications designed around the mobile connection are in short supply, and the display limitations and lack of effective two-way messaging over the phones make them ill-suited in applications where lengthy upstream data communications are required. In certain vertical markets, powerful two-way pager/communicators such as the RIM device continue to enjoy an advantage. The coming of high-speed third-generation (3G) networks will surely lead to the development of a new breed of smart phones that can take advantage of the ability of networks to deliver large quantities of data quickly. One can assume voice activation and voice recognition will begin to come to the fore to provide a more responsive interface. Displays are apt to remain a problem, however. There's simply no way to put more than a few square inches on a pocket phone. It is also possible that the traditional mobile phone could give way to a pure PDA form factor where a speaker phone is substituted for a handset. Manufacturers are already examining such concepts. Mention should also be made of smart phones for in-building use. Spectralink makes phones with messaging capabilities, and Symbol has a product called the DataPhone, which combines a wireless phone with a ruggedized data terminal complete with bar-code scanner. Whether this is the beginning of a trend remains to be seen. Dan Sweeney is technical columnist for Wireless Integration, another PennWell publication.
<urn:uuid:ab9cedff-d110-4e71-9c2e-6efc3957c8b9>
CC-MAIN-2024-38
https://www.cablinginstall.com/home/article/16467331/smart-phone-still-waiting-in-the-wings
2024-09-12T00:27:11Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651405.61/warc/CC-MAIN-20240911215612-20240912005612-00225.warc.gz
en
0.944953
1,002
2.75
3
Bring up the topic of artificial intelligence with friends and colleagues, and you’ll probably hear three different ideas expressed. The visionary entrepreneurial types will talk about the possibilities that exist for data analysis and efficiency. The pessimists will worry about the possibility that human workers will be replaced by robots. And the science-fiction fans will wonder when our phones are going to become self-aware and exterminate all biological life forms. From the visionary perspective, we can only say the potential of machine learning in the world of business – not to mention science and technology – is truly limitless. And unfortunately, we don’t have any opinions on the probability of Earth being overrun by cyborgs. That just leaves us with the not unreasonable issue about AI displacing people who are employed. So, are robots coming for all the jobs? We can see why there is a lot of concern in this area, but we don’t think it’s something to spend too much time worrying about for now. Let’s look at a few of the reasons why… Current AI Has Its Limitations There are a lot of things that computers do very well. However, microchips and robotics aren’t necessarily at a tipping point of functionality yet. As an example, researchers have tried for years to create a machine that can fold laundry and they have yet to be successful. Only time will tell how much better machine learning can get in the next couple of decades, but it may be that many careers and industries aren’t under any immediate threat. Technology Doesn’t Usually Replace Jobs in the Ways People Think Even in situations where technology can make a human employee replaceable, it isn’t necessarily the case that a job will go away. It may just change. For instance, ATMs haven’t reduced staffing in banks. It’s simply the case that financial employees have better (and less repetitive) jobs than they did before. Similarly, there are some things technology can do, but probably won’t. How many of us want to see a robotic therapist, or get a massage from an appliance? AI Could Bring Huge Benefits for Employers and Employees Progress of any kind tends to be messy and disruptive, but when it comes to potential artificial intelligence, the payoffs could be enormous. Imagine a world where instantaneous data streams could help us to manage food, water, and healthcare more effectively. While we can’t promise that no one will lose their job to robotics or AI in the next decade, the risks are probably being overblown. And the advantages could be enormous and life-changing. Want to Know How the Biggest Tech Trends Will Affect Your Business? At Fantastic IT, we excel at providing fast and friendly tech support for our clients. There are very few issues we can’t resolve quickly with our team of certified technicians and help desk professionals. What many of our clients don’t know is that we also provide consultations and virtual CIO services to help them plan for the future. So, if you’re looking for help with tech-related strategy, budgeting, and business planning, we are ready to step in. To learn more, contact us today to schedule a free consultation.
<urn:uuid:5532ec73-8f1b-4d0c-8140-60f202cf9ace>
CC-MAIN-2024-38
https://fantasticit.com/is-ai-coming-for-all-the-jobs/
2024-09-15T13:08:48Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651630.14/warc/CC-MAIN-20240915120545-20240915150545-00825.warc.gz
en
0.950371
674
2.9375
3
The growth of mobile devices has been huge in the past few years. Mobile phones are now being used for everything, from making calls and chatting with your friends on WhatsApp to accessing important organizational documents. Almost everyone who uses mobiles uses at least one application a day. It’s no exaggeration to say that using smartphones without apps is practically impossible. App wrapping means adding an extra layer of security to mobile apps without altering any other feature. This means that the functionality and the appearance of the app remain the same as its normal counterpart but is more secure. Since all the main aspects of the app remain the same, the people using the wrapped app will feel normal, a factor that distinguishes app wrapping from other mobile app management strategies. Once the technology to wrap an app is implemented, the wrapped app can be hosted on an internal store or sent to devices. As mentioned earlier, more offices are now allowing the use of mobiles for office use, so more corporate data gets accessed through apps. So, app wrapping is most commonly used by organizations that want to secure their data which is getting accessed through mobile apps. How is app wrapping done? App wrapping is most commonly done in 2 ways: using SDK from an app or using an EMM or UEM vendor that allows IT admins to distribute APIs through which management policies can be applied. Through the use of app SDKs, apps can be wrapped during or after the development of the app, all that has to be done is to add executable codes to the app via the SDK. As for EMMs, most of them provide an additional tool called an App Wrapper tool to wrap the applications. Using this tool, IT admins can specify which all apps are needed to be wrapped and then specify management policies to the device through the tool itself. What can app wrapping offer? Through app wrapping, IT admins can get control of all apps that are used by the employees in that company. You can send app management policies to wrapped apps (keep in mind that these policies are applied only to the app and not the device). Mainly 3 types of policies are sent to the wrapped app: - Security policies: – These policies are applied with one goal in mind and that is the security of the data accessed through apps. Security policies allow IT admins to specify app settings to define whether the apps are allowed to access corporate data or not and if users can move files from their default location and so on. Security policies also enable app data wiping, jailbreak detection, app-level integrity checks, etc. These features help organizations make self-defending apps that are capable of securing data access through them. - Management policies: – Management policies are mainly used by IT admins to help the users use apps more easily and efficiently. Functions like specifying whom all can access the app, specifying roles to each user of the app, specifying app visibility to different roles and so on come under the management policies. Using these policies, IT admins can not only help the users but also make sure that unwanted personnel does not gain access to the app. - Analytical policies: – These policies give an insight into how the apps are being used, why the app is being used and the details of the users of the app. These policies also include features like setting time limits as well as network limits for apps, and so on. As the name suggests these policies, in short, help IT admins analyze the usage of the app. Hexnode App Management Solution Learn how Hexnode’s App Management solution helps businesses ensure seamless deployment, monitoring, and supervision of enterprise apps on end-user devices. Download datasheetThe cons of app wrapping App wrapping can help IT admins configure the A-Z of apps. Wrapped apps can be constantly monitored and kept in check. In the event of data breaches via wrapped apps, the app data can be wiped by the IT admins. Using app wrapping, access to apps can be specified by IT admins. Even though app wrapping gives this many management capabilities to IT admins, it has its downsides. - One of the most significant disadvantages of using wrapped apps is that it takes up a lot of resources. Since these apps have to be constantly monitored the apps have to be running in the background for smooth functioning. This takes up a lot of battery and may even cause the device to run slower. - Another major problem is that the app wrapping tool itself has to be protected. Even though the apps that are wrapped using these tools are safe, the tool itself might be vulnerable and organizations have to go an extra mile to ensure the safety of the tool itself. - When app wrapping is completed and the wrapped app is uploaded to an organization’s internal app store, confusion may arise if the normal app is also in the store. When there are two versions of an app in the store, users will get confused about which one to install. - App wrapping can take care of data that is accessed through apps, but once data is copied to another location, outside of the app’s control, it is vulnerable and cannot be managed through app wrapping. This is probably one of the biggest disadvantages of app wrapping. Is app wrapping enough? As advancements in mobile technology are happening at a very fast pace, app-level security is the bare minimum that can be done to counter cyber-attacks and app wrapping does just that. As more and more devices are being targeted by cyber-attackers app wrapping alone can’t ensure the security of corporate data that is being accessed through mobile devices. App wrapping with a device-level management solution like an MDM or a UEM can help organizations keep their devices and data safe and secure. As for management uses, app wrapping can be a good starting point. With the features like app config and custom app configurations, MDMs and UEMs can configure apps for both iOS and Android devices easily.
<urn:uuid:3196a59e-0e1c-4ea1-8f87-371ad5c6c8da>
CC-MAIN-2024-38
https://www.hexnode.com/blogs/what-is-app-wrapping/
2024-09-08T09:15:29Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650976.41/warc/CC-MAIN-20240908083737-20240908113737-00625.warc.gz
en
0.935728
1,203
2.578125
3
Manufacturers claim their optical connectors meet industry standards, yet some network failures are still attributed to connector quality issues. Arespected telecommunications service provider experienced random failures within its optical network after upgrading to OC-192 (10 Gbits/sec). The service provider placed the highest priority on determining the source and eventually concluded that poor-quality optical connectors caused the failures. Technicians discovered that each of the failed connectors had some form of surface defect. Determined to eliminate optical connectors as a cause of network failure, the service provider changed the quality requirements for all optical connectors used in its network. The most significant change in the specification was regarding visible defects on the endface, including the outer regions of the connector away from the core (light transmission area). The new specification rejected all optical connectors with any visible defects (viewed at 400x) on the endface. To guarantee that all suppliers conformed to the new specification, the service provider began inspecting every connector purchased. Only connectors that met the current industry standards for insertion loss, reflectance, and endface geometry and had no visible surface defects were placed into service. This approach worked. There have been no failures attributed to optical connectors since the new requirements were implemented well over a year ago. Some service providers view optical connectors as a commodity, purchasing cable assemblies from the lowest-priced bidder then failing to ensure compliance to quality standards. This practice causes at least two problems: First, network managers do not know if connectors in their system meet commonly accepted industry quality standards, and second, manufacturers and suppliers are not held accountable to the standards they profess to meet. Failing to ensure optical connectors meet quality standards is risky, considering the connector is the only place in the optical network where bare fiber can be exposed to the environment. Great care is exercised throughout the remainder of the network to enclose and protect every location where the optical cable is disturbed or opened (splices, transition points, etc.), yet the single place where bare glass can be exposed is often overlooked. With today's networks carrying higher data rates and more optical power than ever before, connector quality is even more critical to network performance. Network failures due to optical connectors can be virtually eliminated by understanding a few key issues surrounding connector quality and implementing some basic practices. Telcordia Technologies (formerly Bellcore) has developed specifications for network components with the goal of ensuring network reliability. The specifications developed for fiber-optic terminations are outlined in GR-326 and specify, among other things, optical performance and endface geometry. GR-326 defines acceptable ranges of insertion loss, reflectance, apex offset, fiber height, and radius of curvature for radius polished connectors.Telecommunications providers generally assume fiber-optic terminations that meet the GR-326 specifications perform reliably in high-speed networks. Therefore, carriers that purchase connectors from a supplier advertising GR-326 conformance expect to avoid any problems related to connector quality. Recent evidence based on field failures and laboratory testing, however, indicates that this is not always the case. To determine the level of industry compliance to GR-326, Telect purchased cable assemblies from five manufacturers that specified GR-326-compliant product. Fifteen singlemode jumpers (30 terminations) were purchased from each vendor and an independent testing laboratory tested all of the connectors. The connectors were tested for insertion loss, reflectance, apex offset, fiber height, and radius of curvature. The results were then tabulated and analyzed.Figures 1 and 2 display the test results for fiber height and reflectance conformance. Based on these test results, there is a reasonable possibility that some of these cable assemblies do not meet the Telcordia GR-326 specification for fiber height and reflectance. Using connectors that do not meet fiber height and reflectance specifications can negatively affect network performance. For example, a connector with a fiber height that is too high can damage the optical endface of the mated connector. A connector with the fiber height too low can fail to make physical contact with the mated connector, increasing reflectance and insertion loss. High reflectance interferes with optical transmission and can cause laser de-stabilization or shutdown. Test results of the other performance and endface geometry specifications demonstrated similar examples of nonconforming connectors. If a connector fails one requirement, such as fiber height, the entire cable assembly is classified as not meeting the GR-326 specification, even if it passes all other parameters. Connectors must meet all performance and endface geometry specifications for the cable assembly to be deemed GR-326-compliant. The test results indicate that most manufacturers are compliant for some specifications but poor at meeting others. The percentage of cable assemblies that passed all GR-326 compliance requirements ranged from 100% for one manufacturer to as low as 40% for another; the rest fell somewhere in between. What happens when GR-326 specifications are not met? Network performance can be adversely affected in multiple ways. A quick look at a few GR-326 specifications amply demonstrates this point. Table 1 provides an overview of the standard's specifications and the possible network effect when a connector does not comply. The telecommunications service provider that upgraded to OC-192 purchased cable assemblies that, according to the manufacturer, met the Telcordia GR-326 specification. Yet, the carrier still had network failures due to poor connector quality. How is this possible? One possibility is that perhaps not all connectors met the specification as advertised. Most manufacturers today use some form of statistical process control (SPC) to monitor manufacturing processes and determine if quality requirements are met. SPC analyzes data taken from random samples to determine the level of compliance. It provides reassurance that the manufacturing process is in control and that a high percentage of the product meets the specifications. SPC is particularly useful in lowering production costs, because it reduces the amount of testing or inspection required. However, it is only an indication, and cannot-nor does it claim to-assure that 100% of the product meets the specifications. Generally, several test samples in a row have to either fail or be outside of the normal "trend" for a manufacturing process to be considered out of control. In other words, today's most commonly used method for assuring quality does not guarantee that 100% of the product meets the relevant specification. Another reason for network failures due to connectors is that GR-326 may not be comprehensive enough to ensure reliability. There are other characteristics under "specified" or "unspecified" in GR-326 that may reduce performance or cause failure. For example, certain types of defects on the optical surface of the connector may not hamper initial performance, particularly if the defects are not in the light-carrying area, but over a period of time and under certain environmental conditions, the defect may change, eventually resulting in a network failure. Defects on connector endfaces may simply be failures waiting to happen.Cabling manufacturers have spent millions of dollars in research and testing to ensure cable designs fully protect the fiber, even under the most extreme conditions. Hardware manufacturers have spent millions to develop cable support and closure hardware that provides a protected and relaxed environment for optical fiber during manufacture, installation, and usage. Optical-fiber companies take great care to ensure fiber is protected as soon as possible after manufacture. During the manufacturing process the fiber is coated with a buffer, a protective coating that is extruded over the glass as it flows from the drawing tower onto the winding spool. The purpose is to protect the fiber from any damage, including moisture in the air. During installation of optical cable, any place that the optical fiber is exposed to air, such as a splice, is quickly coated to protect it from damage. The only possibility of exposed glass in an optical-communications system is at the connectors. Connectors use pressure to make a good connection, increasing the possibility that damage from any surface irregularity will occur. Therefore, it makes sense to pay particularly close attention to the quality of connector faces. Surface defects on connectors can consist of pits, chips, scratches, surface irregularities, and roughness. The defects can occur in the core area, the contact area, or the epoxy ring around the outside of the fiber.Figures 3 and 4 show two different connector endfaces after careful cleaning to remove dust and dirt with typical surface defects. Since the connector in Figure 3 has surface defects in the core area, there is a very real possibility that this connector will cause network problems. A connector with this type of surface imperfection may cause an increase in reflectance, damage to the adjoining connector face, or high insertion loss. Intuitive reasoning suggests that defects in the light-carrying area of a connector endface can cause transmission problems. If the defect is not obstructing the core area and does not impede alignment, it may have no initial adverse affect. Figure 4 shows a connector endface with no visible defects in the core area, but with defects on other areas of the connector surface. Under pressure from the connector springs and over a period of time of varying temperatures, humidity levels, movement, and vibration, defects may grow to a magnitude that interferes with optical performance or at least become reservoirs for contaminants. As time passes, changes in the connector endface are likely to occur because of varying environmental conditions. Surface defects by definition are interruptions in the smooth surface of the glass, and these interruptions can migrate under pressure. Even the movement of air on the bare surface of glass can cause surface imperfections to travel. Studies indicate that the tensile strength of fiber decreases when surface imperfections are exposed to the environment. Moreover, fiber defects can propagate when exposed to humidity, vibration, and temperature swings. The severity, nature, and time of failure due to surface defects are largely unpredictable, although several models have been developed that attempt to predict reliability. Because surface defects demonstrate the likelihood of changing, network failure is possible. Empirical data and scientific research imply that the elimination of surface defects minimizes network failures due to poor connector quality. Surface defects can trap contaminants and become reservoirs of moisture or other types of contamination. Moisture is particularly harmful because of absorption, which reduces mechanical strength. Other types of contamination may be abrasive to the surface, causing scratches, pits, and chips in the endface material. One manufacturer has even suggested that certain contaminants have an affinity for the light-carrying area of the fiber. Any contamination in the optical path can injure the reliability of the transmission. Evidence discovered through research and field experience suggests that communications providers should take steps to ensure optical connectors meet all critical parameters of Telcordia GR-326 and have no surface defects in the core or contact areas. There are several ways to achieve this level of quality. Service providers can inspect and test every connector end for all relevant specifications. This testing requires inspection equipment capable of accurately quantifying each measurement. Equipment that is commonly used to verify these parameters includes interferometers for apex offset, fiber height, and radius of curvature; high magnification measuring microscopes or imaging systems for surface defects; calibrated light sources and power meters for insertion loss; and reflectance meters. In addition, technicians skilled in the handling and cleaning of fiber and the use of the test equipment are required. Thorough inspection will ensure each connector is of sufficient quality to perform reliably in the optical network. Of course, there are drawbacks to this inspection process, including a large investment in test equipment, increased labor costs, and a delay in placing optical cables into service while the inspection takes place. However, consider what the costs are when an optical cable assembly fails. The average time to repair a network failure is estimated at about two hours. An OC-192 network can generate $7,300 an hour in revenue, so just one failed connector can cost upwards of $14,000 in lost revenue, not to mention the intangibles such as customer confidence and productivity. But if inspecting each and every connector is not feasible, there is a simpler way to ensure connector quality. Purchase cable assemblies only from a supplier that guarantees compliance to all relevant specifications through quality practices and 100% inspection. Check to make sure the supplier purchases or manufactures precision parts, has excellent control of the manufacturing process, and performs staged inspection and measurements on 100% of the cable assemblies. Reputable suppliers will be more than happy to discuss these concerns with potential customers. Following this practice has several advantages for the communications provider. Cable assemblies are ready to use immediately upon receipt since no onsite inspection is necessary. The over-all cost of installing high-quality cable assemblies is actually reduced due to quantity efficiencies at the manufacturing facility. The logistics of returns are avoided and the provider is more confident that any network problems are not connector-quality-related. Communications providers that implement and maintain an on-going comprehensive program to ensure connector quality will be ahead in the long run. Either through their own inspection process or by working with a supplier that guarantees compliance to quality standards, service providers can make certain only high-quality optical connectors are used in their networks. While it may require some time, effort, and expense, installing only high-quality connectors is an investment in the long-term performance of the network that will pay for itself many times over. Ken Ditto is product manager, fiber optics, for Telect Inc. (Liberty Lake, WA). His e-mail address is [email protected].
<urn:uuid:51504764-43de-4d32-81cd-8c0ac13abb55>
CC-MAIN-2024-38
https://www.lightwaveonline.com/test/network-test/article/16648728/connector-quality-are-standards-enough-to-eliminate-the-weak-link
2024-09-08T09:00:20Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650976.41/warc/CC-MAIN-20240908083737-20240908113737-00625.warc.gz
en
0.938117
2,736
2.578125
3