text stringlengths 1 1.04M | language stringclasses 1
value | warc_record_id stringlengths 36 36 | url stringlengths 13 4.48k |
|---|---|---|---|
Abstract
abstract = ".",
N2AB | eng | f4c36cfa-fc8b-4cb8-8307-e398e7f054a8 | https://research-repository.uwa.edu.au/en/publications/g-protein-coupled-receptor-kinase-5-regulates-airway-responses-in |
How do you release the Kraken in Virtual Villagers Origins 2?
Once the crafting hut is built, drag a villager to the three empty buckets next to the bag of seeds. The villager will gather water. Once the water is collected, drag a villager to the crafting hut and give them the bucket of water. (If you give them the oil they'll take it to the Kraken statue to be lit.)
How do you get free lava stones in Virtual Villagers Origins 2?
Lavastones are a currency found in Virtual Villagers Origins 2. They can be obtained through free, promotional offers found in-game, or purchased using real money. Lavastones can also be rewarded for completing puzzles.
What is the third puzzle in Virtual Villagers 2?
Puzzle Chronology Crafting Hut is the third puzzle in Virtual Villagers Origins 2.
Where are the Kraken Eyes Virtual Villagers?
The first eye can be found amongst the reeds to the left of the tree at the bottom right of the island. The second eye is found beneath the lab and to the right where the lava connects with the lab. Before you can collect the eye you will have to go to the crafting hut to get water to cool it down.
How do you clear the lagoon on Virtual Villagers?
Your villagers will need to have researched Level 2 of Construction to work on this puzzle. Drag a villager to the pile of rocks blocking the flow of water from the creek. Once the blockage is cleared, the water will flow into a sparkling lagoon that your villagers will use for various work and leisure activities.
What does Giant Head mean in virtual villagers?
Giant Head. Increases the size of the villager's head. Master Builder. The villager becomes a master at building.
Where is the Kraken eyes in Virtual Villagers Origins 2?
How do I get the ruby vv2?
Bring braids to the Clothing Hut and fix them. Gather flowers from the pollinated bush on the right of Kraken statue….
Bring Water from the Crafting Hut to the fire to reveal a diamond and bring it to the statue.
Use Jump Potion to bring ruby to the statue.
Use chief to draw some plans and build the blue statue.
What can the Golden Child do in virtual villagers?
The Golden Child has some significant powers that are unobtainable by any other villager. He is able to magically create food, regrow the berries on the berry bush, and restore the crops in the agricultural area.
How do you push the boulder in virtual villagers
How do you destroy the Rainbow Totem in Virtual Villagers 5?
The Rainbow Totem You'll need to drop 2 builders on both the red and yellow dye almost simultaneously. The villagers will then carry the dye to the hot springs and pour them in. This will drive away the heathens and you can start destroying the Rainbow Totem.
How do you move the boulder in Virtual Villagers 2 in fact, What does the Kraken do in VV Origins 2?
How do you pollinate flowers in Virtual Villagers 2?
Instructions. Wait for the butterflies to follow the Golden Child, then drag him to the strange plant north of the berry bush where the butterflies will pollinate the plant.
What is the latest version of Scene Builder?
Scene Builder is free and open source, but is backed by Gluon. Commercial support offerings are available, including training and custom consultancy services. Scene Builder 17.0.0 was released on Sep 29, 2021. You can use this Scene Builder version together with Java 11 and higher. License: Scene Builder 17 is licensed under the BSD license.
What are the components of the Scene Builder Kit?
The Scene Builder Kit includes the following components: The SceneBuilderKit.jar file included in the JavaFX Scene Builder 2.0 application bundle. When Scene Builder is installed, the JAR file is installed in the following locations: Mac OS X: /Contents/Java/SceneBuilderKit.jar
How do you get the Kraken in Virtual Villagers Origins 2?
Release The Kraken is the fourth puzzle in Virtual Villagers Origins 2 . Two villagers who are both master/adept builders and master/adept gatherers Once the crafting hut is built, drag a villager to the three empty buckets next to the bag of seeds. The villager will gather water.
Where can I find the source code for the Scene Builder application?
The source code is provided and developed through the OpenJFX Project in the OpenJDK Community. Please consult the OpenJFX Wiki for instructions describing how to download and build its source code. Please note that the source code for this application can be found in the "apps/scenebuilder" directory. | eng | 0290fe15-da64-449e-9eba-e45f83523c60 | https://vidque.com/how-do-you-release-the-kraken-in-virtual-villagers-origins-2/ |
Our Content Syndication Course is an immersive and multifaceted educational program that has been meticulously designed to equip learners with a comprehensive suite of techniques and strategies for the successful distribution... | eng | 6941e29c-1fa6-435f-8832-cd47bcf07f1c | https://oyoads.in/category/udemy-coupons/ |
What Is Real Estate Managing?
Real estate management is the professional operation, maintenance, and oversight of real premises assets including residential or commercial properties, land, or perhaps other establishments to generate remarkable returns to get investors. This includes advertising leasing homes to renters, collecting lease, resolving renter issues, developing and employing policies and procedures, along with reporting to owners in the performance of their investment.
In addition to being an expert in the physical facets of property, a good real estate supervisor is usually a excel at at organization and people. Most suitable option evaluate and respond quickly to changing market circumstances, economic trends, regulatory requirements, insurance worries, and other exterior factors that impact the operations with their properties. They are also adept at acquiring additional approaches to increase payment income and keep personal client romantic relationships.
Depending on the sort of property and size, real estate managing fees may range from five per cent to 10% of total rent accumulated. As a result, it is vital to choose an associate that can offer value-added providers that contribute to the success of your investment.
Several landlords take advantage of the interaction with their like it tenants, so they choose to deal with their homes themselves rather than hire a property manager. In some instances, this is an effective and cost-effective solution, but it surely can be a time-consuming endeavor meant for an investor that may have many assets or even just one particular rental property. Others prefer to offload the responsibility and find a reliable asset manager to handle that for them. 37th Parallel Houses has been filling up this part for many passive smaller property investors since 2008. | eng | ab28cf8c-70b4-4565-8c2b-7e24886e4abf | https://bunsbaker.fr/2023/09/05/what-is-real-estate-managing/ |
Social Media Literacy
Our mission is to connect people, brands and communities to create a better future. As part of that commitment, we are working to promote social media literacy and working to get it part of the Irish School Curriculum.
We are dedicated to promoting social media literacy in schools across Ireland. We believe that social media has become an integral part of our daily lives, and it's crucial for young people to develop the necessary skills to navigate it safely and responsibly. That's why we've begun working hard to make social media literacy a part of the Irish school curriculum.
In today's digital age, children and teenagers are exposed to a vast amount of information and content on social media platforms. While social media can provide many benefits, such as connecting with others and sharing ideas, it also presents many risks, such as cyberbullying, online harassment, and exposure to harmful content. Unfortunately, many young people lack the skills to identify and manage these risks, which can lead to negative outcomes for themselves and others.
By incorporating social media literacy into the school curriculum, we can equip young people with the skills they need to navigate social media safely and responsibly. This includes developing critical thinking skills to evaluate online content, understanding privacy and security settings, and learning how to communicate effectively and respectfully online.
We believe that social media literacy should be a fundamental part of the education system, and we need your help to make this a reality. We are calling for support from individuals and organisations who share our vision to join our steering group. Together, we can make a difference and ensure that young people are prepared to thrive in today's digital world. If you would like to get involved, please get in touch with us today.
Our Expertise
We've supported countless SME's and worked with some of the country's biggest companies to grow their business online and develop their brand offline with traditional marketing. We've worked with companies that have both limited marketing budgets and resources and also some that have massive budgets and with both, we can achieve results that exceed your expectations by prioritising your brand requirements. We can help you enhance your digital brand and online presence across all your social media and drive traffic to your website. This will provide the data to help you understand what is working for you, who your customers are and what they want . Then we can tailor your communications to enable you to sell, sell, sell! | eng | 3320e8f2-206c-4b8e-bc76-479f681a8c13 | https://www.pennypop.ie/social-media-literacy/ |
The Different Types of Solar Panels: Manufacturing & The Pros and Cons
Solar energy is revolutionizing the way we generate electricity, and solar panels play a crucial role in harnessing this renewable energy source. As the demand for solar power grows, various types of solar panels have emerged, each with its own unique characteristics, manufacturing processes, and advantages. In this article, we will explore the different types of solar panels, jump into their manufacturing methods, and highlight their pros and cons.
Monocrystalline Solar Panels:
Monocrystalline solar panels are made from a single crystal structure, typically silicon. These panels have a sleek black appearance and high efficiency due to the purity of the silicon used. Manufacturing involves slicing thin wafers from a cylindrical crystal, resulting in a uniform and consistent panel. Pros include excellent efficiency, space-saving design, and longevity. However, monocrystalline panels tend to be more expensive to produce.
Polycrystalline Solar Panels:
Polycrystalline solar panels are manufactured from multiple silicon crystals, which are melted together to form the panel. They have a distinctive blue color and a lower efficiency compared to monocrystalline panels. Manufacturing polycrystalline panels is a simpler process, making them more cost-effective. Although their efficiency is slightly lower, they are still a popular choice due to their affordability.
Thin-Film Solar Panels:
Thin-film solar panels are made using a different technique that involves depositing a thin layer of semiconductor material, such as amorphous silicon or cadmium telluride, onto a substrate. This manufacturing process allows for flexibility and lightweight designs, making thin-film panels suitable for unconventional applications. However, they have lower efficiency and tend to degrade faster than crystalline panels.
Bifacial Solar Panels:
Bifacial solar panels capture sunlight from both sides, increasing their overall energy generation. They can be monocrystalline or polycrystalline, with the rear side exposed to reflected sunlight or ground albedo. Bifacial panels are versatile and can be used in various installations, such as ground-mounted or rooftop systems. They offer increased energy output but may require specific mounting and installation considerations.
Concentrated Photovoltaic (CPV) Solar Panels:
CPV solar panels use lenses or mirrors to concentrate sunlight onto small, high-efficiency solar cells. This design enables CPV panels to achieve remarkably high efficiencies, but they require precise tracking systems to keep them aligned with the sun. CPV panels are suitable for areas with high direct sunlight, such as deserts, but are less common in residential or urban settings due to their specialized requirements.
Conclusion:
Images Source:
Solar panels come in a variety of types, each with its own manufacturing process, benefits, and drawbacks. Monocrystalline panels offer high efficiency but at a higher cost, while polycrystalline panels provide a more affordable option with slightly lower efficiency. Thin-film panels are lightweight and flexible but have lower overall efficiency. Bifacial panels capture sunlight from both sides, enhancing energy generation, while CPV panels achieve exceptional efficiency but require precise tracking mechanisms. Understanding the characteristics and trade-offs of each solar panel type empowers individuals and businesses to make informed decisions when adopting solar energy solutions. With continuous advancements in technology, solar power continues to pave the way towards a cleaner and more sustainable future.
Exploring the World of The Weeknd Merch and Bad Bunny Merch
Merchandise has become an integral part of the music industry, allowing fans to express their love and support for their favorite artists. In recent years, two artists who have captured global attention with their unique sound and style are The Weeknd and Bad Bunny. Both artists have not only made a significant impact on the music scene but have also created an array of merchandise that resonates with their fans. In this article, we delve into the world of The Weeknd merch and Bad Bunny merch, exploring their unique offerings and the cultural significance behind them.
The Weeknd Merch
Abel Tesfaye, famously known as The Weeknd, has carved a distinct niche in the music industry with his soulful vocals and captivating lyrics. The Weeknd Merch reflects his dark and mysterious aesthetic, often inspired by his album artworks and music videos. His merchandise line offers a wide range of products, including clothing, accessories, and collectibles. From hoodies and t-shirts featuring his iconic "XO" logo to bomber jackets and hats adorned with his signature motifs, fans can find an assortment of items that embody The Weeknd's unique style. The merchandise often incorporates a mix of streetwear and luxury fashion elements, appealing to a diverse fan base.
One notable aspect of The Weeknd's merch is the storytelling behind it. Each collection is carefully curated to reflect the theme of his albums, allowing fans to delve deeper into his artistic vision. For example, the "After Hours" merchandise collection features designs inspired by the album's dark and enigmatic concept, with references to the red suit and bloody nose from the album cover. This attention to detail creates a sense of connection between the music and the merchandise, enhancing the overall fan experience.
Bad Bunny Merch
Bad Bunny, the Puerto Rican superstar, has taken the music industry by storm with his infectious reggaeton beats and vibrant personality. His merch reflects his vibrant and flamboyant style, incorporating elements of Latin American culture and street fashion. The Bad Bunny merchandise line offers an eclectic mix of products, including clothing, accessories, and even toys. Fans can find everything from colorful t-shirts and hoodies featuring his iconic bunny logo to bucket hats and earrings inspired by his bold fashion choices.
What sets Bad Bunny Merch apart is its celebration of Latinx culture. His merchandise often incorporates slogans and phrases in Spanish, paying homage to his roots and connecting with his diverse fan base. Additionally, his merchandise showcases a fusion of urban and traditional aesthetics, representing the dynamic cultural landscape from which he draws inspiration.
Collaborations and Limited Editions:
Both The Weeknd and Bad Bunny have collaborated with renowned brands and designers to create limited-edition merchandise, further elevating the exclusivity and desirability of their products. Collaborations with fashion brands like Off-White and collaborations with streetwear icons like BAPE have added a unique touch to their merchandise, appealing to fans who appreciate fashion and collectible items.
These limited-edition collaborations often sell out quickly, creating a sense of urgency and excitement among fans. They also serve as a platform for artists to showcase their creative vision beyond music, blurring the lines between music, fashion, and art.
Cultural Impact:
The merchandise of The Weeknd and Bad Bunny goes beyond simple branding and commercial ventures. It serves as a symbol of cultural influence, representing the artists' impact on their respective genres and fan communities. The popularity and widespread adoption of their merchandise highlight the artists' ability to connect with their fans on a deeper level and create a sense of belonging.
Conclusion:
The Weeknd merch and Bad Bunny merch offer fans an opportunity to express their devotion and connect with their favorite artists on a tangible level. Through their carefully curated collections, The Weeknd and Bad Bunny | eng | 663af2f9-d7de-405b-b801-5b7f511589b1 | https://insiderhubs.com/the-different-types-of-solar-panels-manufacturing-the-pros-and-cons/ |
You've seen The Princess Bride, right? I mean, everyone and their lover has seen it. Anyways, in case you haven't, here is the important part. This guy uses the word "Inconceivable!" all the time, and most of the time, it's not quite right. A funny part of the move is when another person says "I do not think that word means what you think it means."
Why the long lead in? Because I feel like accessible is a word that all sorts of people use, and I just want to look at them and say "I do not think that word means what you think it means."
From a disability centric stand point, accessible is being used more and more commonly. Places say they are ADA accessible, events put accessible on their flyers and more.
However, I'd like to make a few statements about this.
Firstly, ADA accessible has a lot of specifications. Most places that claim to be accessible are not officially ADA accessible. Do you have a handicap MARKED stall in the restroom, with a railing? If not, you're not accessible by ADA standards. Please don't say that you are. Now, it's ok not to be ADA accessible – a lot of smaller locations, especially those in older buildings, can't get up to that code. However, if you can't, don't say that you have done so, as it pisses people off when they show up, and still can't have the access that they need.
Secondly, please take multiple types of disability into account. That's great that you have an elevator, but if you have to go up stairs to get to it, it's NOT accessible. Same goes for doorways that are too narrow to fit most wheelchairs. Not accessible. Are you calling your event accessible? To many people, that means that there will be an ASL interpreter for the hearing impaired. If you don't have one, and it's a spoken event/lecture, how accessible is it really? Please just stop and think for a second before using that word.
Now let's look at other definitions of accessible, particularly within the sex positive community.
Accessible can mean affordable. Don't forget that people of ALL socio-economic backgrounds and income levels deserve to be able to learn about sex education, and/or attend play parties. If all of your events are $30+ (or even $10 with no sliding scale, or way to volunteer to offset the entry fee), you're excluding entire communities.
How about accessible via public transit? Now, not all cities have public transit that goes all places, but if you live in a city that has pretty good mass transit, please consider that when planning events, or opening new stores/centers/dungeons. Lots of people don't have cars (or the ability to pay for cabs) , but still are interested in attending events and learning about sex ed. Let's make sure to be inclusive there.
The list goes on and on. Accessible can mean having bilingual person at events, helping to explain what is going on to people for whom English is not their first language. It can mean advertising in communities outside those that usually get to know about sex education workshops and kink events. In can mean going after under-represented communities to find out what would make an event/class more accessible to them. It can mean many things…but often, I do not think that word means what you think it means. | eng | f9962714-df55-438c-99eb-d123564491e6 | https://www.fearlesspress.com/2013/12/03/i-do-not-think-that-word-means-what-you-think-it-means/ |
Innovative Applications in Smart Cities
036782096X, 9780367820961
This book is a compilation of chapters on scientific work in novel and innovative reference that compiles interdisciplin
Table of contents : Cover Title Page Copyright Page Preface Contents Prologue Part I: Daily Life in a Smart City 1. Segmentation of Mammogram masses for Smart Cities Health Systems 2. Serious Game for Caloric Burning in Morbidly Obese Children 3. Intelligent Application for the Selection of the Best Fresh Product According to its Presentation and the Threshold of Colors Associated with its Freshness in a Comparison of Issues of a Counter in a Shop of Healthy Products in a Smart City 4. Analysis of Mental Workload on Bus Drivers in the Metropolitan Area of Querétaro and its Comparison with three other Societies to Improve the Life in a Smart City 5. Multicriteria analysis of Mobile Clinical Dashboards for the Monitoring of Type II Diabetes in a Smart City 6. Electronic Color Blindness Diagnosis for the Detection and Awareness of Color Blindness in Children Using Images with Modified Figures from the Ishihara Test 7. An Archetype of Cognitive Innovation as Support for the Development of Cognitive Solutions in Smart Cities Part II: Applications to Improve a Smart City 8. From Data Harvesting to Querying for Making Urban Territories Smart 9. Utilization of Detection Tools in a Human Avalanche that Occurred in a Rugby Stadium, Using Multi-Agent Systems 10. Humanitarian Logistics and the Problem of Floods in a Smart City 11. Simulating Crowds at a College School in Juarez, Mexico: A Humanitarian Logistics Approach 12. Perspectives of State Management in Smart Cities Part III: Industry 4.0, Logistics 4.0 and Smart Manufacturing 13. On the Order Picking Policies in Warehouses: Algorithms and their Behavior 14. Color, Value and Type Koi Variant in Aquaculture Industry Economic Model with Tank's Measurement Underwater using ANNs 15. Evaluation of a Theoretical Model for the Measurement of Technological Competencies in the Industry 4.0 16. Myoelectric Systems in the Era of Artificial Intelligence and Big Data 17. Implementation of an Intelligent Model based on Big Data and Decision Making using Fuzzy Logic Type-2 for the Car Assembly Industry in an Industrial Estate in Northern Mexico 18. Weibull Reliability Method for Several Fields Based Only on the 235 Modeled Quadratic Form Index
Preface "Innovation" is a moto in the development of current and future Smart Cities. Innovation understood by newness, improvement and spread, is often promoted by Information and Communication Technologies (ICTs) that make it possible to automate, accelerate and change the perspective of the way economy and "social good" challenges can be addressed. In economics, innovation is generally considered to be the result of a process that brings together various novel ideas to affect society and increase competitiveness. In this sense, future Smart Cities societies' economic competitiveness is defined as increasing consumers' satisfaction given by the right products price/quality ratio. Therefore, it is necessary to design production workflows that maximise the resources used to produce the right quality products and services. Companies' competitiveness refers to their capacity to produce goods and services efficiently (decreasing prices and increasing quality), making their products attractive in global markets. Thus, it is necessary to achieve high productivity levels that increase profitability and generate revenue. Beyond the importance of stable macroeconomic environments that can promote confidence, attract capital and technology, a necessary condition to build competitive societies is to create virtuous creativity circles that can propose smart and disruptive applications and services that can spread across different social sectors strata. Smart Cities have been willing to create technology-supported environments to make urban, social and industrial spaces friendly, competitive and productive contexts in which natural and material resources can be accessible to people, where citizens can develop their potential skills in the best conditions possible. Since countries in different geographic locations, natural, cultural and industrial ecosystems have to adapt their strategies to these conditions, Smart Cities solutions are materialised differently. This book shows samples of experiences where industrial, urban planning, health and sanitary problems are addressed with technology leading to disruptive data and artificial intelligence centred applications. Sharing applied research experiences and results mostly applied in Latin American countries, authors and editors show how they contribute to making cities and new societies smart through scientific development and innovation.
3. Intelligent Fernando Maldonado-Azpeitia and Alberto Ochoa-Zezzatti
22
4. Analysis of Mental Workload on Bus Drivers in the Metropolitan Area of Querétaro and its Comparison with three other Societies to Improve the Life in a Smart City Aarón Zárate, Alberto Ochoa-Zezzatti, Fernando Maldonado and Juan Hernández
34
5. Multicriteria analysis of Mobile Clinical Dashboards for the Monitoring of Type II Diabetes in a Smart City Mariana Vázquez-Avalos, Alberto Ochoa-Zezzatti and Mayra Elizondo-Cortés
7. An Archetype of Cognitive Innovation as Support for the Development of Cognitive Solutions in Smart Cities Jorge Rodas-Osollo, Karla Olmos-Sánchez, Enrique Portillo-Pizaña, Andrea Martínez-Pérez and Boanerges Alemán-Meza
18. Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form Manuel R. Piña-Monarrez and Paulo Sampaio
235
Index
229
267
Prologue Khalid Belhajjame
Nowadays, through the democratisation of internet of things and highly connected environments, we are living in the next digitally enriched generation of social media in which communication and interaction for user-generated content are mainly focused on improving the sustainability of smart cities. Indeed, the development of digital technologies in the different disciplines in which cities operate, either directly or indirectly, alters expectations among those in charge of the local administration and of citizens. Every city is a complex ecosystem with a lot of subsystems to make it work, such as work, food, clothes, residence, offices, entertainment, transport, water, energy, etc. As they grow, there is more chaos and most decisions are politicized, there are no common standards and data is overwhelming. The intelligence is sometimes digital, often analogue, and almost inevitably human. The smart cities initiative aims to better exploit the resources in a city to offer higher-level services to people. Smart cities are related to sensing the city's status and acting in new intelligent ways at different levels: people, government, cars, transport, communications, energy, buildings, neighbourhoods, resource storage, etc. A smart city is much more than a high-tech city, it is a city that takes advantage of the creativity and potential of new technologies to meet the challenges of urban life. A smart city also helps to solve the sensitive issues of its citizens, such as insecurity, urban mobility problems, water resources management and solid waste. They are not the instruments that make a smart city, but everything that is achieved through the implementation of those processes. A vision of the city of the "future", or even the city of the present, rests on the integration of science and technology through information systems. This vision implies re-thinking the relationships between technology, government, city managers, business, academia and the research community. Conclusions and actions are determined by the social, cultural and economic reality of the cities and the countries in which they are located. Therefore, beyond smart cities as an object of study, it is important to think about urban spaces that can host smart cities of different types and built with different objectives, maybe those that have more priority and that can ensure the wellbeing of citizens.
Smart Cities and Urban Computing applications Smart cities where the focus was on how cable, telephones and other wired media were changing our access to services [1,8,22] have been around for a long time. Today, the concept "smart city" refers to a world initiative leading to better exploit the resources in a city to offer value-added services to people [15] considering at least four components: industry, education, participation, and technical infrastructure [5].
viii Innovative Applications in Smart Cities The advent of technology and the existence of the internet have helped transform traditional cities into cities that are more impressive and interactive. There are terms analogous to "smart cities", such as a digital, intelligent, virtual, or ubiquitous city. The definition and understanding of these terms determine the way challenges are addressed and projects are proposed to go towards a "smart urban complex ideal" [2,6,21]. Overall, the evolution of a city into a smart city must focus on the fact that network-based knowledge must not only improve the lives of those connected but also bring those who remain unconnected into the fold, creating public policies that truly see the problems faced by big cities and everyday citizens. Several smart cities in the most important capitals of the world and wellknown touristic destinations have developed urban computing solutions to address key issues of city management, like transport, guidance to monuments, e-government, access to leisure and culture, etc. In this way, citizens of different socio-economic groups, investors and government administrators can have access to the resources of the city in an optimised and personalised manner. Thus, a more intelligent and balanced distribution of services is provided thanks to technology that can improve citizens life and opportunities. This has been more or less possible in cities where the socio-economic and technological gap is not too great. Normally, solutions assume that cities provide good quality infrastructure, including internet connection, access to services (energy, water, roads, health), housing, urban spaces, etc. Yet, not all cities are developed in these advantageous conditions, there are regions in the world where exclusion prevails in cities and urban spaces, where people have little or no access to electricity, technology and connectivity, and where services are not regulated. It is in this type of city that smart cities technology and solutions face their greatest challenges. In Mexico, projects on Smart Cities have been willing to promote sustainable urban development through innovation and technology. The objective of the smart cities project has addressed the improvement of life quality for inhabitants. Areas promoted in Smart Cities in Mexico are quite diverse, ranging from environment, safety and urban design to tourism and leisure. This book describes solutions to problems in these areas. Chapters describing use cases are also analysed to determine the degree of improvement of citizens quality of life, human logistics within urban spaces, of the logistics strategies and the access and distribution of services like transport, health or assistance during disasters and critical events. The experiments described along the chapters of the book are willing to show the way academics, inspired in living labs promoted in other cities, have managed to study major smart cities problems and provide solutions according to the characteristics of the cities, the investment of governments and industry and the willingness of people to participate in this change of paradigm. Indeed, citizen participation is a cornerstone that must not be left aside. After all, it is citizens who are beginning transformation and who constantly evaluate the results of information integration. Citizen satisfaction is the best way to calibrate a smart city's performance. Urban computing1 is defined as the technology for acquisition, integration, and analysis of big and heterogeneous data generated by a diversity of sources in urban spaces, such as sensors, devices, vehicles, buildings, and human, for tackling the major issues that cities face. The study of smart cities as complex systems is addressed through this notion of urban computing [23]. Urban computing brings computational techniques to bear on urban challenges such as pollution, energy consumption, and traffic congestion. Using today's large-scale computing infrastructure and data gathered from sensing technologies, urban computing combines computer science with urban planning, transportation, environmental science, sociology, and other areas of urban studies, tackling specific problems with concrete methodologies in a data-centric computing framework.
Table 1 presents a summary of the families of applications that can be developed in the context of urban computing: urban planning, transportation, environment, social and entertainment, energy, economy and safety and security. Often, these applications can be organised on top of a general urban computer framework reference architecture, enabling platforms that provide the technical underlying infrastructure [16] necessary for these applications to work and be useful for the different actors populating and managing urban territories. In urban computing, it is vital to be able to predict the impact of change in a smart city's setting. For instance, how will a region's traffic change if a new road is built there? To what extent will air pollution be reduced if we remove a factory from a city? How will people's travel patterns be affected if a new subway line is launched? Being able to answer these kinds of questions with automated and unobtrusive technologies will be tremendously helpful to inform governmental officials' and city planners' decision making. Unfortunately, the intervention-based analysis and prediction technology that can estimate the impact of change in advance by plugging in and out some factors in a computing framework is not well studied yet. The objective would be to use this technology to reduce exclusion and make citizens' life more equal. How to guide people through urban spaces with little or no land registry? How to compute peoples' commute from home to work when transport is not completely regulated? How to give access to services through applications that can be accessible for all? For example, Latin American cities that appear as first in the Smart Cities rankings (Buenos Aires, Santiago de Chile, São Paulo, Mexico City) are megacities of more than 10 million inhabitants. For many Smart Cities ideologies, the big urban "spots" are the antithesis of the ideas and values of a truly smart city. Thus, there is room for scientific and technological innovation to design smart cities solutions in these types of cities and thereby tackle requirements that will make citizens' lives better. This book is original in this sense because it describes smart cities solutions for problems in this type of city. It provides use case examples of prediction solutions for addressing not only smart cities issues, but urban computing as a whole.
Smart Cities as Living Laboratories At the beginning of 2013, there were approximately 143 ongoing or completed self-designated smart city projects: North America had 35 projects, Europe 47, Asia 50, South America 10, and
x Innovative Applications in Smart Cities the Middle East and Africa 10. In Canada, Ottawa's "Smart Capital" project involves enhancing businesses, local government, and communities using Internet resources. Quebec was a city highly dependent upon its provincial government because of its weak industry until the early 1990s when the city government kicked off a public-private partnership to support a growing multimedia sector and high-tech entrepreneurship. In the United States, Riverside (California) has been improving traffic flow and replacing ageing water, sewer and electric infrastructure through a tech-based transformation. In San Diego and San Francisco, ICT have been major factors in allowing these cities to claim to be a "City of the Future" for the last 15 years. Concerning Latin America, the Smart Cities council recognizes the eight smartest cities in Latin America: Santiago (Chile), Mexico City (Mexico), Bogota (Colombia), Buenos Aires (Argentina), Rio de Janeiro (Brazil), Curitiba (Brazil), Medellin (Colombia) and Montevideo (Uruguay). Each city focusses on different aspects, including automating pricing depending on traffic, smart and eco-buildings, electrical and eco-subway, public Wi-Fi and public tech job programs, weather, crime, emergency monitoring, university and educational programs. In Mexico, the first successful Smart City project "Ciudad Maderas" (2013–2020) was developed in Querétaro in the central part of the country. This project included the construction of technology companies, hotels, schools, shopping centres, residential areas, churches and huge urban spaces dedicated as a natural reserve in El Marques district. The purpose has been to integrate technological developments into the daily lives of Queretaro's inhabitants. Concerning e-governance, the State Government has launched the Querétaro Ciudad Digital Application. The purpose of this application is to narrow the gap between the citizens and the government. The application is regarded worldwide as second-to-none technology. Cities like Mexico City have focused on key services, such as transportation. A wide range of applications is readily available to residents to accomplish their daily journeys from A to B: Shared Travel Services, Uber, Easy, Cabify. Since 2014, the city of Guadalajara has been working on the Creative Digital City project to promote the digital and creative industry in the region. The city of Tequila, also in the state of Jalisco, promotes the project "Intelligent Tequila" for attracting tourism to the region. One of the smart technologies already in use are heat sensors which help to measure massive concentrations in public places. In Puebla, the Smart Quarter project develops solutions for improving mobility, safety and life quality for the inhabitants. For example, proposing free Wifi in public areas and bike tracks equipped with video – monitoring and alarm systems. The European Union has put in place smart city actions in several cities, including Barcelona, Amsterdam, Berlin, Manchester, Edinburgh, and Bath. In the United Kingdom, almost 15 years ago, Southampton claimed to be the country's first smart city after the development of its multi-application smartcard for public transportation, recreation, and leisure-related transactions. Similarly, Tallinn has developed a large-scale digital skills training program, extensive e-government, and an awardwinning smart ID card. This city is the centre of economic development for all of Estonia, harnessing ICT by fostering high-tech parks. The European Commission has introduced smart cities in line 5 of the Seventh Framework Program for Research and Technological Development. This program provides financial support to facilitate the implementation of a Strategic Energy Technology plan [4] through schemes related to "Smart cities and communities". Statistics of the Chinese Smart Cities Forum report six provinces and 51 cities have included Smart Cities in their government work reports in China [14,17]; of these, 36 are under new concentrated construction. Chinese smart cities are distributed densely over the Pearl and Yangtze River Deltas, Bohai Rim, and the Midwest area. Moreover, smart cities initiatives are spread in all first-tier cities, such as Beijing, Shanghai, and Shenzhen. The general approach followed in this city is to introduce some ICT during the construction of new infrastructure, with some attention to environmental issues but limited attention to social aspects. A modern hi-tech park in Wuhan is considered an urban complex that is multi-functional and ecological. Wuhan is a city that is high-tech and self-sufficient. It is an eco-smart city designed for exploring the future of the city. It is a natural and healthy environment and incubator of high culture and expands the meaning of
Prologue xi
the modem hi-tech park [18]. Taipei City clarified that the government must provide an integrated infrastructure with ICT application and service [10]. China has encouraged the transition to urbanism by improving public services and improving efficiency for transformation to the government model, enhancing the economic development of the city. In 2008, a "digital plateau" was proposed; in 2009, more than ten provinces set goals to build a smart city. China improved the construction of the city, industrial structure, and social development. To start the implementation strategy, a good plan is necessary, as well as knowledge of the importance of smart city construction. Several Southeast Asian cities, such as Singapore [11], Taiwan, and Hong Kong, are following a similar approach, promoting economic growth through smart city programs. Singapore's IT2000 plan was designed to create an "intelligent island," with information technology transforming work, life, and play. More recently, Singapore has extensively been dedicated to implementing its Master Plan in 2015 and has already completed the [email protected] goal of providing free mobile Internet access anywhere in the city [7]. Taoyuan in Taiwan is supporting its economy to improve the quality of living through a series of government projects such as E-Taoyuan and U-Taoyuan for creating e-governance and ubiquitous possibilities. Korea is building the largest smart city initiative in Korea, Songdo, a new town built from the ground in the last decade and which plans to house 75,000 inhabitants [13]. The Songdo project aims at developing the most wired smart city in the world. The project is also focused on buildings and has collaborated with developers to build its networking technology into new buildings. These buildings will include telepresence capabilities and many new technologies. The plan includes installing telepresence in every apartment to create an urban space in which every resident can transmit information using various devices [12], whereas a city central brain should manage the huge amount of information [19]. This domestic offering is only the first step; Cisco aims to link energy, communications, traffic, and security systems into one smart network [20]. At present, there are 13 projects in progress towards the smart city initiatives of New Songdo [9]. Despite an increase in projects and research to create smart cities, it is still difficult to provide cities with all the features, applications, and services required. Future smart cities will require a rethinking of the relationships between technology, government, city managers, business, academia and the research community. This book shows some interesting results of use cases where different communities and sectors interact to find alternative solutions for cities that are willing to become smart and address their problems innovatively and effectively.
Driving Urban Phenomena Foresight Data science and urban computing have been developing and applying a great number of analytics pipelines to model and predict the behaviour of urban spaces as complex systems. For example, there are projects devoted to addressing adaptive street lighting to modulate public lighting, i.e., to locally adjust the luminous intensity of each lamp post according to parameters, and to take into account maintenance of the equipment as precisely as possible (anticipate failures). This regulation is done according to external conditions: luminosity, but also humidity level or even with presence sensors (pedestrians coming, car traffic). Systems for industrial predictive maintenance and prioritisation of interventions can be implemented applying data science pipelines that can use operational research/ multi-criteria optimization applied to light modulation. The expected benefits are a reduction in consumption and maintenance costs and an improvement in the quality of service (feeling of security for citizens, immediate replacement of defective streetlamps). Finally, by adjusting light intensity, cities reduce the level of light pollution, thus improving their aesthetics and their impact on the immediate environment. Multi-channel urban transport projects also introduce phenomena and situations foresight requirements. Urban transport can exploit collected data sharing in order to offer the passenger a global offer based on all the means of transport in a city, i.e., multi-channel transport. An intermodal
xii Innovative Applications in Smart Cities platform can consolidate information on the use and operation of all means of transport at the local authority level (bus, tramway, bicycles, car, transport conditions). Thanks to this consolidated and interpreted information, the city can offer its citizens the most appropriate solution, taking into account the context and requirements of the traveller. In terms of Data Science, just as for parking in the city, intermodality is first and foremost a subject of optimisation of resources under constraints. More broadly, the addition of new data (video, traffic conditions) and the identification of nonlinear patterns (formation of traffic jams, congestion measurement) makes the subject rich and complex. Finally, scoring and customer knowledge algorithms are exploited to take into account user preferences to improve the recommendation. Perhaps it is not necessary to systematically propose bicycle travel times to a daily bus user? Providing fluid displacement of people within urban spaces is also an important problem that can be solved through data science. Focussing on traffic management that is a daily problem in cities, the coordination of traffic lights can be an effective means of regulating road traffic and limiting congestion situations. By smoothing the flow and reducing the number of vehicles passing through bottlenecks, it is possible to increase the daily flow and reduce the level of traffic congestion. For instance, the reduction in speed on the ring road has had the effect of reducing traffic jams and congestion. Despite the counter-intuitive side of this effect, the physical explanation comes from the fact that by reducing the maximum speed, the amplitude of speed variations has been reduced (fluid mechanics enthusiasts already know that it is better to be in laminar rather than chaotic flow situations). Another contribution of the knowledge of traffic conditions is the use of models to test the impact of road works or the construction of new infrastructures. In terms of data science, the aim is to identify forms of congestion and to detect and use the most effective means of contrasting them. Among them, one can generally act on the maximum speed allowed or on the regulation of the timing of traffic lights according to traffic conditions (detected via cameras or GPS signals), visibility conditions and, in general, the weather, the presence of pedestrians, the time of day or other parameters that emerge as significant. Perhaps the most direct benefit expected is the reduction of traffic jams and slowdowns and, indeed, the time required for a trip. Other collateral benefits are the reduction of air and noise pollution and a reduction in the number of accidents caused by traffic problems. This book provides examples of solutions that can be proposed through data science solutions that apply machine learning, data mining and artificial intelligence methods to data intended to make people live better in their daily lives as citizens of cities with unfair distribution of services. It also shows how data science can contribute to human logistics problems, particularly in the presence of critical events happening in cities with few infrastructures or with a huge population. The use cases are issued from the Mexican context, but they can be present in Latin American cities and, thus, solutions can be reproduced in cities with similar characteristics. Having systems and solutions that can promote foresight and planning are key in these kinds of urban places. Book content and organisation The book consists of eighteen chapters organized into three parts, namely, Daily Life in a Smart City (part I), Applications to Improve a Smart City (Part II) and Industry 4.0, Logistics 4.0 and Smart Manufacturing (Part III). These general topics address problems regarding smart cities as environments where citizens behaviour, health, commercial preferences and use of services (e.g., transport) can be observed. As mentioned before, the originality of the chapters is that they address topics regarding cities in Latin American countries, in particular Mexican cities, where citizens behaviour models change given the socio-cultural diversity and the unequal access to services. - Part I: Daily Life in a Smart City consists of seven chapters that focus on important problems in Latin American smart cities. In these smart cities, solutions must deal with massive protocols that can use technology to develop and implement strategies for dealing with diseases common
Prologue xiii
in the Latin American population, like obesity, breast cancer, colour-blindness and mental workload. - Part II: Applications to Improve a Smart City. The way people move around in cities and urban spaces gives clues as to when many events of interest come up and in which hotspots. Part II of the book consists of five chapters that describe technology-based techniques and approaches to observing civilians' behaviour and their quality of life. This part starts with an initial chapter that surveys techniques for dealing with smart cities data, including collection strategies, indexing and exploiting them through applications. Then, the four remaining chapters address the analysis of citizens' behaviour with the aim of proposing strategies for dealing with human logistics in the presence of critical events (human avalanches, floods) and the way services distribution to the population can be improved (distribution of shelter and evacuation routes). - Part III: Industry 4.0, Logistics 4.0 and Smart Manufacturing. Mexico is the Latin American country with the highest number of Smart Cities that offer economic advantages for they represent niches enabling potential economic activities that have not been considered yet; for example, the 4.0 technology sustainable promotion, the energy and agricultural sectors. A wide range of growth opportunities lie ahead for companies falling in these categories. In the long run, Smart Cities push forward towards Mexican economy diversification. This part of the book consists of six that address the important services that activate the economy of smart cities. Indeed, smart manufacturing is an important activator of industry smart cities, and it is activated through techniques that are being developed in Industry and Logistics 4.0 ecosystems. Along with its six, this part addresses algorithms for managing orders in warehouses and supply chains in sectors like automotive industry and retail; and the impact of using technology and data analytics methods in the aquaculture industry. In conclusion, this book is important because it shows that it is possible and important to show how key problems in non-ideal urban contexts, like the ones that characterize some cities in Latin America and particularly in Mexico, can find solutions under the perspective of smart cities. Being smart is maybe a good opportunity for academia, government and industry to work with society and find alternatives to reduce unequal access to services, as well as sustainability and exclusion in the complex urban megapolis.
One of the fundamental aspects of smart cities is an improvement in the health sector, by providing its citizens with better care and prevention and detection of diseases. Breast cancer is one of the most common diseases and the one with the highest incidence in women worldwide. In smart cities, to improve the quality of life of its citizens, especially for women, is to diagnose breast tumors in shorter periods with simpler and automated methods. In this chapter, a new deep learning architecture is proposed to segment breast cancer tumors.
1. Introduction According to the World Health Organization (WHO), breast cancer is one of the most common diseases with the highest incidence in women worldwide, about 522 thousand deaths are estimated annually with data collected in 2012 (OMS, 2019). In smart cities, as regards to the health sector, it seeks to improve the quality of life of its citizens (Kashif et al., 2020; Abdelaziz et al., 2019; Rathee et al., 2019), thus there is a need to diagnose breast tumors in shorter periods with simpler and automated methods that can produce accurate results. The most common method to an early diagnostic is through mammographic images, however, these images usually have noise and low contrast which can cause the doctor to have difficulty classifying different tissues. In some mammogram images, malignant tissues and normal dense tissues are presented, but it is difficult to contrast between them by applying simple thresholds when automatic methods are used (Villalba, 2016). Because of these problems, it is necessary to develop various approaches that can correctly identify the malignant tissues, which represent higher intensity values compared to background information and other regions of the breast. Also, regions where some normal dense tissues have intensities similar to the tumor region have to be excluded (Singh et al., 2015).
2 Innovative Applications in Smart Cities The interpretation of a mammogram is usually difficult, sometimes it depends on the experience of medical staff. In approximately 9% of the cancers detected, tumors were visible on mammograms obtained from two years earlier (Gutiérrez et al., 2015). The key factor for early detection is the use of computerized systems. The segmentation of tumors takes a very important role in the diagnosis and timely treatment of breast cancer. Currently, there are methods to delimit tumors using artificial neuronal networks (Karianakis et al., 2015; Rafegas et al., 2018) and deep learning networks (Hamidinekoo et al., 2018; Goodfellow et al., 2016), but there is the possibility of improving them. In this chapter, a new architecture with aims to segment mammary tumors in mammograms using deep neural networks is proposed.
2. Literature Review There exist several diagnostic methods to perform timely detection, the use of mammography being the method most used by medical staff because of the effective and safe results of the method. The examination is carried out by firm compression of the breast between two plates, using ionizing radiation to obtain images of breast tissue, which can be interpreted as benign or malignant (Marinovich et al., 2018). Here, we review some methods for an automatic segmentation/detection of malignant masses by processing the mammography image. In (Hanmandlu et al., 2010), a comparison of two different semi-automated methods was performed, using level sets method and watershed controlled by markers. Although both methods are not very accurate, they were found to have a short processing time. In the work of (Lempka et al., 2013), two automated methods were presented based on the improvement of region growing and the segmentation with cellular neural networks. In the first stage, the segmentation was carried out through an automated region growing whose threshold is obtained through an artificial neural network. In the second method, segmentation is performed by cellular neural networks, whose parameters are determined by a genetic algorithm (GA). Intensity, texture and shape characteristics are extracted from segmented tumors. The GA is used to select appropriate functions from the set of extracted functions. In the next stage, ANNs are used to classify mammograms as benign or malignant. Finally, they evaluate the performance of different classifiers with the proposed methods, such as multilayer perceptron (MLP), vector support machines (SVM) and K-nearest neighbors (KNN). Among these methods, the MLP produced better diagnostic performance in both methods. The sensitivity, specificity and accuracy indices obtained are 96.87%, 95.94%, and 96.47%, respectively (Lempka et al., 2013). In (Wang et al., 2014), a breast tumor detection algorithm was proposed in digital mammography based on extreme machine learning (ELM). First, they use a median filter for noise reduction and contrast improvement as data pre-processing. Next, wavelet transforms, morphological operations, and the region growing are used for the segmentation of the edge of the breast tumor. Then, they extract five textural features and five morphological features. Finally, they use ELM classifier to detect breast tumors. In the comparison of the detection of breast tumors based on SVM, with the detection of breast tumors based on ELM, not only does ELM have a better classification accuracy than the SVM, but also a much-improved training speed. Also, the efficiency of classification, training and performance testing of SVM and ELM were compared and the total number of errors for ELM was 84 while the total number of errors for SVM was 96, showing that ELM has better abilities than SVM (Wang et al., 2014). In (Pereira et al., 2014), a computational method is presented as an aid to segmentation and mass detection in mammographic images. First, a pre-processing method based on Wavelets transformation and Wiener filtering was applied for image noise removal and enhancement. Subsequently, a method was used for mass detection and segmentation through the use of thresholds, the Wavelet transform, and a genetic algorithm. The method was quantitatively evaluated using the area overlay metric (AOM). The mean standard deviation value of AOM for the proposed method was 79.2% ± 8%. The
Segmentation of Mammogram masses for Smart Cities Health Systems 3
method they propose presented a great potential to be used as a basis for the massive segmentation of mammograms in the craniocaudal and mid-lateral oblique views. The work of (Pandey et al., 2018) presents an automatic segmentation approach that is carried out in three steps. First, they used adaptive Wiener filtering and media clustering to minimize the influence of noise, preserve edges and eliminate unwanted artifacts. In the second step, they excluded the heart area using a set of levels based on the active contour, where the initial contour points were determined by the maximum entropy threshold and the convolution method. Finally, the pectoral muscle is removed through the use of morphological operations and local adaptive thresholds in the images. The proposed method was validated using 1350 breast images of 15 women, showing excellent segmentation results compared to semi-automated methods drawn manually. In (Chougrad et al., 2018) a computerized diagnostic system was developed based on deep convolutional neural networks (CNN) that use transfer learning, which is ideal for handling small data sets, such as medical images. After training some CNN architectures, they used precision and AUC parameters to evaluate images from different databases, such as DDSM, INbreast, and BCDR. The CNN model, named Inception v3, obtained the best results with an accuracy of 98.94%, and so it was used as a basis to build the Breast Cancer Screening Framework. To evaluate the proposed CAD system and its efficiency to classify new images, they tested it in a database different from those used previously (MIAS) and obtained an accuracy of 98.23% and 0.99 AUC (Chougrad et al., 2018).
3. Methodology In this section, we present the methodology used for the development of an architecture based on deep learning neural networks, for segmenting tumors in digital mammography. The schematic of the methodology is presented in Figure 1.
Figure 1: Diagram of the methodology.
The images used in this chapter are from the CBIS-DDSM which is a subset of the Database for Screening Mammography (DDSM) which is a database with 2620 mammography studies. It contains normal, benign and malignant cases with verified pathological information. This database is a useful tool in the testing and development of decision support systems. The CBIS-DDSM collection includes a subset of the DDSM data selected by a trained medical doctor. The images were decompressed and converted to the DICOM format. The database also includes updated ROI segmentation and delimitation tables, and pathological diagnosis for training data (Lee et al., 2017). In ROI annotations for anomalies in the CBIS-DDSM data subset, they provide the exact position of the lesions and their coordinates to generate the segmentation mask. In Figure 2, three images obtained from the CBIS-DDSM database are shown, with tumors. 3.1 Preprocessing of the images The original size of the images obtained from the database was 6511 × 4801 pixels. A mask for the tumor contained in each image was obtained from the database. Then, the region that contained the tumor on both image and mask were trimmed, and the clipping was reduced to a 60 × 60-pixel image, this procedure was done to make the network faster and save memory since the
4 Innovative Applications in Smart Cities
Figure 2: Examples of images obtained from the CBIS-DDSM database; images are shown with false color.
original size was too large to process. The procedure was repeated for each image in the database. In Figure 3, it is shown an image from the database, and the clipping of the region of the tumor and its mask.
3.2 Deep learning architecture design The network architecture is a modification of the architecture of (Guzman et al., 2018). In our modified architecture, we add a channel, using each channel for a different purpose. Thus, our architecture consists of three channels (X1, X2, X3). The input image is the mammogram, and it is directed to the three channels, each containing kernels of different size. The idea is that each channel extracts features of different sizes that help the network with its task. • Channel X1 (for larger size features) consists of a convolutional layer with 25 filters of size 9 × 9. • The second channel X2 (for medium size features) consists of a convolutional layer with 40 4 × 4 filters, followed by a 2 × 2 Maxpooling layer, then another 3 × 3 convolutional layer with 35 filters and finally a 2 × 2 size UpSampling layer. • In the third channel X3 (for small size features), begins with a convolutional layer of 35 filters with a size of 2 × 2, a MaxPooling layer of 2 × 2, a convolutional layer of 50 filters of 2 × 2, a MaxPooling layer equal to the previous one, another convolutional layer with 35 filters of 3 × 3 and a 4 × 4 UpSampling. The three channels are concatenated and then the output goes through three convolutional layers, the first two with a size of 7 × 7 with five and seven filters. The last convolutional layer consists of 1 filter with a size of 1 × 1. The network has as output, a mask over the tumor. Figure 4 shows the architecture and an example of network input and output. Note that several other state-of-the-art architectures (Karianakis et al., 2015; Noh et al., 2015) were tested without having a favorable result. For the training of the network, a batch of 330 images was used, and we train the network with 3500 epochs. Cross entropy was used as a loss function. The optimization algorithm used was Adam. The architecture was implemented using the Keras library (Chollet, 2018; Gulli et al., 2018; Cortez, 2017).
Segmentation of Mammogram masses for Smart Cities Health Systems 5
Figure 4: Architecture for the segmentation of masses in mammograms.
For segmentation evaluation, the Intersection over the Union Metric (IoU) was used. This is an evaluation metric commonly used to measure the accuracy of an object detector in a particular data set, calculating this metric is as simple as dividing the area of overlap between bounding boxes by the area of the joint (Palomino, 2010; Rahman et al., 2016). The metric is also frequently used to assess the performance of convolutional neural network segmentation (Rezatofighi et al., 2019; Rosebrock, 2019). The formula to calculate the IoU is IoU=(Area of Overlap)/(Area of union)
(1)
Another metric used is the true positive value (PPV) as (Hay, 1988; Styner, 2008). PPV =
TP (TP + FP)
(2)
Where: • TP are the true positive pixels, i.e., pixels that are part of the mass (object) and detected as mass; • FP are false positive pixels, i.e., pixels that are not part of the mass (background) and detected as mass. We also used the true positive rate (TPR), which is a true positive that represents a pixel that is correctly predicted to belong to the given class (González García, 2019). Its formula is: TPR =
TP TP = p (TP + FN)
(3)
Where: • FN are false negative pixels, i.e., pixels that are part of the mass (object) but are classified as background.
4. Results This section presents the results obtained from the network on the test set, which consists of 150 images. In Figures 5 and 6, two images of the test set and the output of the proposed network are presented. In Figure 7, we show three graphical examples of the evaluation of the network. The green color represents the true positives, the red color the false positives and the blue color the false negatives.
Figure 7: Each row is the input and output for the same mammogram. First column is the input mammogram, second column is the true mask, third column is the output mask from the network and finally the fourth column shows the TP, FP, and FN in green, red and blue, respectively.
Table 1 shows the IoU metric values obtained for the first eight images of the database processed by the network. Table 2 presents the precision or positive predictive value (PPV) obtained from first eight images of the database.
The average of the metrics for all tests was as follows: IoU as an average of 0.77, PPV averages 0.85, while TPR has an average of 0.88. In general, the proposed architecture shows promising results; the total of TP is much greater than the sum of FP plus FN. From Figure 7, it can be seen how most of the tumor mass is correctly detected and segmented giving high TPR values. This is enough for a specialized doctor to note a possible mass in the mammography. In addition, the calculation of the total mass could serve as a quantitative measure to evaluate the response of the tumor to the treatment. On the other hand, it is also necessary to improve the network to reduce as much as possible the amount of FN and FP.
8 Innovative Applications in Smart Cities
5. Conclusions and Future Work This chapter presented the design of architecture to create a network architecture that was able to segment breast tumors. In the main structure of the architecture, three-channel convolutional neural networks with six different layers and different filters were used. The network evaluation and validation were done with the IoU metric and values PPV and TPR and indicated that the network correctly segmented the tumors with an efficiency of 88%. As future work, it is planned to improve by the following aspects: • Search and test more databases to obtain more variability in images. • Improve network performance by making it more efficient by eliminating some layers. • Evaluate the network with medical doctors.
A new report on childhood obesity is published every so often. The bad habits of food and the increasingly sedentary life of children in a border society has caused an alarming increase in the cases of children who are overweight or obese. Formerly it seemed a problem of countries with unhealthy eating habits, such as the United States or Mexico in Latin-America, where junk food is part of the diet during childhood. However, obesity is a problem that we already have around the corner and that is not so difficult to fight in children. In the present research, the development of an application that reduces the problem of the lack of movement regarding the children of a smart city is considered a future problem. The main contribution of our research is the proposal of an improved type of Serious Game, coupled with the achievement of an innovative model to practice an Olympic sport without the complexity of moving physically in an outside space and having to invest in a space with high maintenance costs, considering the adverse weather conditions such as wind, rain and even a dust storm. We use Unity to model each Avatar associated with a set of specific sports, such as Water polo, Handball, Rhythmic Gymnastics and others.
1. Introduction The increase in childhood obesity, a problem of great importance in a smart city, determines the challenges that must be addressed with respect to applications that involve Artificial Intelligence. Computer games to combat childhood obesity are very important to reduce future problems in our society. Children increasingly play less on the street and spend more time with video games and computer games, so they lead a more sedentary life. This, together with bad eating habits, increases the cases of obese children every year. What can parents do to avoid their children being overweight? A bet that comes to us from the University of Western Australia, Liverpool John Mores University and the University of Swansea in the United Kingdom is "exergaming", an Anglicism that comes from joining the word "exerdizze" in Turkish (exercise in English) with "gaming" (game). These are games that run on consoles such as Xbox Kinect or Nintendo Wii in which you interact through physical activity in tests in which you have to run, bike, play bowling or jump fences. The researchers tested children who performed high and low-intensity exergaming and measured their energy expenditure. The conclusion reached was that the exergaming generated an energy expenditure compared to exercise of moderate or low intensity, depending on the difficulty Universidad Autónoma de Ciudad Juárez, Av. Hermanos Escobar, Omega, 32410 Cd Juárez, Chihuahua, México. Universiad Simón Bolívar, Sartenejas, 1080 Caracas, Distrito Capital, Venezuela. * Corresponding author: [email protected] 1 2
Serious Game for Caloric Burning in Morbidly Obese Children 11
of the game. In addition, the game was satisfactory for the children, who enjoyed the activities they did. It is a tool that parents can take advantage of to prevent children from spending so many hours sitting in front of the console as it has been shown to offer long-term health benefits. In any case, it must always be one of the means we can use to encourage children to do some physical activity but not the only one. Going out the street to play, run, jump, must always be on the children's agenda, as is shown in Figure 1.
Figure 1: Intelligent application using Kinect.
The Serious game represents a practical idea of how to solve problems associated with caloric intake because they allow performing ludic aspects of a game and the regulations associated with a specific sport, which is why the research conducted took into consideration a set of sports with high mobility associated with a control group that has morbid child obesity. The remainder of this chapter is structured as follows: In Section §2, the approach of a serious game for caloric burning is presented. Methodological aspects of the implementation of serious games are presented in Section §3, where psychological and technological factors are considered to guide their development. Section §4 introduces the method for estimating caloric burning in the implementation of a serious game. Technical aspects for modeling of avatars in a serious game for caloric burning are given in Section §5. Finally, the analysis of results and the conclusions are presented in Sections §6 and §7, respectively.
2. Serious Game for Caloric Burning After the advent (more than two decades ago) of video games that implement technologies that allow the active physical interaction of the user (active video games), there has been an increased interest regarding research into estimating the amount of energy consumed in the gaming sessions conducted by users and whether these video games promote the physical activity of players. Compared to traditional non-physically interactive video games, active video games significantly increase energy consumption to levels similar to those of moderate-intensity physical activity [1]. It has been found that a child's energy expenditure during the activity of a video game such as the Boxing and Dance Dance Revolution (level 2) of the Nintendo Wii console, is compared to the energy expenditure experienced on a treadmill at about 5.7 km/hr [2]. Studies reveal that the continuous practice of active video games generates a calorie-burning equivalent to a physical activity that is able to cover the recommendations, in terms of energy expenditure per week (1000 kcals), of the American College of Sports Medicine (ACSM) [1,3]. In 2011, Barnett et al., in a systematic review, found that the average of metabolic equivalent (MET) in young subjects during active video games was estimated at 3.2 (95% CI: 2.7, 3.7), whose value is considered to be moderate-intensity physical activity, although none of the papers reviewed in that study found that the MET reached the value of 6, which is considered to be the threshold for intense physical activity [4]. Later in 2013, Mills and co-
12 Innovative Applications in Smart Cities workers found that the Kinect Sports 200 m Hurdles video game generated increases in heart rate and energy expenditure patterns consistent with intense physical activity in children, which was also related to effects that can be considered beneficial on vascular function [5]. Gao et al. conducted a study to compare the effect of physical activity performed by children in physical education classes using active video games and those who performed physical activity in regular physical education classes. The authors concluded that the positive effects of a regular physical education class can be achieved with a physical education class using active play in children with light to vigorous physical activity behavior, with similar energy expenditures in both class modalities [6]. Studies show a positive impact of active video game use on body mass index in children and adolescents [7,8]. All this reveals the potentialities of the use of video games in the control of obesity and the prevention of illnesses associated with this condition. A remarkable aspect of active video games is the fun and entertaining nature of the video game itself, which makes it attractive to children and teenagers and represents a motivating component for physical activity. Research reveals that engaging in physical activity through an active video game is significantly more enjoyable than other traditional physical exercises, such as just walking [9] or using a treadmill [10]. On the other hand, a study showed that adolescents with obesity who were included in a physical activity program using active games reported an increase in physical activity and showed a high motivation intrinsic to the use of active video games [11]. The goal of the present research is the development of serious games based on active sports video games for increasing the burning of calories in morbidly obese children. In addition to the application of active games, the serious game incorporates metabolic equivalent analysis for the estimation of caloric burning based on the metabolic intensity indices described in the most recent compendium of physical activities for young people [12]. 2.1 Components of a serious game Let's go into detail, serious games are meant to teach. They are video games and applications with a didactic background. What does it mean? That users of serious games learn while having fun, which translates into a doubly positive experience. But what are the elements to achieve these objectives? Narrative: A story that engages the users with a storyline will encourage them to get involved in the project, so that they do not abandon it and get more involved. Having a good argument guarantees a greater immersion and motivation, which translates into better results. Interactivity: User participation in serious games allows communication between the tool and the user. Moreover, it is done with immediacy since the results can be seen as soon as the test is done. This provides valuable feedback, through which you can learn from mistakes and improve efficiency. Training: The main objective. This is the virtue of serious games, to create an experience in which the user has fun, where although he plays with his eyes, in the end he learns. With these key elements in mind, betting on serious games will provide added value in the current digital context and training in any organization that wants to build a genuine effort to help others, in this case, with the impact on promoting the exercise that helps improve the health of children in a Smart City.
3. Methodological Aspects There are many transcendental topics when considering to establish a specific serious game associated to an aspect of the technology that allows to take the caloric control of the exercise carried out as in:
Serious Game for Caloric Burning in Morbidly Obese Children 13
3.1 Emotions in children The science-based child psychology emerging in the second half of the nineteenth century promised to provide a rational basis for education for the overall development of the child. All this promising area in development offers new and interesting proposals such as the study of children in their environments associated with the improvement of empathy towards exercise, not only to be used by child psychology, that is why the creation of a science focused on the child that goes beyond traditional psychology [2]. This thanks to William Preyer, who is considered the father of child psychology, but the authors of the study did so with the same rigorous exactitude as that used in their observations of their work, characterizing it as being of a very systematic type [1]. 3.2 Emotional disorders in children Mental disorders of children affect many children and their families. Children of all ages, ethnic or racial backgrounds, and from all regions of the United States have mental disorders. According to the report by the National Research Council and Institute of Medicine in the division (Prevention of mental, emotional and behavioral disorders in young people: progress and possibilities, 2009) who gathered findings from previous studies, it is estimated that 13 to 20% of children living in the United States (up to 1 in 5) have a mental disorder in a given year, and about 247,000 million dollars a year are spent on childhood mental disorders [3]. In the literature, we find several definitions to refer to emotional, mental or behavioral problems. These days, we find that they are referred to as "emotional disorders" ("emotional disturbance"). The Education Act Individuals with Disabilities Education Act ("IDEA" for short) defines emotional disorders as "a condition exhibiting one or more of the following characteristics over a long period of time and to a marked degree that adversely affects a child's educational performance" [4]. The lack of ability to learn that is inexplicable by intellectual, sensory or health reasons. (a) A lack of ability to maintain personal relationships on good terms with their classmates or teachers. (b) Having inconsistent behaviors or feelings under normal circumstances, including anxiety attacks. (c) Having recurring episodes of sadness or depression. (d) Developing physical symptoms, including hypochondriacal symptoms or fears associated with personal or school problems. Hyperactivity: This type of behavior manifests as the child being inattentive, easily distracted and impulsive. Assaults: When the result of the behavior ends in injury, either to themselves or their neighbors. Withdrawal: The social life shows signs of delay or the individual shows an inability to relate to their environment. This includes excessive fears or anxiety. Immaturity: Unwarranted crying spells and inability to adapt to changes. Learning difficulties: Learning does not develop at the same pace as the average in their environment. It remains at a level below their peers. There are even more young children with serious emotional disturbances, i.e., distorted thought, severe anxiety, uncommon motor acts, and irritable behavior. These children are sometimes diagnosed with severe psychosis or schizophrenia [4]. 3.3 Play therapy Play therapy is defined as a therapeutic model of formal recognition of the child and has also proven its effectiveness in children with emotional stress problems that contribute to and manifest in each child during their normal development. Play therapy builds on child's play as a natural means of
14 Innovative Applications in Smart Cities self-expression, experimentation, and communication. While the child plays, they learn about the world and explore relationships, emotions and social roles. It also gives the child the possibility to externalize his personal history, thus releasing negative feelings and frustrations, mitigating the effects of painful experiences and giving relief from feelings of anxiety and stress [5]. The play therapist is specialized, trained in play and uses professional technical therapeutic methods adapted to the different stages of child development. The importance of a specialized Serious Game is to capture and understand the child's emotions, as well as to get involved in the child's game to create a true relationship for the expression and management of the child's internal conflicts, an aspect of great relevance and that favors Serious Games, as well as to download and understand their emotions in order to properly manage them and not get trapped in them, making them able to recognize and explore the problems that affect their lives [6]. A serious game does not replace the play therapy of your therapist, but seeks to provide an auxiliary and also effective support tool for those who are in contact with children who at some point show signs of moodiness. It could also be used by health professionals as part of their therapeutic tools to control the emotions of their patients and to be able to give a more adequate follow-up to their therapy, especially occupational therapy. 3.4 Serious games An old definition of the 70's that so far has remained is that serious games are those that have an educational or therapeutic purpose and are not just for fun. Within this definition is the real challenge for developers, i.e., maintaining this balance between fun and fulfilling the purpose of carefully planned learning. Once the construction of serious game is focused on learning or some specific therapeutic purpose, it often happens that the fun part of the game (so important to it) is neglected by not determining the quality and impact that serious play should have on users. The applications are varied, with the most common being education, management, policy, advocacy, and planning, among others [7]. Why games like support? The answer is simple and intuitive as these are part of the training of people who are going to use it. Nowadays most of the serious games are multiplayer, that is, collaborative, so it is very important to learn in a collaborative way. Most of these serious games are based o n cultural aspects that allow children to associate with their environment and with the people around them. The use of technology has improved the effects associated with the visual aspect of serious games [8].
Figure 2: Design of a speed skating rink in an open and arboreal space to improve the ludic aspect of the practice of this sport.
3.5 The importance of properly building the components of a Serious Game An important aspect of Serious Games is determining the child's correct progress in it and how he adapts to changes in the stages of serious play [9]. Another aspect to consider is the environment associated with Serious Game, which should be as realistic as possible and according to the scenario where the child's learning skills are to be developed. Another great challenge is to properly organize the set of rules to follow to advance in the development of the game and achieve an intuitive understanding of it [8].
Serious Game for Caloric Burning in Morbidly Obese Children 15
3.6 A serious game associated with the appropriate heuristics for continuous improvement Artificial intelligence, using adequate heuristics, will allow to demonstrate the correct functioning of the avatar in the built environment for learning, achieving an adequate link with the avatar associated with the child, as can be seen in Figure 2.
4. Estimation of Caloric Burning during the Practice of Serious Game When a person is at rest, his or her body consumes energy for the maintenance of vital functions; this energy consumed is called the resting metabolic rate (RMR). Physical activity increases energy consumption above resting levels, and the number of calories expended is related to the intensity and duration of the activity. The metabolic equivalent (MET) is an index used to measure or express the intensity of physical activity, so it can be used to estimate the amount of energy consumed during some type of exercise. A young adult of 70 kg at rest (seated and still) normally consumes about 250 mL/min of oxygen; this is equivalent to ≈3.5 mLO2/kg·min, which represents 1 MET (standardized value), and similarly corresponds to a consumption of 1 kcal/kg·h [13]. When carrying out a physical activity, oxygen consumption increases, so this activity has a MET > 1; for example, if an activity has 4 METs, it means that 14 mLO2/kg·min are required to carry out the activity, that is to say, 4 times the consumption of energy that is presented at rest. The Physical Activity Guide for Americans classifies physical activities as light intensity with METs < 3, moderate with METs between 3 and 5.9, and intense with METs > 6 values. A walk at 2 mph represents a light physical activity of 2.5 METs; if the walk is at 3 mph it has a MET = 3, which classifies the activity as moderate intensity; and running at a speed of 10 mph corresponds to an intense physical activity with a MET = 6. Also, the guide qualifies a sedentary behavior as behavior or activity that presents low levels of energy expenditure, equivalent to MET < 1.5. It is suggested that children and adolescents aged 6 to 17 years should do 60 minutes (1 hour) or more of moderate to vigorous daily physical activity, such as Aerobic, Muscle Strengthening or Bone Strengthening [14]. Since the standardized value of 1 MET is derived from oxygen consumption for a subject with particular characteristics (healthy adult male, 40 years old, 70 kg) in resting conditions, their relationship between calorie consumption 1 MET = 1 kcal/kg·h is subject to variations depending on age, sex and body composition, which is primarily due to the fact that the energy consumed by a resting person (RMR) depends on such factors, in addition to the health condition, stress level, and others. In this sense, it should be mentioned that the RMR has higher values in men than in women, increases with height, weight and body composition of a person, and decreases with age. Research has found that the use of the standardized MET value can cause misclassification of the intensity of physical activity [15], and, in turn, inappropriately estimate the true values of oxygen consumption and energy costs during the activity [16]. This is why the compendium of physical activities (update 2011) proposes a correction factor for the calculation of the metabolic equivalent (MET corrected) based on a better estimate of the RMR, which uses the Harris-Benedict equation which takes into account the age, height, weight, and sex of the subject [17]. In [18] a correction factor that allows a more accurate MET calculation in overweight adults is proposed. On the other hand, because children have a higher resting basal metabolism (BMR) per unit body mass than adults, adult MET values do not apply to children. In addition, as the child grows, the BMR decreases gradually. The factors that cause this decrease in BMR in children are mainly changes that occur in the mass of the organs and in the specific metabolism of some organs, and changes in muscle mass and fat mass, which in turn are linked to the sex of the child. A child's BMR may be underestimated by using the standard adult MET, since the BMR of a 6-year-old child is on average ~ 6.5 mLO2/kg·min (1.9 kcal/kg·h) and approximately 3.5 mLO2/kg·min for 18-year-olds [12]. Because of BMR behavior in children, calorie intake from physical activity is not constant
16 Innovative Applications in Smart Cities during childhood. At the same time, for the same physical activity, a child has a higher energy expenditure per body mass than an adult or adolescent. Thus, the most recent compendium of physical activities for youth establishes a metric for youth MET (METy) that is significantly different from that of adults, and that is age-dependent [12]. The compendium presents the METy values of 196 physical activities commonly performed by children and young people, for the following discrete age groups: 6–9, 10–12, 13–15 and 16–18 years. For the calculation of the BMR, the Schofield equations according to age groups and sex are used: Age
For the present work, it is proposed the use of the METy values of the young compendium for the different physical activities that are implemented in the designed serious games, which belong to the category of active full body video games [12]. Table 1 shows the different METy values for the age groups 6–9, 10–12 and 13–15, for which the use of serious games is intended. Table 1: METy values of active video games (full body) for the physical activities of the serious games [12]. Code
Specific Activity
METy by age-group (years)
15120X
Baseball
6–9 3.7
10–12 4.7
13–15 5.7
15140X
Boxing
3.0
4.0
4.9
15180X
Dance
2.3
3.3
4.1
15260X
Olympic games
2.6
3.6
4.5
15320X
Walking on treadmill and bowling
2.8
3.9
4.8
15400X
Wii hockey
1.4
2.4
3.2
15480X
Wii tennis
1.6
2.5
3.2
Then, knowing the BMR and METy for age group, and duration of a physical activity, the energy expenditure is calculated by: EE = METy × BMR (kcal/min) × duration (min)
(5)
For example, if a 10-year-old girl (37 kg), with a BMI greater than the first three quartiles according to the population of her age, plays the serious game of Rhythmic Gymnastics (like Dance in table 1, METy = 3.3) for 15 minutes twice a day, her daily caloric burning due to the practice of that physical activity can be determined as follows: • Using Schofield equation 3, BMR = [20.315 × 37 + 485.9]/1440 = 0.86 kcal/min. • Total energy expenditure (EE) for this physical activity: EE = 3.3 × 0.86 kcal/min × 30 min = 85 kcal 4.1 Using a Serious Game to determine efforts in a Virtual Sport As mentioned in this document, this project is currently in the construction phase. This prototype of a Serious Game has the firm intention of taking the next phase of complete and functional construction. A foundation sponsored by one of the most important children's hospitals in the
Serious Game for Caloric Burning in Morbidly Obese Children 17
city of Paso Texas in the United States has shown a genuine interest in the project, for which it mentioned providing the necessary support for its realization. In Mexico, the National System for Integral Family Development has childcare programs with interesting and very efficient strategies for reducing morbid obesity in overweight children. That is why it is intended to integrate this intelligent tool in their support programs for these children. For this, it will soon be formalized by both parties committed to supporting the project shown in this document. In future research, we try to modify a game based on collaborative work in a group—we are choosing rugby seven—with high intensity of pressure for each child and modify the importance related to the support of this type of pressure related to the responsibility of a Collective activity, an approximation will be related to what is implied for Water polo, as shown in Figure 3. A very relevant aspect is to consider that if someone asks why he likes to use our Serious Game, this user will be able to respond: because he has had a playful scope and of adequate selection with the avatar, so he could have empathy for our proposal. By analyzing in more detail the group of people who used our Serious Games, we determined that, like the role-play, it is a hobby that unites them and gives them opportunities to help each other and their videogame community. It is a safe environment in which you can experience social interactions, something fundamental when the climate does not allow it, in Bw-type climates (according to the Köppen climate classification scale) as the place of our study. This group of users of our Serious Game says that they have witnessed the personal growth of individuals in terms of their self-esteem and the expansion of their social interactions as a result of the game. This is just one of the benefits of the game. Our research showed that it was discovered that everyone can find some hours a week to "save the universe, catch the villains or solve mysteries" while learning to practice Water polo, and that playing with the computer is as fun as any other activity in our research Playing our Serious Game can strengthen a variety of skills such as math and reading online recommendations. Increase the ability to think and speak clearly and concisely, when formulating and implementing their plans, cooperating and communicating with others, as well as increasing the ability to analyze written and verbal information. Placed on the market, our Serious Game will determine that players are cohesive members of the group in multiplayer games, and it can help people develop leadership skills and promote cooperation, teamwork, friendship, and open communication. In another study related to this kind of Kinetic Serious Game, we try to compare with our colleagues of Montenegro who propose and develop an innovative Serious Game which involves a model to Fencing practitioners, as this sport is reaching high popularity in this society. A representative model can be shown in the next Figure 4. What would users expect from our proposal of a Serious Game in this Kinetic model to learn and practice Water polo? Improve the mood of serious game users through the components that will be used as background music, with the purpose of using music therapy techniques to lift the spirits of the players as it develops [11]. Another element of the strategy and ploy in Serious Game is the colors of the scenery. By taking into account the effect of color on mood, a better color experience can be predicted to keep the player in a positive emotional state [10]. The third element is through the sounds of the game at each stage of development. With every success and failure, the sounds in your environment can represent the stage you are playing on. And the fourth element is the recognition of achievements. Through badges, medals, trophies, scores, and appointments you want the player to have a feeling of satisfaction with the recognition of each achievement [12].
Figure 3: Our Kinetic Serious Game Model using collaborative task to play Water polo.
18 Innovative Applications in Smart Cities
Figure 4: Use of Kinetic software to improve performance in Modern Pentathlon.
5. Modeling of Avatars in a Serious Game for Caloric Burning In addition to the BDI methodology, the physiological traits of the agents intervene to improve each aspect of an avatar. The scenes structured associated with the agents cannot be reproduced in general, since they only represent a small part of the population in space and time of the different societies. These individual behaviors represent a unique and innovative form of global adaptive behavior that solves a computing problem that does not attempt to group societies only with a factor associated with their external appearance (phenotype) and therefore sports that could be practiced much more, but it tries to solve a computer problem that implies a complex change from the perspective of sport that has opted for a better empathy in relation to its practice, in order to improve competitiveness in children's health among the existing relationships with the population that practices these sports. The generated configurations can be metaphorically related to the knowledge of the behavior of the community with respect to an optimization problem (to conform to cluster social and culturally with other similar people, without being of the same sport [4]). In Table 2 is shown a sample of seven sports, describing each of the analyzed characteristics in order to determine which were the most viable to develop in a Serious Game. K = [C + CS+ GAL] ± CBS
(6)
where: C = represents if this sport can be practiced in any climate. For example, in the case of Chess, it is not an indispensable condition. CS = represents the Symbolic Capital associated to the perspective of a society in a Smart City. The practice of Fencing is considered sophisticated and therefore has more prestige. G = is defined as the gender in the population range. In Juarez City the population pyramid is different from the rest of Latin America because violence causes deaths and exodus from the city; that is why the values are different and their representativeness as well. AL = Ludic aspect related to the age of the children who practice sports. For example, the trampoline is associated with a fun experience, and its playful aspect is associated with high promotion. CBS = Social Benefits-Cost related to each sport. In the case of Rhythmic Gymnastics, it is associated with improving the health of sedentary children and helps them lose weight quickly. We use equation 4, where K represents the performance of the practice for each of the sports in the city and their respective promotion in various collaborative spaces. 5.1 Unity implementation To achieve everything mentioned previously, from the implementation of algorithms, we use Unity to obtain the greatest potential of the serious game. The most relevant aspect of this research is to consolidate the practice of sports in children and young people who, due to weather conditions in the practice of normal sports which is not the case of Serious Games, cannot agree to join together to
Serious Game for Caloric Burning in Morbidly Obese Children 19
practice a sport together with the self-confidence generated in the players, which allows improving their performance in other areas of their daily life through a model of emotional support in children, which entails a commitment and intrinsic complexity in their motor development. Considering that childhood is a vulnerable stage where they are also in full development and any event or occurrence may be able to cause negative effects and may leave the child permanently marked by the rest of their life [13, 15, 16, 17, 18, 19, 20], it is very important to focus more on how to obtain results associated with group performance sponsored by the individual. That is why the future of the Serious Game will require a deeper investigation that allows being of great impact by having an opportunity to help children and youth who do not have access to sports for various reasons. In the future, the Serious Games may present a natural opportunity due in large part to the acceptance that videogames have in this age and even more so with the advantage that these generations (both Gen Z and now the
Figure 5: Representation of different virtual sports.
children of the so-called Alpha α generation) have easy access to technology. The implementation of Unity in diverse virtual sports is presented in a Collage of them, as is shown in Figure 5.
6. Analysis of results And Discussions of Our Research As mentioned in the present investigation, this prototype wishes to continue consolidating in order to establish diverse teams associated mainly with the avatar. This prototype regarding the Serious Game to achieve the complete and functional construction phase [21, 22, 23, 24, 25, 26, 27]. At this time and due to a project with funding from the European Union and the collaboration of FINA, we want to make a Kinetic application that allows being inclusive through interesting and very efficient strategies associated with mobility. FINA is very interested in this type of application that could diversify the practice of Water polo in developing countries. Table 2 presents the data for the multivariable analysis with the information on the number of spaces to recover for the practice of these virtual sports and their representations by gender (aspects to be covered in each sport)—specific values for both Men and Women, the social status for its practice, the fun for use of the Wii application that is being proposed, the external climate to complement with the Serious Game, the improvement of health to change the paradigms to sedentary children, and finally the relationship between the social cost/benefit associated with its practice.
20 Innovative Applications in Smart Cities Table 2: Multivariable analysis to determine the relationship between the timely improvement of some sports using our Kinect proposal and an avatar associated with its performance coupled with an intelligent system for the control of heat burn in morbidly obese children. Sport Aquatic Sky
Virtual space
Gender
Social status
Fun
Climate
Increase Health
CostBenefit
8
m-.50, f-.45
0.498
0.914
0.774
0.715
0.387
Judo
6
m- 40, f- 30
0.857
0.714
0.851
0.879
0.568
Baseball
5
m-50, f-.40
0.617
0.658
0.385
0.712
0.514
Syncronized Swimming
3
f-.47
0.275
0.637
0.416
0.748
0.885
Water polo
14
m-.45, f-.40
0.578
0.784
0.925
0.627
0.879
Bowling
3
m-.30, f-.30
0.620
0.631
0.715
0.802
0.744
BMX Bike
5
m-.40, f-.40
0.562
0.748
0.611
0.303
0. 448
Rolling Sport
4
m-.48, f-.42
0.877
0.917
0.459
0.897
0. 574
f-.49
0.718
0.897
0.427
0.928
0. 927
Rhythmic Gymnastics
7. Conclusions and Future Challenges The main experiment consisted of detailing each one of the 47 sports, with 500 agents, and one condition of unemployment of 50 generations, this allowed us to generate different scenarios related with Time Horizons, which was obtained after comparing different cultural and social similarities in each community and to determine the existing relations between each one in relation with the Mahalanobis Distance (the number of dots indicated each sport and the size of people represents the number of people which determine the magnitude related with the society). In future research, we will try to improve the practice of a sport associated with a game and based on collaborative work in a group with high intensity of pressure for each child as professional tennis and modify the importance related to the support of this type of pressure related to the responsibility of a collective activity [14], as shown in Figure 6.
Figure 6: A serious game based on collective activities and related with the increase of social skills.
Intelligent1 Fernando Maldonado-Azpeitia1,* and Alberto Ochoa-Zezzatti2
1. Introduction Around the world, we can find data about food waste and some of its most important causes. To mention some data, every year in the Madrid region, 30% of products destined for human consumption are lost or wasted by improper handling in the food supply chain (CSA) comments Gustavsson et al. (2011). A study in the United States by the Natural Resources Defense Council (NRDC) found that up to 40% of food is lost from the producer's farm to the consumer's table, Gunders (2012). Losses of perishable products vary among countries around the world, in some countries, such as in China, they even increase. Reports indicate that only 15% of fresh products are transported under optimum temperature conditions, despite the knowledge that these types of products require refrigerated handling (Pang et al., 2011). They also comment that fruits and vegetables are the most affected type of food, where 50% of what is harvested is not consumed and this is mostly due to insufficient temperature control. Approximately one-third of the world's fruits and vegetables are discarded because their quality has fallen and because of this it lacks acceptance and puts food safety at risk. 1.1 Situation of perishable foods In Mexico, 20.4 million tons of food is wasted annually, these data only correspond, according to the World Bank (2018), to 79 foods representative of Mexico's food basket, which implies large environmental impacts due to the excessive use of water and carbon dioxide generation. This represents the waste of about 34% of the national food production, which if considered the rest of the food could reach about 50% of the total national production produced. Additionally, it was Universidad Autónoma de Querétaro, Mexico. Doctorado en Tecnologia, UACJ; Mexico. * Corresponding author: [email protected] 1 2
Intelligent Selection of Best Fresh Products 23
observed in this study that approximately 72% of losses occur between pre-harvest and distribution. That is, in the early stages of the production chain and by taking them to their target market, in retail, which could be the result of bad consumption habits. Much of these losses are related to inadequate management of temperature control during CSA processes (production, storage, distribution and transport, and at home) (Jedermann et al., 2014). Similar studies have shown that, in many cases, food security is frequently affected by poor temperature management (Zubeldia et al., 2016). Environmental conditions, mainly temperature, have a great impact on the overall quality and shelf life of perishable foods, according to Do Nascimento Nunes et al. (2014). These are just some statistics that reveal a scenario where CSAs have deficiencies, in addition to providing sufficient support to strengthen the importance of control and monitoring of the cold chain, not only to solve the problem of food spoilage but also to address general challenges associated with world food security. Good temperature management is the most important and easiest way to delay the deterioration and waste of these foods (Do Nascimento et al., 2014). Franco et al. (2017) comment that there is no doubt that our way of life depends on our ability to cool and control the temperature of storage spaces and means of food distribution. 1.2 Cooling technology Today there are several alternatives in refrigeration systems that can be implemented either in commerce or industry. Gauger et al. (1995) classified the different refrigeration systems by an average range according to 6 criteria: state-of-the-art, complexity, size and weight, maintenance, useful life, and efficiency. Table 1: Definition of numerical ratings for the evaluation criteria. Evaluation Criteria
The 6 criteria to evaluate the refrigeration systems, each one of them was qualified for the different turns where it is necessary, such as commercial air conditioning, domestic and mobile, commercial and domestic refrigeration. For commercial refrigeration, which is the focus of this work, Gauger et al. (1995) evaluated refrigeration technologies from best to worst according to the criteria mentioned as shown in the following Table. The refrigeration technologies with the best qualification and, therefore, the most suitable for application in the commercial sector are steam compression and absorption. Steam compression technology is currently the most widely used refrigeration system for food preservation and air conditioning, both for domestic, commercial and mobile use. This system uses gases, such as chlorofluorocarbon (CFC) and hydrochlorofluorocarbon (HCFC), as cooling agents. These types of gases have excellent thermodynamic properties for cooling cycles, as well as being economical and stable (Gauger et al., 1995). The favourable environment for the storage of fruits and vegetables is low temperature and high humidity. This is reasonably achievable by steam compression cooling with relatively low investment and lower energy consumption. Dilip (2007) reports that this type of refrigeration
24 Innovative Applications in Smart Cities Table 2: Classification of technologies in commercial refrigeration from best to worst. Classification
system achieves a favorable environment for the storage of fruits and vegetables since the shelf life of perishable foods stored under these circumstances increases from 2 to 14 days compared to storage at room temperature, so for CSA this technology is very favorable. However, the gases used by this technology, such as the CFCs and HCFCs used as refrigerants for many years, have depleted the ozone layer, while fluorocarbons (FC) and hydrofluorocarbons (HFCs) have a high global warming potential (GWP) and cause global warming phenomena. For this reason, the use of alternative technologies, such as absorption, Lychnos and Tamainot-Telto (2018), has been targeted. The absorption cooling system is attractive for commercial refrigeration and air conditioning. If levels of complexity and maintenance can be reduced, it could also be attractive for domestic applications (Gauger et al., 1995). For Wang et al. (2013) this type of cooling is considered as a green technology that can provide cooling for heating, ventilation and air conditioning, especially when silica gel is adopted due to its great suitability in effective contributions to reduce greenhouse gas emissions. Absorption systems use natural refrigerants, such as water, ammonia and/or alcohols, that do not damage the ozone layer and have little or no impact on global warming (Lychnos et al., 2018). However, certain drawbacks have become obstacles to their actual applications and commercialization. For example, the discontinuous operation of the cycle, the large volume and relative weight of traditional refrigeration systems, the low specific cooling capacity, the low coefficient of performance, the long absorption/desorption time, and the low heat transfer efficiency of the adsorbent bed, explain Wang et al. (2018). On the other hand, Bhattad et al. (2018) reflect that one of the greatest challenges in today's world is energy security. There is a great need for energy in refrigeration and air conditioning applications. Although, due to limited energy resources, research is being conducted in the area of improving the efficiency and performance of thermal systems. According to Zhang et al. (2017), today the economic growth and technological development of each country depends on energy. Heating, ventilation, air conditioning and domestic and commercial refrigeration consume a large amount of energy. The refrigeration system has great potential for energy savings. Lychnos et al. (2018), present in their work the development of a prototype with hybrid refrigeration systems that combines steam compression and absorption technologies. Preliminary tests showed that it can produce a maximum of 6 kW of cooling power with both systems running in parallel. It is designed as a water cooler with an evaporating temperature of 5ºC and a condensing temperature of 40ºC. For countries such as Mexico, promoting energy savings would become a competitive advantage, even more so in the commercial sector for the micro-enterprise which, as mentioned above, often have limited electricity supply at their points of sale and marketing of their perishable products.
Intelligent Selection of Best Fresh Products 25
2. Methodology The design methodology to be used to carry out the project will be the "Double Diamond". The Design Council, an organization that advises the English government on the fundamental role of design as a creator of value, argues that designers from all disciplines share perspectives at various points during the creative process, which they illustrate as "The Double Diamond". This design process is based on a visual map that is divided into 4 stages: Discover, Define, Develop and Deliver. The purpose of choosing this methodology is to discover the best solutions by testing and validating several times since the creative process is iterative and with this, the weakest ideas are discarded.
Figure 1: Double Diamond Methodology. Source Desing Council UK.
2.1 Discover The first stage of this model describes how to empathize with users to deeply understand the problem or problems they are seeking to solve. For this purpose, field visits were made with different microenterprises within the food sector in the municipality of San Juan del Río, Querétaro, to which an interview was conducted to obtain data on the management and marketing of their products. In order to obtain relevant data for the application of this interview and to calculate the size of the sample with an 80% confidence level, a visit was made to the Secretary of Economic Development of San Juan del Río to investigate the number of micro-enterprises operating in the municipality; it was found that there are no data on the number of micro-enterprises at either the municipal or state level because they are such small enterprises that the vast majority are not registered with the Ministry of Finance and Public Credit, making it very difficult to have reliable data on microenterprises. Due to this, the interview was applied to 10 micro-enterprises in the municipality of San Juan del Río, Querétaro. The questionnaire is presented below. 1. 2. 3. 4. 5. 6.
Company and business. Place of operation of the company. What kind of products do you sell? Do you know at what temperature range they should be kept? Do you produce or market the products you handle? If you sell, do you receive the products at storage temperature?
7. What type of packaging do perishable products have? 8. How long does it take to move perishable products from where they are manufactured to where they are exhibited or delivered?
26 Innovative Applications in Smart Cities
Figure 2: Conceptual diagram of the implementation of an intelligent system that determinates the greatest freshness in the presentation threshold and color analysis of various sándwich issues in a stock of a store selling healthy products. Source own preparation.
9. 10. 11. 12. 13. 14. 15. 16.
How much product do you store? How much product do you handle during distribution? What type of vehicle do you use to distribute your products? Do you have the option of home delivery? What tool do you use to preserve the product during shipment? Is there any product that is wasted during distribution? What factors cause this decline? What do you do with the product that is not sold during your working day? What is the amount you plan to invest in a specialized tool to help keep the product better during distribution?
With this interview, we sought to know the situation of micro-enterprises in relation to the distribution, handling and storage of perishable products, as well as the level of loss of them. 2.2 Define For this stage, the objective was to carry out an analysis of the perception of the quality of perishable products among consumers in the municipality of San Juan del Río, in the state of Querétaro, in order to obtain data that will allow us to know how relevant the freshness, good presentation and first instance perception of the quality of food products are for consumers, and whether this impacts on the purchasing decision. As a measuring instrument, a questionnaire was designed to assess whether the quality, freshness and presentation of food is relevant to people when purchasing raw
Rate from 1 to 5, 1 being irrelevant and 5 quite relevant, the aspects you consider when buying a food for the first time. • Correct handling of the product during the distribution chain. • Information about the product and how it was produced. • Affordable price. • The quality of the product can be observed (colour, texture, etc.) • The packaging of the product is not in poor condition (knocked, dented, broken, torn, etc.)
28 Innovative Applications in Smart Cities and prepared foods. In the same way, the aim is to find out what consumers trust so much in the different types of businesses that sell food as raw material for preparing dishes, which we call "basic basket" in the questionnaire and the other type we call prepared foods, which are already prepared dishes that sell the types of business. This questionnaire was applied online through the platform of "google surveys" for consumers in the municipality of San Juan del Rio at random. In order to determine the sample, INEGI data were taken, which indicates that the number of inhabitants between 18 and 60 years of age in 2015 is 268,408, which we can consider as the total number of potential consumers in our population. The calculation was made with a confidence level of 90% and a margin of error of 10% so the result was 68 to obtain a reliable sample of the total population. The first part of the questionnaire consists of 4 questions to know the demographic data of the participants which help us to categorize them according to their age range, gender, maximum degree of studies and professional occupation. The second part of the questionnaire is made up of 6 questions that focus on obtaining data that provided us with a panorama to better understand whether there is consumer concern regarding the quality of the perishable products they buy, the level of confidence in the businesses that commercialize these products, more relevant aspects for the purchase decision, level of preference in the different businesses that commercialize both raw and prepared foods. For each of the reagents that make up the previous survey, answers with Likert scales were established to make the way of answering the respondents more dynamic. 2.3 Develop For this area of the second diamond, a hybrid cooling system with steam compression and absorption systems was tested to validate its operation. These tests were carried out in a laboratory of the company Imbera, located in the municipality of San Juan del Río in the state of Querétaro, with a controlled environment with a maximum temperature of 34°C, and a relative humidity of 59%. A prototype was developed for the tests. An evaluation protocol was developed for the tests to obtain the following information: -
Pull down time (time it takes the system to generate ice blocks for temperature conservation). Temperatures reached by the cooling system during pull down. Duration of the ice blocks without power supply. Temperatures reached by the cooling system without electrical energy.
The objective of this protocol was to delimit the categories of perishable products that can be optimally conserved by a hybrid refrigeration system. After this, a preliminary cost analysis of materials and manufacturing processes was carried out in order to know which are the best adapted to the investment capacities of the users. In this exercise, two specific manufacturing processes were analyzed: roto-moulding and thermoforming, since the plastic material is the best option for the manufacture of the conservation tool due to the variety that exists, and therefore the versatility of qualities that it offers. The following table describes prices of materials and tools. Table 4: Comparison of manufacturing processes. Roto Moulding
Thermoformed
Cost Moulds
$415,000
$44,000
Cost Parts
$8,531.27
$2,118.04
Cost of Refrigeration System
$3,404.34
$3,404.34
Cost of Electrical System
$6,981.83
$6,981.38
Total
$18,917.44
$12,504.21
Source: Prepared by the authors.
Intelligent Selection of Best Fresh Products 29
Figure 3: Conservation tool concept. Source Own preparation.
The most appropriate manufacturing process is thermoforming because the estimated price is within the range that users are willing to spend for the conservation tool. With this information, the design of the conceptual proposal of the conservation tool was developed with the considerations described above. The concept of the conservation tool for small businesses is made of high-density polyethylene. This plastic is commonly used in packaging, safety equipment, and construction, to offer lightness to maneuver and resistance to withstand vibrations and shocks during distribution routes. The dimensions of the equipment are 100 cm wide, 40 cm deep and 40 cm high, with a capacity of 74 liters. These dimensions make it easy to transport, i.e., it can be placed in any compact vehicle or cargo vehicles. As for the hybrid refrigeration system, it will be composed of a steam compression system and an absorption system, the cooling agent of the absorption system will be water which will be contained in two plastic tanks. Inside the plastic tanks passes a copper tube that is part of the refrigeration system by steam compression to freeze the water. In order for the steam compression
Figure 4: Refrigeration system concept. Source Prepared by the authors.
30 Innovative Applications in Smart Cities system to work, it must be connected to the electrical energy during the night and the correct formation of the ice blocks is guaranteed so that the absorption refrigeration system works correctly during the distribution routes. The internal walls of the equipment, as well as those of the water containers, have slots through which metal separators can slide to generate different spaces for different product presentations. To close this stage, a prototype was made that complies with the functionality of the concept in order to be able to evaluate it in the next stage of the methodology.
Figure 5: Product Organization. Source Prepared by the authors.
2.4 Delivery For this last stage of the double diamond model, the perishable food logistics strategy was validated with the prototype of the conservation tool with the products developed in the Amazcala campus of the Autonomous University of Querétaro and marketed in the tianguis that is established at the central campus of the same university. The products that were placed to be evaluated within the prototype of the conservation tool are containers with 125 ml of milk and a milk-based dessert (custard), which are produced on that campus. As can be seen in the Figure above, the samples are identified so that one is placed inside the prototype of the conservation tool and the other outside it. For these milk samples, the following initial data were taken before placing them in the prototype of the conservation tool. In the case of the milk-based dessert, the measurement will be visual since this product will present syneresis (expulsion of a liquid in a mixture), losing its texture if the cold chain is not maintained correctly during distribution. The validation protocol for the conservation tool prototype consists of the following steps: -
The conservation tool prototype is mounted on the truck at 8:30 am. The products to be marketed are mounted on the truck for a period of 30 minutes. Products leave the Amazcala campus for the point of sale at the downtown campus at 9:00 am. The van arrives around 10:15 am at the Centro campus, at the engineering faculty, unloads the products at the point of sale. After this, it is withdrawn to make other deliveries to different points of sale. - At 1:00 pm the van returns to the engineering faculty point of sale to pick up the unsold product. - At 3:15 pm the van arrives at the Amazcala campus, until then the samples are taken and analyzed. Before this validation protocol, the equipment was left connected the night before from 1:30 a.m. to 8:30 a.m., a period of 7 hours to ensure that the ice blocks were formed correctly.
The objective was to make comparisons between the use of the conservation tool prototype and the dry box of the truck with which the products are transported from the Amazcala campus to obtain data from the specialized tool and fine-tune the strategy and be able to launch it to the market. 2.5 Ethical considerations For the last stage of the methodology that is delivery where it is planned to make tests with the tool specialized in micro-enterprises, the perishable foods that will be used during these activities will not be consumed by people and will be available in an appropriate manner. On the other hand, the information obtained during the discovery and definition stages will be data related only to the working activity of micro-enterprises and not personal data or confidential information about microentrepreneurs.
3. Project Development The micro-enterprises that were considered for research and application of the project are those that are categorized as micro-enterprises, which are formed by no more than 15 workers including the owner according to INEGI, within the trade sector in the branch of food and/or perishable products. These companies will be approached with an interview questionnaire to learn about and analyze the tools and procedures they use during their supply chain. With this information, it is sought that the candidate companies to the project have the following or at least one of the following characteristics: - They do not have a clear and established logistics for the distribution of their perishable products. - The tools with which they distribute and maintain the conservation of perishable products are not adequate and mistreat the presentation of their products. - Distribution logistics are complex, either because they don't know the ideal tools for conservation and optimal temperature ranges for their products, or because they don't have the economic justification to invest in specialized refrigeration transports.
32 Innovative Applications in Smart Cities Having explained the process of selecting companies for the development of the project, below is a list of the activities to be carried out in order to design a logistics strategy for perishable products for the microenterprise: 1. A problem is discovered through the observation of the actors. 2. Review of literature on issues associated with the problem. 3. Application of interviews to population of micro-entrepreneurs on the application of CSA in their business. (The sample population are businesses with turn in the commercialization of prepared or perishable foods in the state of Querétaro, specifically in the municipality of San Juan del Río). 4. Case studies and similar projects. 5. Analysis of the information collected. 6. Exploration of users' needs and aspirations. 7. Definition of the design problem. 8. Definition of project specifications. 9. Stage of development of potential solutions. 10. Weighting of solutions to find the most feasible. 11. Conceptualization of the possible solution. 12. Design of support material for the communication of benefits of the strategy. 13. Prototyping of the product. 14. Execution of the strategy based on the prototype. 15. Validation of the strategy according to KPI. - Quantity of perishable products that arrive in optimal conditions of presentation and conservation. - Time of handling and organization of the perishable products so as not to break the CF in the CSA. - Increase in sales due to adequate management of the conservation and presentation of perishable products. 16. Analysis of results. 17. Conclusions. 3.1 Human resources - Designer: Development, application and evaluation of the strategy based on the established methodology. - Industrial Designer: Product design based on the requirements and wishes of the user. - Refrigeration Engineer: Design and development of a refrigeration system that adapts to the requirements of the product and the project. 3.2 Material resources -
Laptop. Internet. Work space (table, desks). Automobile for field visits. Automobiles and user vans for tests and validations.
Intelligent Selection of Best Fresh Products 33
- Smartphone for communication with users, as well as for taking videos and photographs as a record. - Various materials for prototype.
4. Results and Discussion The main objective of this stage was to empathize with micro-entrepreneurs and to analyze the problems they encounter when distributing, handling and storing the perishable products they market. The questionnaire consists of 16 questions with the purpose of shedding light on the deficiencies in the operation of the company and to know how they are solved. It was found that the products marketed by these companies range from pastries and confectionery (dairy products) to flowers, sausages, fruits and vegetables. The current logistics process in the distribution channels of the micro-enterprise, in general terms. The strategy for logistics in the distribution of perishable foods based on a specialized conservation tool, in addition to guaranteeing the quality and presentation of the products, helps micro-entrepreneurs to find added value in their economic activities. Like the applications for the distribution of prepared foods (uber eats, rappi, etc.) that through their services generate an increase in their profits in businesses through home delivery. This distribution strategy for perishable foods transfers this same service model, through the specialized conservation tool to successfully promote the incursion of new sales channels for micro entrepreneurs and increase their income.
Analysis of Mental Workload on Bus Drivers in the Metropolitan Area of Querétaro and its Comparison with three other Societies to Improve the Life in a Smart City Aarón Zárate,1 Alberto Ochoa-Zezzatti,2 Fernando Maldonado1,* and Juan Hernández2
1. Introduction Mental workload is investigated in ergonomics and human factors and represents a topic of increasing importance. In working environments, high-cognitive demands are imposed on operators, while physical demands have decreased (Campoya Morales, 2019). These figures make it possible to measure the serious public health problem that causes road accidents in the world and in our country and the strong negative impact that it generates in the society and the economy. Hence, in 2011, the WHO generated a program that is called the Decade of Action for Security Vial 2011–2020, through which it summoned several countries to generate actions with the purpose of mitigating this problem. In our country, in 2011, the National Road Safety Strategy 2011–2020 and 2013 was promoted within the National Development Plan, the Road Safety Specific Action Program 2013–2018 (PAESV). The goal of reducing the mortality rate caused by road accidents to 50% was proposed, as well as minimizing injuries and disabilities through 6 strategies and 16 lines of action concentrated in 5 main objectives: 1. To generate data and scientific evidence for the prevention of injuries caused by road accidents 2. To propose a legal framework on road safety that includes the main risk factors present in road accidents
3. To contribute to the adoption of safe behaviors of road users to reduce health damage caused by road accidents 4. To promote multisector collaboration at the national level for the prevention of road accident injuries 5. To standardize prehospital medical emergency care of injuries.
2. Implementing Case-based Reasoning to Improve our Order Picking Model According to the INEGI data projected in the National Road Safety Profile of the Ministry of Health, in the period from 2011 to 2015, it is observed that the mortality rate has not increased, nor significantly decreased, which denotes apparent control as a result of the actions implemented in various programs. However, these results are insufficient for the fulfillment of the goal established in 2011 (to reduce the rate of mortality by 50%). On the other hand, the Mexican Institute of Transportation conducted a study called "Efficiency and/or effectiveness of road safety measures used in different countries (2013)" (Dominguez, Karaisl, 2013), in which, through a questionnaire applied to 22 countries, it is possible to identify 23 main implemented security measures and the effects obtained by conducting a Cost-Benefit analysis. For the identification of road safety strategies, of the 22 questioned countries, only one (Japan) does not use economic analysis methods liked Cost-Benefit Analysis (CBA) and Cost-Effectiveness Analysis (ACE). The rest of the countries go to accident records and compare the cost of implementing road safety measures against the impact coming from the human capital or the costs derived from the accidents (low productivity, costs per hospitalization, costs per repair and replacement, etc.). In this study, the lack of data is considered as the biggest barrier for a Cost-Benefit analysis, which ultimately results in the implementation of security measures with insufficient effectiveness.
3. Methodology The development of this project will be based on a variant of the methodology of the Double Diamond, proposed by the Green Dice company of the United Kingdom, which they called Triple Diamond.
36 Innovative Applications in Smart Cities Unlike the Double Diamond methodology, the Triple Diamond Methodology incorporates a third intermediate stage, moving the stage of Development to the intermediate diamond and adding two stages, the Distinction and the Demonstration. This methodology is selected because it integrates a series of steps typical of the development and implementation of a project, an essential stage in any innovation project: Distinction. In this stage, the aspects that show that the project has the character of innovation are highlighted. Next, each of the stages of the Triple Diamond methodology process are described in more detail. Discovery The discovery stage is the first in the methodology. Start with the idea initial and inspiration. In this phase, the needs of the user are stated and the following activities are executed; • • • •
Market research User research Information management Design of the research groups
Definition In the definition stage, the interpretation of the user needs is aligned with the business or project goals. The key activities during this phase are the following: • Project planning • Project Management • Project closure Development During this phase is when the planned activities for the development of the project are performed, based on the established plan, iterations and internal tests in the company or group of developers. The key activities during this stage are the following: • • • •
Multi-disciplinary work Visual Management of the project development Methods Development Tests
Distinction The stage of distinction reveals and establishes the characteristics that distinguish the project proposal of the rest of the proposals, also determines the strategy to continue to ensure that the target customers will actually choose to seek the product or service that is developed in the project. The key tasks during this stage are the following: • Definition of critical and particular characteristics • Definition of introduction strategy • Development of market strategy Demonstration In the demonstration stage, prototypes are made to evaluate the level of fulfillment of the project's purpose and to ensure that the design meets the problem for which it was created. Unlike the iterations
Analysis of Metropolitan Bus Drivers Mental Workload 37
that are made in the stage of development, the tests that are carried out during the demonstration stage are already made with the final customer. The key tasks in this phase are the following: • • • • •
Requirements compliance analysis Prototype planning and execution Execution of tests with end users Evaluation of the test results Definition of design adjustments
Delivery The delivery is the last stage of the Triple Diamond methodology. During this stage, the product or service developed is finalized and launched to the market and/or is delivered to the final customer. The main activities that occur during this stage they are the following: • Final tests, approval and launch • Evaluation of objectives and feedback cycles
4. Project Development Process The following describes the specific activities that are part of the project's development based on the triple diamond methodology. Discovery Stage: • Identification of the problem. • Dimensioning of the problem found and its effects, based on international, national and local statistics. • Research on the methods currently used to try to attack the problem, the resources used and their effectiveness. The investigation has been conducted through interviews with experts on mobility of the UAQ faculty of engineering, and it is planned to also attend the Mobility Secretariat of the municipality of Querétaro to capture the specific needs for consultation. • Identification of project actors, including experts in mobility and development strategies of data management platforms. The inclusion of these actors will be agreed voluntarily and their participation will be limited only to the consultation and audit of the proposal. • For the research of the users of the mobile application, meetings of focus groups will be developed to evaluate the usability of the same. Two rounds of evaluation of the participation of the users in the generation of the reports will be carried out through the implementation of a Beta version. Users will voluntarily download the application, and inside it they will find the information concerning the privacy warnings of their data, as well as an informed consent button where it is confirmed that they know the purpose of the investigation and the conditions of use: a. Design of a solution proposal to the problem, raise details and technical specifications of the project. In this stage, the relevant parameters are defined to fulfill the project objective: • High-level project design for determination of phases and parts b. Definition of scope and project partners c. Planning of project development d. Definition of resources for project development. The resources will be basically for the development activities of the platform and will be supported by personal contacts and also financed with personal resources.
38 Innovative Applications in Smart Cities 4.1 Stage of Development • Structuring of the database. It is taken as an initial part of the development as a strategic step to optimize the development of the mobile application and its performance designed for the user (data usage, optimization of device memory and processing resources) • Development of the first mobile application prototype to evaluate the user's experience during its use through focus groups • Definition of Story Boards for the mobile application • Development of the consultation platform • Platform performance tests • Establishment of servers and information management methods. 4.2 Stage of Distinction • Development of the gamification strategy for the mobile application. • Definition of the business scheme. The proposal will be based on a flexible scheme, which will allow expanding the scope and personalization of the platform depending on the geographical area where it is to be implemented. The platform will be developed as an independent entity, without any link with governmental institutions or private interests, but it will be open to adaptations in order to respond to the needs and particularities of whoever decides to adopt it. • Definition of modality for presenting the information in the consultation platform (georeference). Demonstration stage: • Carrying out the first integration tests of the platform • Analysis of the generated information, its processing and presentation to assess the level of value and practicality that they present • Analysis of compliance with initial requirements • Project validation
5. Development 5.1 Nature of driver behavior They will be analyzed in a general demographic factor that influences the behavior of the drivers. • The problem from the perspective of public health. The programs that have arisen worldwide and their adoption through programs to attack the problem. • Information Technology applied to Road Safety. In this section the available tools of information and how they have been used in terms of road safety will be reviewed. • Big Data. A brief overview of the term and its application to the draft. It will also explain how this concept is increasingly important in terms of commercial value and business opportunity. • Mobile applications focused on driver assistance and prevention of accidents. We will give a tour of the main applications currently available in the market and their contribution to the solution of same problem.
Analysis of Metropolitan Bus Drivers Mental Workload 39
5.2 Analysis of the mental workload of driver behavior Several factors influence the behavior of driver's moment of driving in their vehicles on public roads, they all influence directly or indirectly in the way in which drivers live in the Road environment in which they move: • Cultural factors. The culture is directly involved with the behavior of drivers. A general culture where respect for others is promoted allows that way of proceeding to be transferred to a road culture of greater harmony. This mentality, in turn, is influenced by a series of social and economic factors, which is why, in countries considered the first world, you can aspire to a road culture of greater respect and harmony. • Infrastructure. A city whose infrastructure was developed without proper planning generates a conflict environment for drivers, generating emotional stress that results in the modification of their behavior. The road infrastructure of a city exerts an influence on the motorist, modifying his temperament, behaviors, and answers, making him participate or propitiating the road chaos (Casillas Zapata, 2015). • Type of vehicle. The features and dimensions of the vehicle that is being driven influence the way the driver reacts. Characteristics such as the response to acceleration or maneuverability, as well as large dimensions of the vehicle, generate a state in the driver of courage prone to intrepidity (List, 2000). • Public politics. The use of public policies that regulate effectively the behaviors that generate road accidents generates an environment of the permissiveness of unsafe behaviors, propitiating conditions of a high probability of accidents. Considering a study conducted directly with the bus drivers of the four samples, we took into account those that were designed with several focus groups, which are concentrated in Figure 2, this information collected, collects various aspects related to mobbing, work stress and social isolation in each group of bus drivers, these samples consisting of 217 individuals (57 women and 160 men): Sample 1—Querétaro: 42 (F: 6; M: 36), Sample 2—Salvador de Bahía: 62 ( F: 15; M: 47), Sample 3—Palermo:
Figure 2: Visual representation of the analyzed sample, characterizing the diverse socio-economic aspects, the mobbing including the social blockade and the reflected labor performance.
40 Innovative Applications in Smart Cities 57 (F: 14; M: 33) and Sample 4—Metropolitan Area of Milan: 167 (F: 23; M: 44). In the case of Queretaro bus drivers (See Figure 2), which is the group that presents greater differences in the salary relation concerning the working day and this lies in its place of origin, the bus drivers who suffer more mobbing come from Oaxaca, Guerrero, and Veracruz; an intermediate group of bus drivers from Coahuila, Zacatecas and Durango try to group to negotiate with the majority group and finally the children of the most recent wave coming from the Federal District, State of Mexico and Morelos, they even become intimidators in their respective routes of transport, because they have the greatest social capital of the group and tend to be accustomed to longer working days, so this group can be considered completely heterogeneous in its relations with the majority group. Through public policies of the European Union, it is very easy to identify that the group of Milano bus drivers is first found, but it is very dispersed in the Salvador de Bahia samples where the work stress is greater, in Queretaro where the work-life relationship is not the most appropriate and in the little or no recognition that exists for the bus drivers in Palermo, which implies that there are more strikes than in the rest of the groups. 5.3 The problem from the perspective of Public Health The effects of car accidents worldwide are devastating from many perspectives, but especially that of Public Health, since it affects society emotionally, functionally and economically. In many cases, deaths and chronic injuries end up generating dysfunctional families, by losing one of their members, or by allocating part of their patrimony to the support of any of them (Ponce Tizoc, 2014). Similarly, in the economic aspect, automobile accidents have an economic impact of 1% to 3% in the respective GNP of each country, which amounts to a total of more than $ 500 billion (Hijar Medina, 2014), considering the material costs and the decrease in the productivity of the population. 5.4 Information Technology applied to Road Safety The use of Information Technology to solve many social problems has played an important role in the development and improvement of conditions of life. In the case of Road Safety, they have tried to implement different strategies based on Information Technology solutions, such as the use of urban monitoring and surveillance elements and the processing of data from mobile phones for the determination of congested roads and re-definition of routes, mainly in fleets of cargo vehicles. These types of benefits have also been raised through network vehicles, a technology that is growing and that will eventually allow an exchange of data between vehicles and urban networks, allowing to improve the traffic dynamics and street safety. The only drawback for them at the moment is a technology in development, which will begin to be implemented in some cities of the world, and eventually in countries like Brazil and Italy.
6. Mobile Applications Focused on Driver Assistance and Accident Prevention Currently, there are endless applications focused on the assistance of drivers in order to prevent accidents and help you reach your destination without inconveniences. Most applications have been developed to generate a network of information that allows you to be alerted to various circumstances, such as the volume of traffic in the user's route, dangers, obstacles and even checkpoints or presence of patrols. Some others, focus their functionality on the assistance to the user before an emergency. The rest of the applications available in the market are currently focused on education about traffic rules and best practices of driving. These types of applications have characteristics that can make them attractive for children and adolescents, providing a means of education, which may have a
Analysis of Metropolitan Bus Drivers Mental Workload 41
positive impact on the road culture of future generations at the steering wheel, as is proposed to a Smart City in Figure 3.
Figure 3: Comparative of technology use in a model of Smart City to improve the Mental health in bus drivers of four societies including the human factor.
Of all these applications, the most common worldwide is the so-called "Waze", which refers to the ways that a user can take establishing its destination within the application, taking into account the conditions of said route determined mainly by reports of other users of the same application. Recently, Volvo became the first automaker to launch radar-based security systems in automobiles with the launch of the XC90 Hybrid in India in 2016. Features such as Airbags and ABS have already begun to become a standard safety feature and the future; we expect more of these characteristics to become standard. The most specific implementation is associated when generally more than one vehicle is moving at different speeds, and the collision points can be higher than two, as can be seen in Figure 4. In the end, the facility that exists today to generate applications with various purposes, and the great boom that the use of mobile phones has had in different societies, allows us to visualize these types of conditions to use them as an excellent source for data mining. This data was provided to map them over the street map view of the city. We found that around 3,000 total accidents in the city
Figure 4: Model of IoT based on sensors for the control of the car and a radar added to identify obstacles in the proximity of up to one kilometer to identify automatic braking and implementation of a multi-objective security model for a smart city.
42 Innovative Applications in Smart Cities involve at least one motorcycle vehicle in them. Using a Kriging model, it is possible to determine the correct tendencies map in the future, as is shown in Figure 5.
Figure 5: A kriging model used to determine changes in an Ecological model and its Visual representation associated with the modeling of loss of forest in a landscape. Source: how-kriging-works.htm.
We fetched exclusively this data to map it in QGIS. To make all this information fit in the map in an ordered way, we took the spreadsheet file and randomly generated the latitude and longitude of the points of the map. All the other data fields were preserved as they came. Also, the spreadsheet software we used, LibreOffice Calc, to generate the random positions needed to be limited to all the points to fit in the area of Juarez. Calc doesn't generate decimal point random numbers, required for the cartographic precision in QGIS. To solve this problem, we created an algorithm to create positions. We used two different equations which represent each possible point in our model, as is possible to show in Equation 1 and Equation 2. latitude =
rand* (norther–souther) 10000
(1)
rand* (wester–easter) 10000
(2)
longitude =
The formulas above are the integer random generator for the positions of the points. The actual cardinal limits for the city are in the order of the same integer number (31.5997 south, 31.7845 north, –106.3077 east, –106.5475 west). To solve the problem, we multiplied all numbers by 10,000 to open the range and to let the Calc random algorithm to calculate with a more flexible grid the position of the points. After the number is calculated, the number is divided by 10,000 again. Once all the points were calculated, the spreadsheet file was parsed as a CSV (comma separated values) file, so the QGIS software can read the data of the records and print the layer of the locations. Once this was done, the Kriging method was applied. QGIS has a method to perform this by itself. Then the points were manually positioned to the closest road one by one. Also, some of the points were in zones that compared to OPM were outside the area of Juarez. These were allocated inside the desired area and sometimes deleted. As is shown in Figure 6. Once this was done, the Kriging method was applied. QGIS has a method to perform this by itself. The kriging method was applied to adjust the grid, the area size, removing the OSM beneath, so the new calculated layer would not be big. As they are still too small, indicated by the absolute
Analysis of Metropolitan Bus Drivers Mental Workload 43
Figure 6: Results of each point in a possible traffic accident in Juarez City.
positions of the points. The new kriging calculated layer was difficult to adjust, as is shown in Figure 7.
Figure 7: Final analysis of each point where a traffic accident will occur involving deaths associated with Italian scooter. The layer beneath the yellow dots (reddish colors) is the kriging generated layer that shows the most probable places in lighter colors. Some sites outside the city were shaded in the light; these could be errors in the algorithm.
44 Innovative Applications in Smart Cities If we compare this map above with the OSM, we can see the areas where the motorcycle accidents are frequent. Besides, we are proposing a tool for decision making using ubiquitous computing to help parents in locating children during the tour in case they had forgotten some object that is required for the class, as can be seen in Figure 8:
Figure 8: Representation of each node and their visualization in the streets related with our Kriging Model.
In the Google Maps city screenshot of traffic on Friday in the afternoon one can see that the places in dark red, red and orange are the places where traffic is worse than the areas in green, which we can say are fluid and empty. If one compares the areas in reddish with the kriging prediction in Figure 4, for example, the already mentioned Pronaf section and also the crossing bridge to the neighboring city, El Paso, which are very crowded due to security protocols in the United States. In Figure 9, we showed all avenues where the majority of traffic accidents related to Italian Scooters occur. The kriging applied to the map does not directly predict traffic, but traffic it is directly related to accidents.
7. Conclusions • Develop a mobile application (first stage) that serves as the interface for capturing reports of risk incidents. Citizens will be able to go through this application installed on their mobile phones, report risk situations that are observed during their daily journeys. • Design a scheme based on gamification that motivates users to make constant captures of reports of incidences of risk of transit. This scheme must allow a constant feeding of the database. • Configure a database (second stage) that contains relevant parameters for the definition of specific patterns of traffic risk. • Process the information contained in the database to present it in a geo-referenced way, classified by type of incident, schedules, type of vehicle and incidence coordinates through a consultation platform (third stage), graphically representing the red dots on a map, as in (Jiang et al., 2015). • Implementation of a predetermined warning function, which allows to notify other registered users directly about situations that they can put them or other users of public roads at risk, and that have probably gone unnoticed before the notification (conditions of lights, tires, etc.).
This innovative application can be used in other smart cities, such as Trento, Kansas City, Paris, Milan, Cardiff, Barcelona, and Guadalajara.
8. Future Research Through the implementation of this project it is intended to generate a platform of consultation that can provide accurate and relevant data in the study of urban mobility research, in such a way as to allow those in charge of generating proposals for improvement in this area to have the information available for its analysis. The configuration of the database will contain the data that are most representative and important to describe the risk factors: date, time, type of vehicle (through the capture of the license plates), coordinates and weather conditions. The information will be presented in the consultation platform in a georeferenced manner. This format will make it possible to identify areas of conflict and a specific analysis can be made through the implementation of filters that help to visualize the specific parameters of study in the map of a specific area. The platform will be limited to providing statistical information, as a way that can be easily tropicalized and adapted to local systems and legislation. Concerning users of the mobile application, to allow a constant feeding of the data, the system based on the gamification will motivate continuously, generating a status scheme within the application and giving them access to various benefits for their collaboration.
46 Innovative Applications in Smart Cities Since it is an information platform, the project can be used in different ways: • Access to the consultation platform by public security authorities may allow better planning and optimization of resources intended for the implementation of accident prevention measures vehicles, since more timely identification of the specific patterns to attack and the statistical data that they locate in temporality and frequency the most representative incidents. For example, it will be able to statistically identify the pathways where they occur with higher recurrence reports for speeding. This pattern will allow them to decide if it is necessary to concentrate patrolling in the detected areas and schedules or days in which most of the events reported are concentrated or if it is necessary to invest in the installation of cameras for speed/red light tickets. In this way, the planning of the resources destined for accident prevention will be greatly optimized. The instances of Government will be able to decide if they should increase the fleet of patrols or make a better distribution according to the incident statistics, generate benefit programs for a good rating in the platform, install surveillance systems at certain points, install traffic lights or speed reducers, and install specific signs in the zones detected, among others.
Multicriteria analysis of Mobile Clinical Dashboards for the Monitoring of Type II Diabetes in a Smart City Mariana Vázquez-Avalos,1 Alberto Ochoa-Zezzatti1 and Mayra Elizondo-Cortés2,*
In 2015 alone, 23.1 million people were diagnosed with diabetes, according to data gathered by the Centers for Disease Control and Prevention (CDC). This sum of people joined the estimated 415 million people living with diabetes in the world. The fast rise of diabetes prevalence and its life-threatening complications (kidney failure, heart attacks, strokes, etc.) has made healthcare and technology professionals find new ways to diagnose and monitor this chronic disease. Among its different types, type 2 diabetes is the most common found in adults and elderly people. Anyone diagnosed with diabetes requires a strict treatment plan that includes constant monitoring of physiological data and self-management. Ambient intelligence, through the implementation of clinical dashboards and mobile applications, allows patients and their medical team to register and to have access to the patient's data in an organized and digital way. This paper aims to find the most efficient mobile application for the monitoring of type II diabetes through a multicriteria analysis.
1. Introduction For a disease to be considered chronic, it has to have one or more of the following characteristics: They are permanent, leave residual disability, are caused by irreversible pathological alteration, require special training of the patient for rehabilitation, or may be expected to require a long period of supervision, observation, or care [1]. The last characteristic involves self-management and strict monitoring of the patient's health data, to avoid the development of life-threatening complications [2]. Diabetes checks all these characteristics; therefore, it is considered a chronic metabolic disorder characterized by hyperglycemia caused by problems in insulin secretion or action [3]. Type II diabetes, previously known as non-insulin-dependent diabetes or adult-onset diabetes, accounts for about 90% to 95% of all diagnosed cases of diabetes [4]. Over the years, its prevalence has been increasing all over the world and as a result, it is becoming an epidemic in some countries [5] with the number of people affected expected to double in the next decade. Furthermore, as it was previously mentioned, diabetes, as a chronic illness, requires constant monitoring of a patient's health parameters. Patient monitoring can be defined as "repeated or Universidad Autónoma de Ciudad Juárez, Av. Hermanos Escobar, Omega, 32410 Cd Juárez, Chihuahua, México. Universidad Nacional Autónoma de México, Ciudad Universitaria, 04510, CDMX, México. * Corresponding author: [email protected] 1 2
48 Innovative Applications in Smart Cities continuous observations or measurements of the patient and their physiological function, to guide management decisions, including when to make therapeutic interventions, and assessment of those interventions" [6]. However, most patients have difficulty adhering to the self-management patterns that diabetes monitoring requires. This current problematic, as well as the rise of other chronic diseases, has forced the healthcare system to transform how it manages and displays information to meet its needs [7]. New emerging systems for diabetes care have the potential to offer greater access, provide improved care coordination, and deliver services oriented to higher levels of need. The development of tools and implementation of ambient intelligence has brought technical solutions for doctors, as wells as help for the patients in the management of their health [8]. Ambient intelligence in healthcare Nowadays, new ways to interact with technology in our everyday life continue to develop. Ambient intelligence is an emerging discipline that brings intelligence to our everyday environments and makes those environments sensitive to us [9]. In the ambient intelligence environment, individuals are bounded by embedded intelligent devices' networks to collect information nearby their physical places, in the healthcare environment, these devices are used in medical informatics, decision support, gathered electronic health record and knowledge representation [10]. Health records and the development of clinical dashboards In the monitoring of diabetes, the continuous measurement of parameters such as blood glucoseare very important. The purpose of an electronic health record system (EHR) is to improve the wellbeing of patients and to avoid the organizational limitations set by paper-based records [11]. Patients are asked to track their key measures like blood pressure, heart rate, diabetes-relevant data, well-being, or side effects of the medication by taking daily notes on a piece of paper, called a health-data diary. The captured data are expected to show trends in the illness patterns and to help the doctor guide the patient to the best possible health status. Paper base diaries lack proper data representation, feedback, and timely delivery data. Therefore, an easy-to-use and patient-centered data-acquisition system is essential to guide the patient through data capturing and the complex process of self-management [1]. Healthcare is an environment that has been experiencing dramatic progress in computing technology in order to process and distribute all relevant patient information electronically and overall to improve the quality of care [12]. An EHR allows physicians to have easier access to the patient's parameters. Furthermore, the development of clinical dashboards made it faster to read this information. Dashboards are a tool developed in the business sector, where they were initially introduced to summarize and integrate key performance information across an organization into a visual display as a way of informing operational decision making. A clinical dashboard is designed to "provide clinicians with the relevant and timely information they need to inform daily decisions that improve the quality of patient care. It enables easy access to multiple sources of data being captured locally, in a visual, concise and usable format." [13]. These technologies have been implemented to mHealth (mobile health) apps, which have been defined as software that is incorporated to smartphones to improve health outcome [14], that make it possible for the patient to visualize their information and promotes self-monitoring to record their healthcare diary. A great variety of apps have been developed to assist patients in the management of diabetes mellitus type2 [15]. Some of the features involved in these self-management apps are medication adherence, self-monitoring of diabetes, insulin dose calculators and promoting physical activity and healthy eating [16].
Multicriteria analysis of Diabetes Clinical Dashboards
49
Some of the barriers that can be found in the use of mobile applications for diabetes management are cost, insufficient scientific evidence, not being useful in certain populations, data protection, data security, and regulatory barriers. In the first one, the cost is a potential barrier for any new medical technology. It can be hard for some patients to use these technologies due to the high cost of smartphones and lack of internet services. Regarding the insufficient scientific evidence, there aren't enough studies that show the effectiveness of these apps. On not being useful in certain populations, it's mentioned that a lot of these apps may not be useful for the elderly, non-English speakers and physically challenged. Also, the protection of the information uploaded to an application, as well as the proper use of this software and digital tools are important [16]. Application domain The different devices that exist in ambient technology, such as mobile apps that ease the monitoring of diabetes in everyday life, are helpful for patients and their medical team. Each software has different features that make one better than the other, in the following work seven different mobile applications will be analyzed to determine which one has the best performance. Problem statement When searching on their phone, a person with type II diabetes must choose among a large number of applications. To choose the best one, four evaluation criteria (functionality, usability, user information, and engagement with user), which were then divided into subsequent sub-criteria, were considered for the selection between seven different mobile applications available in online mobile stores.
2. Methodology There are different mobile applications (apps) that are used to keep track of the important parameters of a patient with diabetes, as well as encourage self-monitoring. The seven chosen to be compared by a multicriteria analysis (the analytical hierarchy process) are shown in Table 1. Diabeto Log is a mobile app developed by Cubesoft SARL, and it allows the user to see their blood glucose tests and medications intake in a single screen (Figure 1). This app was designed specifically to help the user see evolutions and compare data from one day to another. It also registers parameters such as the number of hypoglycemias, number of hyperglycemias, estimated A1c, number of insulin units used. Table 1: Mobile apps reviewed using multicriteria analysis. Mobile Applications Name
Developer
Diabetes Tracker: Diabeto Log
Cubesoft SARL
gluQUO: Control your Diabetes
QUO Health SL
BG Monitor
Gordon Wong
Bluestar Diabetes
WellDoc, Inc.
OnTrack Diabetes
Vertical Health
Glucose Buddy+ for Diabetes mySugr
Azumio Inc. mySugr GmbH
50 Innovative Applications in Smart Cities
Figure 1: Diabetes Tracker: Diabeto Log.
The purpose of GluQuo, an application developed by QUO Health, is to avoid hyperglycemia andhypoglycemia, control sugar levers, integrate all exercise data, and note carbohydrate rations. It also can connect to Bluetooth glucometers (Figure 2). In addition, it generates a report, allows the user to set up insulin reminders, keeps track of exercise and food intake with graphics, and helps in tracking diabetes with glucose dashboards.
Figure 2: GluQuo: Control Your Diabetes.
BG Monitor is a diabetes management app that has a clean user interface and filtering system that allows the user to find what they are looking for (Figure 3). It provides statistics that show blood glucose levels and can help to identify trends and make adjustments to insulin dosages. This app also has reminders to check blood or give insulin and creates reports so the user can email them from within the app. Bluestar diabetes is advertised as a digital health solution for type 2 diabetes. It provides tailored guidance driven by artificial intelligence (Figure 4). It collects and analyses data to provide precise, real-time feedback that supports healthier habits in users and more informed conversations
Multicriteria analysis of Diabetes Clinical Dashboards
51
Figure 3: BG Monitor.
Figure 4: Bluestar Diabetes.
during care team visits. In addition to glucose, it tracks exercise, diet, lab results, symptoms and medication. Ontrack Diabetes is a mobile app developed by Vertical Health. Some of its features are that it tracks blood glucose, haemoglobin A1c, food, and weight. It generates detailed graphs and reports that the user can share with their physician. It allows to easily keep track of a user's daily, weekly and monthly glucose levels (Figure 5). Glucose Buddy + helps in the management of diabetes by tracking blood sugar, insulin, medication,and food (Figure 6). The user can get a summary for each day as well as long term trends. It can be accessed as a mobile application, but it is also available for desktop and tablet. MySugr is an application that allows a digital logbook, shows personalized data analysis such as the estimated A1C (Figure 7). It also has Bluetooth data syncing to glucometers. A) Analytic Hierarchy Process (AHP) The analytic hierarchy process was developed by Thomas L. Saaty as a tool to manage qualitative and quantitative multicriteria [17]. There are several steps involved in the AHP. The first step is to define the problem and determine the kind of knowledge sought [18]. Several factors where evaluated to determine the best option among the different diabetes apps. The criteria and sub-criteria are listed in Table 2. These same criteria and sub-criteria are displayed
52 Innovative Applications in Smart Cities
Figure 5: Ontrack Diabetes.
Figure 6: Glucose Buddy +.
in the diagram of Figure 8. The information is structured with the goal of the decision at the top, through the intermediate levels (criteria on which subsequent elements depend) to the lowest level (which is usually an asset of the alternatives). When searching about these mobile applications, each one has different parameters in the subcriteria mentioned above. The values for each criterion are shown in Tables 3–6, where each letter represents an application: A is Diabeto Log, B gluQUO, C is OnTrack Diabetes, D is Bluestar Diabetes, E is BG Monitor, F is Glucose Buddy+, and G is mySugr.
In the next step, with the information available it is possible to construct a set of pairwise comparison matrices. To construct a set of pairwise comparison matrices, each element in an upper level is used to compare the elements in the level immediately below concerning it. To make comparisons, we need a scale of numbers that indicates how many times more important or dominant one element is over another element concerning the criterion or property for which they are compared [18]. Each number represents the intensity of importance, where 1 represents equal importance and 9 extreme importance (Table 7). With this scale of numbers, it is possible to construct the pairwise comparison matrices for the criteria and each sub-criteria, this is showed in Table 8. Once the comparative matrix of the criteria is done, the following step is to obtain the weight trough a standardized matrix. This is done by calculating the sum of each column, then the normalization of the matrix by diving the content of each cell by the sum of its column, and finally the calculation of the average rows, see Table 9.
54 Innovative Applications in Smart Cities
Figure 8: Diagram of The Criteria and Sub-criteria. Table 3: Sub-criteria of the Functionality criterion. Criterion: Functionality A
Table 7: Scale used to determine intensity of importance in pairwise comparison. Definition
Explanation
1
Equal Importance
Two activities contribute equally to the objective
2
Weak or slight
3
Moderate importance
4
Moderate plus
5
Strong importance
6
Strong plus
7
Very strong or demonstrated importance
Experience and judgement slightly favor one activity over another. Experience and judgement strongly favor one activity over another. An activity is favored very strongly over another; its dominance demonstrated in practice.
Table 8: Pairwise comparison matrix of the main criteria with respect to the goal.
As a conclusion, the user information is the least important criterion with a 0.060479 weight, followed by the engagement with user with a 0.11765 weight, then usability with a 0.28530 weight, and finally functionality is the most important criterion with a 0.53656 weight (Figure 9).
Figure 9: Weights of the criteria with respect to the goal.
To know if the comparison done and the weights obtained are consistent it is important to check the system consistency. The first step of this check is to calculate the weight sums vector: {Ws} = {M} . {W} The weight sums vector is in Table 10. λ
The value of consistency index is 0.068, while the value of consistency ratio is 0.06. The consistency ratio is CR < 0.1, thus it is acceptable, and the system is consistent. Table 10: Values obtained when calculating weight sums vector. M.W 2.34353 1.23710 0.24617 0.47769
Multicriteria analysis of Diabetes Clinical Dashboards
57
Now, to choose the best mobile application among the seven alternatives we must obtain the weight for each of the seven alternatives in each criterion. The first criterion analyzed is functionality (Table 11 and Table 12). Table 11: Pairwise comparison matrix for the sub-criteria with respect to functionality. Functionality Subcriteria
O/S
O/S
Int.
Mem.
Update
Rank
1
6
3
5
7
Interfaces
1/6
1
1/6
5
1/6
Memory
1/3
6
1
6
7
Update
1/5
1/5
1/6
1
1/4
Ranking
1/7
6
1/7
4
1
1.84
19.20
4.48
21.00
15.42
Total
Table 12: Normalized matrix of the pairwise comparison of the sub-criteria with respect to functionality. Subcriteria
Normalized Matrix
Weight
O/S
0.5426
0.3125
0.6702
0.2380
0.4540
0.4434
Interfaces
0.0904
0.0520
0.0372
0.2380
0.0108
0.0857
Memory
0.1808
0.3125
0.2234
0.2857
0.4540
0.2913
Update
0.1085
0.0104
0.0372
0.0476
0.0162
0.0440
Ranking
0.0775
0.3125
0.0319
0.1904
0.0648
0.1354
With the values obtained, the next step is to obtain the prioritization of the functionality subcriteria (Table 13 and Figure 10). Table 13: Prioritization of the sub-criteria with respect to functionality. Subcriteria
Weight
Prioritization
O/S
0.4434
0.2379
Interfaces
0.0857
0.0459
Memory
0.2913
0.1562
Update
0.0440
0.0236
Ranking
0.1354
0.0726
Figure 10: Prioritization obtained of each sub-criterion of functionality.
58 Innovative Applications in Smart Cities For the sub-criteria of usability, the same steps as before are used (Tables 14–16 and Figure 11). Table 14: Pairwise comparison matrix of the usability criterion. Usability Subcriteria
Language
Language
Cost
Target Group
1
6
Cost
1/6
1
7
Target Group
1/5
1/7
1
1.36
7.14
13
Total
5
Table 15: Normalized matrix of the comparison matrix. Sub-criteria
Normalized Matrix
Weight
Language
0.7352
0.8403
0.3846
0.6533
Cost
0.1225
0.1400
0.5384
0.2669
Target Group
0.1470
0.0200
0.0769
0.2439
Table 16: Prioritization of the usability criterion. Sub-criteria
Weight
Prioritization
Language
0.6533
0.1590
Cost
0.2669
0.0649
Target Group
0.2439
0.0593
Figure 11: Prioritization obtained of each sub-criterion of the usability criterion.
The same steps are followed for the user information criterion (Tables 17–19 and Figure 12). Table 17: Pairwise comparison of the user information. User Information Sub-criteria Rating No. Rating No. Down. Total
Figure 12: Prioritization obtained of each sub-criterion of the user information criterion.
Finally, for the criterion of user information the weights and prioritization are obtained (Tables 20–22 and Figure 13). Once the priority values are obtained the next step is to get the weight of each mobile application for each sub-criteria. The weight of each alternative is obtained following the same steps as before. Table 20: Pairwise comparison matrix for the sub-criteria with respect to the engagement with user. Engagement with user Sub-criteria Alerts
60 Innovative Applications in Smart Cities Table 22: Prioritization of the sub-criteria with respect to the engagement with user. Sub-criteria
Weight
Prioritization
Alerts
0.0688
0.0047
Passcode
0.1807
0.0125
Parameters
0.7557
0.0524
Figure 13: Prioritization of each sub-criterion of the engagement with user criterion.
The criterion of functionality has the sub-criteria of operating systems (Table 23 and Figure 14), interfaces (Table 24 and Figure 15), capacity that it occupies in memory (Table 25 and Figure 16), last time it was updated (Table 26 and Figure 17), and the ranking it has among other medical applications (Table 27 and Figure 18). Table 23: Comparative matrix of the alternatives with respect to the operating system. Sub-criteria: Operating System Alt.
A
B
C
D
E
F
G
Weight
A
1
1/7
1/5
1/7
1/5
1
1/7
0.02960
B
7
1
7
1
7
7
1
0.26396
C
5
1/7
1
1/7
1
1/5
1/7
0.05014
D
7
1
7
1
7
7
1
0.26396
E
5
1/7
1
1/7
1
1/5
1/7
0.05014
F
1
1/7
5
1/7
5
1
1/7
0.07823
G
7
1
7
1
7
7
1
0.26396
Figure 14: Weight of each alternative with respect of the operating system.
Figure 15: Weight of each alternative with respect of the interfaces. Table 25: Comparative matrix of the alternatives with respect to the memory it occupies. Sub-criteria: Capacity that it occupies in memory Alt.
B
C
D
E
F
A
A 1
1/6
1/7
1/6
1/7
5
G 5
0.06967
Weight
B
6
1
1/6
1/3
1/6
5
5
0.10696
C
7
6
1
6
3
5
5
0.35485
D
6
3
1/6
1
1/6
5
6
0.13595
E
7
6
1/3
6
1
7
7
0.26886
F
1/5
1/5
1/7
1/5
1/7
1
1/3
0.02497
G
1/5
1/5
1/6
1/6
1/7
3
1
0.03870
Figure 16: Weight of each alternative with respect of the capacity it occupies in memory.
61
62 Innovative Applications in Smart Cities Table 26: Comparative matrix of the alternatives with respect to the last update. Sub-criteria: Last Update Alt. A
A
B
1
C 1
D 5
1
E
F
6
G 1
Weight 1
0.1841
B
1
1
5
1
6
1
1
0.1841
C
1/5
1/5
1
1/5
5
1/5
1/5
0.0519
D
1
1
5
1
6
1
1
0.1841
E
1/6
1/6
1/5
1/6
1
1/6
1/6
0.0272
F
1
1
5
1
6
1
1
0.1841
G
1
1
5
1
6
1
1
0.1841
Figure 17: Weight of each alternative with respect of the last update. Table 27: Comparative matrix of the alternatives with respect to the ranking. Sub-criteria: Ranking C
D
E
F
G
Weight
A
Alt.
A 1
B 3
8
1/6
8
1/7
1/7
0.0897
B
1/3
1
8
1/6
8
1/7
1/7
0.0735
C
1/8
1/8
1
1/8
1
1/8
1/8
0.0176
D
6
6
8
1
8
1/4
1/4
0.1593
E
1/8
1/8
1
1/8
1
1/8
1/8
0.0176
F
7
7
8
4
8
1
3
0.3210
G
7
7
8
4
8
3
1
0.3210
Figure 18: Weight of each alternative with respect of the rank.
Multicriteria analysis of Diabetes Clinical Dashboards
63
The next criterion is usability, it has the sub-criteria of languages available (Table 28 and Figure 19), acquisition costs (Table 29 and Figure 20), and target user groups (Table 30 and Figure 21). Table 28: Comparative matrix of the alternatives with respect to the languages available. Sub-criteria: Languages available Alt.
A
A
B 1
C
D
E
F
G
Weight
4
5
5
5
1/8
1/7
0.1315 0.0718
B
1/4
1
3
3
3
1/8
1/7
C
1/5
1/3
1
1
1
1/8
1/7
0.0338
D
1/5
1/3
1
1
1
1/8
1/7
0.0338
E
1/5
1/3
1
1
1
1/8
1/7
0.0338
F
8
8
8
8
8
1
3
0.4179
G
7
7
7
7
7
1/3
1
0.2769
Figure 19: Weight of each alternative with respect of the languages available. Table 29: Comparative matrix of the alternatives with respect of the acquisition costs. Sub-criteria: Acquisition Costs Alt.
A
B
C
D
E
F
Weight
G
A
1
1
1
1
6
5
1
0.1840
B
1
1
1
1
6
5
1
0.1840
C
1
1
1
1
6
5
1
0.1840
D
1
1
1
1
6
5
1
0.1840
E
1/6
1/6
1/6
1/6
1
1/6
1/5
0.0279
F
1/5
1/5
1/5
1/5
6
1
1/5
0.0558
G
1
1
1
1
5
5
1
0.1800
64 Innovative Applications in Smart Cities
Figure 20: Weight of each alternative with respect of the acquisition costs. Table 30: Comparative matrix of the alternatives with respect of the target user groups. Sub-criteria: Target user groups Alt.
A
B
C
D
E
F
G
Weight
A
1
1
1/5
1/5
1/5
1
1
0.0526
B
1
1
1/5
1/5
1/5
1
1
0.0526
C
5
5
1
1
1
5
5
0.2631
D
5
5
1
1
1
5
5
0.2631
E
5
5
1
1
1
5
5
0.2631
F
1
1
1/5
1/5
1/5
1
1
0.0526
G
1
1
1/5
1/5
1/5
1
1
0.0526
Figure 21: Weight of each alternative with respect of the target user groups.
The user information criterion has the sub-criteria of user rating (Table 31 and Figure 22), number of user ratings (Table 32 and Figure 23), and number of downloads (Table 33 and Figure 24).
Figure 22: Weight of each alternative with respect of the rating. Table 32: Comparative matrix of the alternatives with respect of the user ratings. Sub-criteria: Number of user ratings Alt.
B
C
D
E
F
G
Weight
A
A 1
2
1/8
1/5
1/5
1/4
1/6
0.0315
B
1/2
1
1/8
1/5
1/5
1/4
1/6
0.0244
C
8
8
1
8
8
8
8
0.4559
D
5
5
1/8
1
4
6
1/5
0.1342
E
5
5
1/8
1/4
1
4
1/5
0.0928
F
4
4
1/8
1/6
1/4
1
1/6
0.0595
G
6
6
1/8
5
5
6
1
0.2014
Figure 23: Weight of each alternative with respect to the number of user ratings.
65
66 Innovative Applications in Smart Cities Table 33: Comparative matrix of the alternatives with respect of the number of downloads. Sub-criteria: Number of downloads Alt.
A
B
C
D
E
F
G
Weight
A
1
2
1/9
1/6
1/5
1/3
1/7
0.0281
B
1/2
1
1/9
1/7
1/6
1/4
1/8
0.0208
C
9
9
1
9
9
9
9
0.4357
D
6
7
1/9
1
4
5
1/7
0.1459
E
5
6
1/9
1/4
1
5
7
0.1425
F
3
4
1/9
1/5
1/5
1
7
0.0899
G
7
8
1/9
7
1/7
1/7
1
0.1366
Figure 24: Weight of each alternative with respect of the number of downloads.
For the engagement with user criterion, the sub-criteria to be analysed are the following: if the application send alerts or reminders to the user (Table 34 and Figure 25), if it allows a passcode before seeing the information that the user uploads on the mobile application (Table 35 and Figure 26), and the number of parameters that can be registered to maintain a management of diabetes (Table 36 and Figure 27). As the last step of the analytic hierarchy process, the values obtained before, shown in Table 37, are analyzed. Each sub-criterion is numbered by the order of appearance in Table 2. Table 38 presents the priority values of each sub-criterion. Table 34: Comparative matrix of the alternatives with respect of the reminders. Sub-criteria: Reminders/alerts Alt. A
A 1
B 1
C 7
D 1
E 1
F 1
G 1
Weight 0.1627
B
1
1
7
1
1
1
1
0.1627
C
1/7
1/7
1
1/7
1/7
1/7
1/7
0.0235
D
1
1
7
1
1
1
1
0.1627
E
1
1
7
1
1
1
1
0.1627
F
1
1
7
1
1
1
1
0.1627
G
1
1
7
1
1
1
1
0.1627
Multicriteria analysis of Diabetes Clinical Dashboards
Figure 25: Weight of each alternative with respect of the reminders. Table 35: Comparative matrix of the alternatives with respect of the passcode. Sub-criteria: Passcode Alt.
A
E
FD
1
1
1
1
7
7
1
0.1891
E
1/7
1/7
1/7
1/7
1
1
1/7
0.0270
F
1/7
1/7
1/7
1/7
1
1
1/7
0.0270
G
1
1
1
1
7
7
1
0.1891
A
1
B C
B
C
D
G
Weight
Figure 26: Weight of each alternative with respect of the passcode. Table 36: Comparative matrix of the alternatives with respect of the number of parameters. Sub-criteria: Number of parameters Alt.
A
B
C
D
E
F
G
Weight
A
1
1
2
1
4
1
1/4
0.1208
B
1
1
2
1
4
1
1/4
0.1208
C
1/2
1/2
1
1/2
4
1/2
1/5
0.0759
1/4
0.1208
D
1
1
E
1/4
1/4
F
1
1
G
4
4
2 1/4
1
4
1
1/4
1
1/4
1/5
0.0360
2
1
4
1
1/5
0.1178
5
4
5
5
1
0.4075
67
68 Innovative Applications in Smart Cities
Figure 27: Weight of each alternative with respect of the number of parameters. Table 37: Weights obtained for each alternative in each sub-criterion.
mySugr
Buddy+
BG
Bluestar
OnTrack
gluQUO
Diabeto
0.2639
0.0782
0.0501
0.2639
0.0501
0.2639
0.0296
1
0.1891
0.0270
0.0270
0.1891
0.1891
0.1891
0.1890
2
0.03870
0.0249
0.2688
0.1359
0.3548
0.1069
0.0696
3
0.1841
0.1841
0.0272
0.1841
0.0519
0.1841
0.1841
4
0.3210
0.3210
0.0176
0.1593
0.0176
0.0735
0.0897
5
0.2769
0.4179
0.0338
0.0338
0.0338
0.0718
0.1315
6
0.1800
0.0558
0.0279
0.1840
0.1840
0.1840
0.1840
7
0.0526
0.0526
0.2631
0.2631
0.2631
0.0526
0.0526
8
0.2495
0.2495
0.1366
0.0800
0.0365
0.0247
0.2228
9
0.2014
0.0595
0.0928
0.1342
0.4559
0.0244
0.0315
10
0.1366
0.0899
0.1425
0.1459
0.4357
0.0208
0.0281
11
0.1627
0.1627
0.1627
0.1627
0.0235
0.1627
0.1627
12
0.1891
0.0270
0.0270
0.1891
0.1891
0.1891
0.1891
13
0.4075
0.1178
0.0360
0.1208
0.0759
0.1208
0.1208
14
Table 38: Prioritization for each sub-criterion. Sub-criteria
Prioritization
Operating System
0.2379
Interfaces
0.0459
Memory
0.1562
Update
0.0236
Ranking
0.0726
Languages
0.1590
Costs
0.0649
User Groups
0.0593
User rating
0.0793
No. user ratings
0.0237
No. downloads
0.0105
Reminders
0.0047
Passcode
0.0125
Parameters
0.0524
Multicriteria analysis of Diabetes Clinical Dashboards
69
In Table 39 are the values obtained on the ultimate prioritization, each value was calculated with the weight the alternative for each sub-criterion and with the priority values of that same subcriterion. Table 39: Ultimate prioritization for each alternative. Alternatives
Last Prioritization
Diabeto Log
0.1015827
GluQUO
0.1365309
OnTrack
0.1361640
Bluestar
0.1620315
BG Monitor
0.0973677
Glucose Buddy+
0.1539829
mySugr
0.2144584
Therefore, using the values obtained with the analytical hierarchy process as reference (see Figure 28), the best alternative for a mobile application used to manage diabetes is mySugr, followed by Bluestar, then Glucose Buddy+. Meanwhile, the least preferred option is BG Monitor.
Figure 28: Ultimate prioritization values for each alternative.
B) Grand Prix The Grand Prix model is a tool that allows choosing amongst an array of alternatives considering the human and economic factor. It can be applied to the selection of the best mobile application for the monitoring of diabetes. As mentioned before, according to the AHP method, the best mobile app is mySugr with a score of 0.2144. The second-best option is Bluestar with a 0.1620 score, followed by Glucose Buddy+ with a 0.1539 score. It can be observed that, for these three options, the values obtained aren't far apart from each other. These three options are shown in Figure 29. Considering the economic factor, mySugr and Bluestar are freeware, whilst Glucose Buddy+ is a paid option. Analysing this information, it can be observed that mySugr is still the best option from an economic standpoint. On the other hand, considering the human factor, we sought the application that is easy to use for a wide demographic, especially older adults who are the main population affected by this disease. Also, it is available in a variety of languages and in both operating systems, so that it can reach a bigger amount of people. From these three options, the one that fits these characteristics is mySugr.
70 Innovative Applications in Smart Cities
Figure 29: Three best options with AHP.
However, if considering the languages, Glucose Buddy+ is available in 31 different languages, while Bluestar only in one. This means that Bluestar is not as accessible as Glucose Buddy+. With the Grand Prix model, it can be observed that amongst the top three options mySugr is still the best, but from the other two are still good choices. C) MOORA Method The MOORA (Multi-Objective Optimization on the basis of Ratio Analysis) is a method introduced by Brauers and Zavadskas. It consists of two components: The Ratio System and Reference Point approach. The basic idea of the ratio system part of the MOORA method is to calculate the overall performance of each alternative as the difference between the sums of its normalized performances which belongs to benefit and cost criteria. The first step of the MOORA method is to construct a decision matrix of each alternative and the different criteria: x11 x1n X = xij mxn xm1 xmn
(4)
This decision matrix shows the performance of different alternatives concerning the various criteria (Table 40). Next, from the decision matrix, the normalized decision matrix is obtained (Table 41). The following equation is used to obtain the values. xij =
xij
∑
; i =1,2, ..., m and j = 1,2, ..., n
m2 i =1
(5)
xij
The normalized decision matrix is weighted and shown in Table 42. The overall performance of the alternatives is measured by g
Where, g and (n-g) are the number of criteria to be maximized and minimized, respectively. And wj xij is the weight of the criterion. The results are in Table 43 and Figure 30. Table 43: Overall performance of the alternatives. g ∑wjxij j=1
n ∑wjxij j=g+1
yi
Ranking
A1
0.1642678
0.0784590
0.085808
4
A2
0.2248360
0.0503174
0.174518
3
A3
0.1341652
0.1054533
0.027811
5
A4
0.2828456
0.0041372
0.278708
1
A5
0.0656681
0.1921217
-0.12675
7
A6
0.2311133
0.2134249
0.017688
6
A7
0.3268881
0.086612
0.210276
2
Figure 30: Value of y for each alternative.
Multicriteria analysis of Diabetes Clinical Dashboards
73
3. Multivariate Analysis Cluster analysis groups individuals or objects into clusters so that objects in the same cluster are more like one another than they are to objects in other clusters. The attempt is to maximize the homogeneity of objects within the clusters while also maximizing the heterogeneity between clusters. Cluster analysis classifies objects on a set of user-selected characteristics. The resulting clusters should exhibit high internal homogeneity and high external heterogeneity. Thus, if the classification is successful, the objects within clusters will be close together when plotted geometrically, and different clusters will be far apart. The process followed is to start with all observations as their cluster, use the similarity measure and combine the two most similar clusters into a new cluster, repeat the clustering process and continue combining the two most similar clusters into a new cluster. The results in the hierarchical clustering can be represented as a dendrogram as shown in the Figure 31, which uses a single link (re-escalated cluster combination).
Figure 31: Dendogram.
4. Discussion There is a wide array of health mobile applications on the market. Since diabetes continues to rise in numbers, mobile apps for management of this disease are popular. Among all these apps seven alternatives were analysed using the AHP and Grand Prix method. From the first set of criteria, using the AHP method, it was concluded that functionality is the most important factor when looking for a mobile application, while the user information wasn't as important. These values were also used to find the priority values of each sub-criterion. From the functionality sub-criteria, the operating system had the higher values, thus it was deemed as the most important. From the usability sub-criteria, languages had the higher value. On the user
74 Innovative Applications in Smart Cities information criteria, the user rating sub-criterion was the most important. Furthermore, from the engagement with user criterion, the parameters that could be registered were the most important. Of each alternative, mySugr was considered the best alternative for the management of diabetes, Bluestar the second and Glucose Buddy+ the third option. The alternative with the lowest value was BG Monitor. The MOORA method was also used to analyse which of the seven alternatives is the best option. The results from this method are like the ones obtained through the AHP method. The number one option from AHP is the second option using MOORA, while option number two from AHP is the first and best option from MOORA.
5. Conclusions and Future Research The multiple-criteria decision analysis (MCDA) is an important and useful tool for a variety of fields. The AHP method allows us to compare many alternatives with respect to a set of criteria. The MOORA method is also another type of multiple-criterion decision analysis, and like the AHP method, it compares alternatives to certain criteria. These two methods showed similar results and, as a conclusion, it can be said that the mobile application mySugr is the best amongst the diabetes monitoring apps.
Electronic Color Blindness Diagnosis for the Detection and Awareness of Color Blindness in Children Using Images with Modified Figures from the Ishihara Test Martín Montes,1,* Alejandro Padilla,2 Julio Ponce,2 Juana Canul,3 Alberto Ochoa-Zezzatti4 and Miguel Meza2
Color blindness is a condition that affects the cones in the eyes; it can be congenital or acquired and is considered an average disability that affects about 10% of the world's population. Childrens with color blindness have particular difficulties when entering an educational environment with materials developed for people with normal vision. This work focuses on modifying the Ishihara test to apply it to preschool children. The proposed test helps to identify children who suffer from color blindness so that the teacher who guides them in the school can attend them.
1. Introduction The sense of sight in human beings as in other organisms depends on their eyes; these use two types of cells for the perception of images, i.e., rods and cones (Richmond Products, 2012). The rods are used to identify the luminosity, i.e., the amount of light received from the environment and the cones are used to identify the color or frequency in the spectrum of light received (Colorblindor, 2018). In most people there are three types of cones, each one to perceive a basic color, these can be red, green or blue, and the other colors that are generated are the result of various combinations that are received from amounts of light in tune with the frequencies of these basic colors (Deeb, 2004). The world around us is designed to work with the colors that are perceived with three cones, since most people can perceive the environment with three basic colors, i.e., they are trichromats, however, there are data of people with a fourth type of cone, which allows them to perceive more colors than the average person visualizes, however, these people often have problems describing the environment and tones they perceive, since the world is not made with their sensory perceptions in mind (Robson, 2016). Universidad Politécnica de Aguascalientes, Calle Paseo San Gerardo, Fracc. San Gerardo, 20342 Aguascalientes, Aguascalientes, México. 2 Universidad Autónoma de Aguascalientes. México. 3 Universidad Juárez Autónoma de Tabasco. México. 4 Universidad Autónoma de Juárez. México. * Corresponding author: [email protected] 1
76 Innovative Applications in Smart Cities On the other hand, there are also cases of people with a lower color perception, this condition is called color blindness and is considered a medium disability, since the colors with trichromat perception are used in various activities, such as to identify objects in a conversation, to identify dangerous situations, know when to advance in a traffic light, to decide what clothes to buy and to enjoy art forms such as painting or photography (Kato, 2013). Color blindness can be catalogued in four variants according to the cones available to perceive the environment, these variants can be anomalous trichromacy, dichromacy, and monochromacy or achromatopsia (Colorblindor, 2018). The most common variant of color blindness is anomalous trichromacy, in which there are all the cones for color perception, but there is a deficiency in some of them. Anomalous trichromacy can be separated depending on its severity into mild anomalous trichromacy, medium or strong and depending on the color in which the deficiency is presented, trichromacy can be divided into deuteranomaly or deficiency in the perception of green, protanomaly in the perception of red and tritanomaly in the perception of the blue (Huang et al., 2011). Another variant of color blindness that occurs less frequently, but more severely, is dichromacy. In this, one type of cone is absent, i.e., the person cannot perceive one of the basic colors, which causes problems with all colors that have this tone in their constitution, for example, a person who has problems with the green cone will have problems with all forms of green, but also with yellow and brown because they are constituted with the color green as a base. Dichromacy can also be classified depending on the absent cone, it is deuteranopia when the green cone is absent, protanopia when the red cone is absent and tritanopia when the blue cone is absent (Colorblindor, 2018). Monochromacy, or achromatopsia, is the rarest condition of color blindness, in this, all the cones are absent, therefore, the environment is only perceived in gray scales or luminosity and, although it is very rare, it represents a major difficulty for people who suffer from it, since they cannot live a normal life without the assistance of healthy people. These people cannot drink a liquid from a bottle without first looking for a label confirming its contents, they cannot identify if a food is in good condition before eating it, they cannot choose their clothes or identify a parking place, among other difficulties (Colorblindor, 2018). About 10% of people suffer from some deficiency or blindness to color, that is, about 700 million people suffer from color blindness, considering that the world population exceeds 7,000 million inhabitants. Table 1 shows the percentages of incidence in the world for men and women with each of the variants of color blindness (Colorblindor, 2018). As often as in adults worldwide, 1 in 10 children is born colorblind, facing a world that is not designed for him, generating various difficulties even in their learning and school performance (Pardo Fernández et al., 2003). Mexico has a similar situation in color blindness, as detailed in the study of the prevalence of color blindness in public school children in Mexico (Jimenéz Pérez et al., 2013). Table 1: Prevalece of color blindness in the world (Colorblindor, 2018). Type
Variant
Prevalence H/M
Monochromacy
Acromatopsia
0.00003%
Dichromacy
Deuteranopia
1.27 %
0.01%
Protanopia
1.01%
0.02%
Tritanopia
0.0001%
Deuteranomaly
4.63%
0.36%
Protanomaly
1.08%
0.03%
Tritanomaly
0.0002%
Anomalous Trichromacy
Detection of Color Blindness in Children 77
Children with visual difficulties associated with color blindness may have school problems when performing activities that involve color identification, such as using educational material with colored content, for example, relating content in their textbooks and participating in games, among other difficulties related to visual tasks (Pardo Fernández et al., 2003). Despite all the difficulties that children with color blindness are exposed to, this condition is one of the vision anomalies that take longer to be detected by parents and teachers, because by the intelligence or support they receive from other people they manage to get ahead with their education (Pardo Fernández et al., 2003). Teaching materials are often designed for people with normal vision, so it is important to detect vision impairments before school age so that children receive appropriate assistance in their educational processes (Jimenéz Pérez et al., 2013). The detection of color blindness can be done through the application of various tests that are designed according to the variant that is presented and the colors that are confused in this condition, for this, it is important to reproduce the perception of a colorblind people so that the tests be correctly designed. Several algorithms adjust the model of a digital image represented with red, green and blue parameters (RGB), to simulate color blindness and then allow people with a normal vision to see as people with this medium disability do (Tomoyuki et al., 2010). Figure 1 shows the color spectrum seen by an average trichromate, Figure 2 by a person with protanopia, Figure 3 by deuteranopia, and Figure 4 by tritanopia, all of them obtained applying color blindness simulation models.
Figure 1: Full color spectrum.
Figure 2: Full color spectrum perceived by a protanope.
Figure 3: Full color spectrum perceived by a deuteranope.
78 Innovative Applications in Smart Cities
Figure 4: Full color spectrum perceived by a tritanope.
The most common test used by ophthalmologists, based on the principle of detecting confusing colors, is the Farnsworth-Munsell test, in which the patient is presented with discs with colors of the entire color spectrum and is asked to make their arrangement in the correct order, this test is highly accurate in identifying any variant of color blindness and the severity with which it is presented, however, its application is highly complicated and in order to carry it out correctly, specific ambient conditions are required, such as a specific brightness of 25 candles at 6740 K, which describe the lighting conditions at midday (Cranwell et al., 2015). Figure 5 shows a simplified Farnsworth-Munsell test manufactured by Lea Color Vision Testing. The most commonly used test for the detection of color blindness are Ishihara plates, designed for the detection of protanopia, protanomaly, deuteranopia and deuteranomaly, however, any other variant of color blindness cannot be detected by Ishihara plates as they are designed to detect the colors that confuse people with these variants of color blindness (Ishihara, 1973). Figure 6 shows plate 42 that is seen by people with protanopia or strong protanomaly as a 2 and by deuteranopia and strong deuteranomaly as a 4. The tests to perform the Ishihara test are widely known and simple to evaluate, since the difficulties to perceive the color are observable when the individual who is submitted to the test, cannot see the number inside or it is difficult to visualize it, in addition, depending on the plate with which you have problems, it is identified which variant of color blindness is presented. Table 2 is used as a comparison list to identify the type of color blindness variant presented with the Ishihara test. The initial problem with this type of test is the purchase of the same, however, currently, there are several websites of organizations which allow a rapid assessment when there are suspicions of color blindness. An age-appropriate color blindness test (Jouannic, 2007) includes the possibility of detecting color blindness in toddlers when dealing with figures; however, given the options presented for each plate, the test can still be complicated for a preschooler. One of the images presented in this test is
shown in Figure 7. In this slide, the child is asked whether he sees a letter B behind the mesh, a star behind the mesh, or the mesh itself. Developing a test that can be done at the preschool level with a group of children would make it possible to raise awareness of the difficulties presented by some of the peers in that group and allow the preschool teacher to identify students who have problems with color perception, in order to adjust the activities to children who might face this type of condition in a preschool group. The aim of this chapter is to review the issue of color blindness, the difficulties, and guidelines when present in preschool children, while proposing images that can be identified through sight by groups of healthy children and disagree with those perceived by people with some form of
80 Innovative Applications in Smart Cities
Figure 7: Star test plate in (Jouannic, 2007).
color blindness, in this chapter focusing primarily on the identification of protanopia, protanomaly, deuteranopia, and deuteranomaly, to recommend that the child's parent go to a specialist to confirm the evaluation and present a specialist test such as Farnsworth's test. The tests are also mounted on an application so that they can be taken from home by the preschool-age child's parent or tutor.
2. Backgrounds The concern for color blindness in children is not an issue that has begun to be studied recently and there are several papers that are presented around this average disability. Several of these papers focus on identifying the incidence of color blindness in children in certain parts of the world, such as at (Jimenéz Pérez et al., 2013), where color blindness is detected in school-age children in Mexico. The same incidence is being studied in eastern Nepal (Niroula and Saha, 2010). In (Moudgil et al., 2016) a similar study is conducted on children between the ages of 6 and 15 in Jalandhar, in all cases using Ishihara plates. Another group of works seeks to identify problems that children with color blindness have and their detection. One of these studies shows that children with color blindness problems often have difficulties in identifying colors, but these are less than those expected in the color blindness model, as indicated in (Lillo et al., 2001). The work proposed in (Nguyen et al., 2014) shows the development of a computer interface for school-age children, it uses images suitable for children in a game that can be solved correctly by children with normal vision, however, they need instructions and supervision, making it difficult when working with a group of children.
3. Development Considering that the most used diagnostic tests and simple to evaluate are linked to the identification of confusing colors, images are designed with these colors using similar plates to those used in the Ishihara test, but instead of using numbers and figures, it is proposed to use drawings similar to those recognized at an early age. Another difficulty that should be kept in mind is to keep the indications simple, asking the child to confirm if what he or she sees is correct or not. For this purpose, the Microsoft Net Assembly speech interface is used to make the computer tell the child what it would have to see in the developed images. When opening the application, the first thing shown is a description addressed to the applicator or the teacher, which indicates what color blindness is, the purpose of the test and the instructions to follow in the application. Once the applicator clicks the button that starts the test, the images are presented to the children for three seconds, this following the indications of the original Ishihara test, since it is difficult for
Detection of Color Blindness in Children 81
children with color blindness to use the contrast and brightness conditions to try to identify the images; at the same time, with the speaker, the child is told what it should see and these indicate to the applicator whether or not they could see the figure. It is hoped that, with this, the applicator or teacher can identify which children have problems with the colors, inform the parents and take the child to a specialist in order to assess the child's condition, furthermore, the teacher should consider the special condition of the affected children when preparing the material for their classes.
4. Results The plates developed based on the colors used in the Ishihara plates, with designs that are known to preschool children are shown in Figure 8 to Figure 13. Table 3 shows the images obtained using a dichromacy color blindness simulation model (protanopia, deuteranopia, and tritanopia), they show how each plate looks like in each of the most critical color blindness variants.
Figure 8: Face pseudochromatic plates for children.
Figure 9: Tree plates pseudochromatic for children.
82 Innovative Applications in Smart Cities
Figure 10: Sweet pseudochromatic plates for children.
Figure 11: Boat pseudochromatic plates for children.
Figure 12: Sun pseudochromatic plates for children.
Detection of Color Blindness in Children 83
Figure 13: House pseudochromatic plates for children. Table 3: Perception of proposed pseudochromatic plates for children with different variants of dichromacy Normal perception
Deuteranopia
Protanopia
Tritanopia
84 Innovative Applications in Smart Cities The diagrams of the application generated at each of the moments of the test are shown from Figure 14 to Figure 20. Initially, instructions are shown when opening the application to start the test. When the applicator or teacher clicks the start test button, images begin to be shown (Figure 15 to Figure 20) as the sound produced with Microsoft Net Assembly tells the children which image they should see and then the applicator selects from the drop-down list whether the plate was viewed correctly or incorrectly.
Figure 14: Instructions shown in the application when opening the test.
Figure 15: Face shown in the application when the Microsoft Voice Assistant says "You should see a Face"
Detection of Color Blindness in Children 85
Figure 16: Tree shown in the application when the Microsoft Voice Assistant says "You should see a Tree"
Figure 17: Candy shown in the application when the Microsoft Voice Assistant says "You should see a Candy".
Figure 18: Ship shown in the application when the Microsoft Voice Assistant says "You should see a Ship".
86 Innovative Applications in Smart Cities
Figure 19: Sun shown in the application when the Microsoft Voice Assistant says "You should see the Sun".
Figure 20: House shown in the application when the Microsoft Voice Assistant says "You should see a House".
When the test is completed, a chart is received indicating that a doctor should be visited visually and by audio, in case of failure in the identification of the plates. The user has the possibility to do the test again.
6. Conclusions In this work, we manage to design plates with figures that preschool children can identify, as well as using them in an application that presents an audio aid that tells children what they should see on each plate. So, with the help of an applicator that could well be the teacher, children can take
Detection of Color Blindness in Children 87
Figure 21: Conclusion of the test indicating that there is deficiency in color perception and a doctor should be visited.
the test. The design of each of the plates is intended to make it difficult for people with problems in the different variants of color blindness that are detected by the Ishihara test, i.e., protanopia, protanomaly, deuteranopia, and deuteranomaly. The graphs show different figures depending on the variants of color blindness that are presented as shown in the results section. 6.1 Future Research Apply the pilot test with preschool children, to verify that they are familiar with the figures, as well as seek to detect real cases of color blindness using the proposed test.
An Archetype of Cognitive Innovation as Support for the Development of Cognitive Solutions in Smart Cities Jorge Rodas-Osollo,1,4,* Karla Olmos-Sánchez,1,4 Enrique Portillo-Pizaña,2 Andrea Martínez-Pérez3 and Boanerges Alemán-Meza5
This chapter presents a Cognitive Innovation Model that formalizes the basic components and the interactions between them for the establishment of a Cognitive Architecture (CA). The convenience of walking in the sense of achieving an archetype as support for the implementation of innovative intelligent solutions in Smart Cities and having the client as a means of convenient validation of the representation and processing of the knowledge expressed in the CA in comparison with those executed by humans in their daily activities.
1. Introduction Smart cities are a vital issue in this constantly changing world, dynamically shaped by science, technology, nature, and society, which implies people still face many challenges, both individual and social, where innovation plays a vital role. These constant challenges are now addressed more frequently through Cognitive & Innovative Solutions (CgI-S) which establish new schemes— innovation—of how to address them. Our current technological world uses a lot of pieces of knowledge, this means high valuable information, useful to solve a problem or satisfy a particular need and, of course, to drive innovation. Thus, the satisfaction of who has the problem or need is achieved when the knowledge is capitalized by CgI-S. Hence, the importance of finding out how to use and take advantage of as much of the creative expertise as possible, including imagination, even though its use in a systematic way is a complex challenge, even if only to share it through traditional ways, requires it to be made explicit. Even though dominating the challenge is a fundamental key for the cognitive era to progress and its artificial intelligence technologies, machine learning, cognitive computing, etc., coexist daily with humans. In the cognitive era, Cognitive Architects (Cg.Ar) together with specialists, from the domain to be treated, make up the Cognitive & Innovative Solution's Architects & Providers team (CgI-
90 Innovative Applications in Smart Cities SAP team) to provide CgI-S using highly specialized information, experience, creativity, coming from an ad hoc Collaborative Network (ahCN); which allows the team to do an adequate job even with innovation. Also, the CgI-SAP team applies science and technology to take advantage of this knowledge in order to achieve the Capitalization of Experience or Knowledge in solutions or innovation. It is undeniable that the above represents a complex situation [1, 2] since it requires a complete orchestration of the process, on the part of the Cg.Ar, which results in a CgI-S which requires technological developments and changes in the processes of the organization where the Cg.Ar must work side-by-side with ahCN. This arduous labour must be supported by a Cognitive Architecture, particularly apt, when cognitive approaches are required to meet the challenges of the cognitive era. This document is an effort to match situations or needs that should be faced with intelligent technologies and innovation processes, at times when the environment is extremely dynamic being this characteristic very typical within what is now called a cognitive era. The above motivates to provide a Conceptual Model of Cognitive-Innovation (CgI-M) as an Archetype that has formal support that consists of the Systematic Process for Knowledge Management (KMoS-REload); which formalizes the interaction between an ahCN, a Cognitive Architecture, and the CgI-S implementation process or particular treatment. The remainder of this chapter is structured as follows: In Section §2, sensitive concepts and related work to the subject are described. A general proposal Conceptual Model of Cognitive Innovation is presented in Section §3 where the ad hoc Collaborative Network, the Cognitive Architecture, and the dynamics of the Systematic Process for Knowledge Management (KMoS-REload) and its main characteristics are also presented. As an application, Section §4 introduces the start-up of the KMoS-REload process through a client study to describe the benefits of using the Conceptual Model of Cognitive Innovation and then presents the results of this study. A brief discussion is given in Section §5. Finally, the conclusion and future challenges are presented in Section §6.
2. Sensitive Concepts and Related Work In the present section, concepts and works are pointed out, both related to the model that is made known in this article, which is essential and summarizes the support towards the establishment of the archetype. 2.1 Informally structured domain A situation wherein an individual or company must face a challenge of adaptation to the cognitive era is treated in this chapter as a problem or need, and can consist of how to do it (processes), the incorporation of technologies, or both. Generally, who suffers from a situation, problem or need, belonging to the cognitive era, is aware of this situation but does not have the time, ability, or knowledge to determine the nature of the problem and less to give the appropriate treatment or implement actions that resolve it because the activities related to the dynamics and environment of the problem are constantly changing, which implies that the problem cannot be stopped. The organization and processes of such activities could be carried out in acceptable conditions, but to survive in the current environment, innovation is required. This innovation must start from the fact that there is no a knowledge-base where knowledge is formal and explicit, which generates gaps between the dynamics of processes and, even the communication between them. The knowledge of the environment is uncertain, ambiguous, and only some decision-makers and specialists in the domain have it but incomplete and with different degrees of specificity. The Conceptual Model of Cognitive Innovation (CgI-M) general proposal establishes how to deal with the situations or needs, mentioned above, through particular treatments in dynamic environments of the Informal Structure
Cognitive Innovation Archetype for Smart Cities Applicaitons 91
Domain (ISD). An ISD is a complex domain that can be described by characteristics of how its data, information, and knowledge are, and how are the representation and communication between them in the following way: • heterogeneous data and information; specialized knowledge with a high degree of informality, partial and non-homogeneous; and • knowledge that is mostly tacit and without structure. Besides, the ISD interacts with an ahCN that must understand the problem, need or business, identify application opportunities and obtain the knowledge requirements of this intricate knowledge
Figure 1: An overview of Informal Structured Domain's Eco System.
ecosystem to propose a convenient, viable and valuable CgI-S. In Figure 1, an ISD is characterized by exemplifying the context or environment of whoever requires a CgI-S. Finally, in the context of the ISD, in particular, the External Knowledge under the business concept must include the market and consumers; it is very important to understand from the beginning of the user experience under the integration approach of a value chain. The pace of business development, the amount of data and knowledge they handle from their clients and the need to insist on the concept of strategic adaptation—as opposed to the traditional strategic planning approach—have forced companies to think about a new approach that is different from the traditional "B2B". We know the importance of the consumer-oriented approach (Business to Client, B2C), we also know how the focus changes when we talk about collaboration between companies (Business to Business, B2B); however, under the current optics of handling artificial intelligence, machine learning, and cognitive technologies, it is necessary to evolve the last concept to a new one: Business to Business to Client (B2B2C). Under this concept, the need to understand the biometric profiles of final consumers adds an additional element to the appropriate handling of data or knowledge and its impact on the development of more efficient knowledge or predictive models. Companies supplying goods and services to other companies must now insist and collaborate in understanding the factors that motivate the choice of the offer of one company or another. That is, to be able to add value in a value chain, it is now not only necessary to understand the dynamics of the companies that are served, but the factors that motivate their respective market niches.
92 Innovative Applications in Smart Cities 2.2 Natural cognition process The natural cognition process can be managed as the capacity of some living beings to obtain information about their environment and, from its processing by the brain, interpret it, give it a meaning and further acted upon. In this sense, cognitive processes depend on both sensory capacities and the central nervous system. Therefore, Cognition is understood as the processing of any type of information through mental functions. It is important to mention that this conceptualization is derived from the traditional separation between the rational and the emotional; however, today emotion is seen as a cognitive process too. Thus, the faculties that makeup cognition are multiple from attention, language, metacognition (or knowledge about one's cognition) to emotion: • Perception. This is a cognitive process by which a mental representation of this information and its interpretation is generated. The process includes the capture of stimuli from the environment by the sensory organs and their transmission to higher levels of the nervous system. • Attention. It is a general ability to focus cognitive resources, such as selection, concentration, activation, monitoring or expectations, in stimuli or specific mental contents; therefore, it has a regulatory role in the functioning of other cognitive processes. • Learning and memory. Learning is defined as the acquisition of new information or the modification of existing mental contents (together with their corresponding neurophysiological correlates). Different types of learning have been described, such as classical and operant conditioning models, which are associated with synaptic potentiation mechanisms. Memory is a concept closely related to learning, since it covers the coding, storage and retrieval of information. In these processes, key structures of the limbic system are involved, such as the hippocampus, the amygdala, the fornix, the nucleus accumbens or the mammillary bodies of the thalamus. • Language is the faculty that allows certain living beings to use methods from simple to complex communication, both oral and written. • Although emotion, traditionally treated separately from cognition, is understood as equivalent to thought, it has been established that these two processes work in a similar way both at the level of the sympathetic nervous system and the motivation to approach or move away from a stimulus. • Reasoning and problem solving. Reasoning is a high-level cognitive process that is based on the use of more basic ones to solve problems or achieve objectives around complex aspects of reality. There are different types of reasoning according to how we classify them; if it is done from logical criteria, there is a deductive, inductive and abductive reasoning. • Social cognition. It includes several models that integrate the theories of attribution and the theory of schemes on the representation of knowledge. • Metacognition. The faculty that allows one's own cognitive processes to be known, by themselves, and reflected upon. Special attention is paid to metamemory, since the use of strategies to improve learning and memory is very useful to improve cognitive performance. 2.3 Innovation process Innovation is the change or series of changes, however minimal, through inventions that will always improve and add value to a product, a process, or service. The innovation can occur incrementally or disruptively and has to be tangible in the same process, or product or service resulting from it. Therefore, an innovation process is important due to: • Opportunities for problem-solving: When innovation is fostered, brainstorming arises from attempts to solve existing problems or needs.
Cognitive Innovation Archetype for Smart Cities Applicaitons 93
• Adapting to change: In the technological world, where the environment changes drastically, change is inevitable and innovation is the means, not only to keep a company afloat, but also to ensure that it remains relevant and profitable. • Maximization of globalization: Innovation is a necessity to solve the needs and challenges and take advantage of the opportunities that open markets around the world. • Be in competition: Innovation can help establish or maintain the vanguard of a company, compete strategically within a dynamic world and make strategic moves to overcome the competition. • Evolution of the dynamics of the workplace: Innovation is essential for the use of demographic data in the workplace, which constantly change, and ensure the proper functioning of the product, service or process. • Knowing the changing desires and preferences of clients: Currently, clients have a wide variety of products, services or processes at their disposal and are well informed to make their choices. Therefore, it is imperative to keep up with changing tastes and also forge new ways to satisfy the clients.
3. Conceptual Model of Cognitive Innovation Archetype: A General Proposal In this section, the CgI-M model is presented as a starting point to achieve an archetype: an original mould, or pattern, that links the process where elements or ideas intervene in order to establish an architecture to support a Cognitive Solution, a Cognitive Innovation. The archetype proposes models of knowledge representation that include experiences, behaviours and ways of thinking, collectively shared, which are constructed from theoretical tools—by imitation or 185 similarity to human ones—assimilating and codifying knowledge, defining the existing relationships between concepts of the Informally Structured Domain given. Consequently, the archetype produces a physical or symbolic solution, tangible or intangible things or processes, or parts of them, that could generate something more of themselves. Figure 2 shows a general outline of the CgI-M model. It 3.1 ad hoc collaborative network As previously pointed out, cognitive innovations use creative experience or specialized knowledge from various sources that can be entities, agents, systems, etc., who possess knowledge or information, and together set up an ad hoc Collaborative Network (ahCN). An ahCN is compound by a triplet where: ahCN = (IK, IS, EK)
(1)
Equation (1): The ahCN is equivalent to a triplet compounded of knowledge or information sources, internal or external, from a given domain; where: • Internal and External Knowledge (IK & EK) are pieces of Knowledge or Experiences that are present in the Informally Structured Domains (ISD) and that compound a set of abstract representations with the useful purpose of solving or addressing something that happens in their environment that they are stored through experience or they are acquired through the senses. Such pieces are obtained, or come, from Internal or External Agents who belong to different fields of the same domain, and they know about the problem, their environment and the actions that must be carried out in it and they usually can be: specialists, decision makers,
Figure 2: A general outline of the Cognitive and Innovation Model.
94 Innovative Applications in Smart Cities
Cognitive Innovation Archetype for Smart Cities Applicaitons 95
stakeholders, competition, workforce, clients, knowledge requirements engineers, cognitive engineers, cognitive architects, etc. • Information Systems (IS) is the information or data from a system what can be understood as a set of components that interact to collect, organize, filter, process, generate, store, distribute and communicate data. The interaction occurs with a certain number of users, processors, storage media, inputs, outputs and communication networks. A system with access to selected data clouds, databases or research sites about a given domain; and it is important to emphasize that: 1. the data, information or knowledge sources of the ahCN can be largely autonomous, geographically distributed and heterogeneous in terms of their operating environment, culture, social capital or goals, but they work in close collaboration to achieve the best common goals or, at least, compatible ones, and whose interactions could be internal, external or both to ensure proper functioning of the ahCN [3]; 2. the Knowledge or Experience belonging to agents, from different fields, is capitalized in the Cognitive/Innovative Solution (CgI-S); 3. the Pieces of Internal Knowledge (IK) are considered as the foundation of the solution and the Pieces of External Knowledge (EK) are considered as the feedback the solution and influence, motivating the CgI-SAP team, to provide the best solution. It is important to note that some EK pieces come from neuroscience, biometric profiles, often trivialized by Artificial Intelligence, but generate updated perceptions from the user's evolutionary experience, traditionally presented as insights. 3.2 Cognitive architecture In this cognitive era of surprising changes that have taken place in an extremely fast period, the idea of Cognitive Architecture must be properly delimited. Therefore, the authors consider it convenient to achieve the homogenization of concepts or paradigms, related to the cognitive field and hypotheses about the nature of mind, among those who work in this area. The task is hard because every day something new arises, but it is worth trying to go in the same direction. Consequently, we agree with [4] when they point out that Cognitive Architectures are hypotheses about fixed structures, and their interactions, intelligent behaviour and underlying natural or artificial systems. In essence, a Cognitive Architecture must have a Semantic Base, derived from a Cognitive Analysis; which, in turn, is the essential component of the Cognitive System that must support a CgI-S. Semantic Base. The semantic base formalizes, through a consensus, the relationships between concepts or terms and their attributes belonging to the domain related to CgI-S. The terms are registered constituting knowledge through an extended lexicon (KDEL) that classifies them into objects, subjects and verbs and is based on LEL [5]. The externalization of this knowledge allows the achievement of a consensus among the interested parties and, consequently, minimizes the symmetry of ignorance. The concepts and relationships identified generate a matrix called Piece of Knowledge (PoK). It also facilitates the construction of a conceptual graphic model that provides a visual medium for the semantic base of the domain and facilitates its validation, where an entityrelationship model can be used. Generally, after forming a semantic base, it is common to find that a good amount of terms used in the domain are ambiguous, are not unified and are particular to those who use them. It is important to bear in mind that, although the domain specialists validate the description of the concepts of the lexicon, the graphic conceptual model provides a very complete description of the knowledge of the domain that allows domain specialists to identify possible errors and what lacks in the semantic base; particularly, between the relations of the concepts. This is very important since this model is essential for the design of a Cognitive Architecture.
96 Innovative Applications in Smart Cities Cognitive System. Set of entities, definitions, rules or principles that when interrelated in an orderly manner contribute to formalizing a cognitive process, at least the irreducible set of components that are used to explain or carry it out. 3.3 Knowledge management on a systematic process (KMoS-REload) The Systematic Process for Knowledge Management KMoS-REload (Figure 3, all details in [5]) is specially designed to interact with Informal Structure Domains (ISD), supporting the Cognitive Analysis, and provides a formal procedure for obtaining, structuring and establishing formal knowledge relationships that serve as a guide for the cognitive architect to: (a) integrate the Cognitive Architecture that supports a Cognitive and Innovative Solution and avoid ambiguity, incompleteness and inappropriate links between pieces of knowledge in the context of a given Informally Structured Domain; and (b) coordinate and operate the CgI-M model. In particular the process performs three sequential phases: 1. Conceptual Modelling Phase which models the CgI-S 's domain using a linguistic model and a graphic conceptual model; 2. Strategic Model to visualize the general functionality of the CgI-S 's domain; and 3. Tactical knowledge phase, which is in charge of obtaining, discovering, giving structure and enrichment to the knowledge of the CgI-S. In addition, cross-cutting activities are included to identify tacit knowledge, and, once this knowledge is explicit, the wrong beliefs are recorded and the relationships between the concepts and their behaviors are traced. Three activities complement the models used in the process:
Figure 3: General overview of the KMoS-REload process is represented by activities flow diagram.
Cognitive Innovation Archetype for Smart Cities Applicaitons 97
1. The tacit identification of knowledge, where a discourse analysis is carried out with the objective of identifying the knowledge that is hidden behind the linguistic indicators as presuppositions; 2. The capture and updating of specialized knowledge of the matrix, throughout the process, knowledge is associated with those involved in the domain, which forms a matrix that captures experience in the domain; and 3. The assumption record, a phenomenon that occurs when learning a new domain is to associate our mental scheme with new concepts and relationships, therefore, false assumptions will be made clear as the process progresses, these assumptions must be recorded to facilitate the learning of new members in the project. The process begins with an initial interview between the Solution's Architects and Providers (CgI-SAP team) and the Internal or External Knowledge (Domain Specialists) in a session where socialization predominates. Then, the Tacit Knowledge Identification, the Expert Matrix Update and the Assumptions Record are developed in parallel by C.Ar (or the CgI-SAP team)—a Cognitive Analysis is done by them in a socialized way—in order to verify the artifacts and decide if they should continue with the following phases or require a validation of them. In fact, under a lean or agile innovation approach, living iterative processes exist that allow adding value from the validation of their elements and proposals. The validation requires that the CgI-SAP team explain the models in order to validate the knowledge. The process-in-turn generates more knowledge, then the cycle starts again, and the process may end when all those involved in the CgI-M Model reach an agreement. Finally, the process makes the team aware that, in order to develop a CgI-S, it is necessary to understand and formally define the knowledge requirements and the domain that circumscribes them [6]. The details of the KMoS-REload process application can be found in [15]. 3.4 The cornerstone of the cognitive/innovative solution What is a solution? In the CgI-M model context, a solution means solving a situation, problem or need of an individual or company (the client), through the experience and talents of a highly specialized team of people. Despite the fact that the concept of a solution is simple, the cornerstone of a Cognitive/Innovative Solution (CgI-S) is the result of processes and actions, obtained collaboratively. Cognitive/Innovative Solution. The CgI-S is the result, given by Solution Cognitive Architects and Providers (CgI-SAP team), of solving a problem or cognitive need taking into account the connections and relationships of the models obtained from Cognitive Analysis (CgAn), making use of the Internal Knowledge (IK) and External Knowledge (EK), and any other feedback from the Informal Structure Domain. Thus, CgI-S can be represented as a function of three parameters that can be represented by the Equation (2). CgI − S = F (CgAn, IK, EK)
(2)
Equation (2): The CgI-S is the result of the development and implementation function carried out by the CgI-SAP team. At this moment, it would be convenient that two concepts are in mind: Open Innovation and Corporate Venturing. Today, companies are learning that their innovation models and proposals can find more value—and much faster—if they find a way to integrate the approach and proposals of their potential clients into their innovation models, by the way, users of their technologies can usually express their needs more easily with respect to technology itself. Today, reorienting its research and development efforts, originally armoured towards an open innovation approach that includes the multiplied vision of its clients, adds much greater innovation potential and more variation. Companies such as Telefónica are leading worldwide collaboration initiatives such as these; the term that has been coined to name this type of effort is that of "Corporate Venturing",
98 Innovative Applications in Smart Cities where companies allocate resources to encourage start-ups or small businesses to develop new concepts, indeed much more economically accessible. Cognitive Innovative-Solution Architects & Providers (CgI-SAP) team. This is a team of human talent that performs consultation and analysis of information technology systems, intelligent and cognitive. The CgI-SAP team supports all its activities, within the CgI-M model, in a Systematic Process for Knowledge Management (KMoS-REload) to develop cognitive and, therefore, innovative solutions that bring great value to clients. It is well known how engineers or scientists become obsessed with past solutions and how the process of scientific discovery and the engineering design process can lead them to new solutions. However, there is still much to understand about the cognitive and innovative processes, particularly with respect to the underlying natural cognitive processes. Behind the KMoS-REload process, there are theories and methods of several disciplines related to cognition and knowledge, such as cognitive psychology, social psychology, knowledge representation, machine learning to analyze, structure and formalize the complex cognitive processes that occur in the real world, the world of the Informal Structure Domains. It implies that the CgI-SAP team is highly trained to be empathetic and solve problems of a given Informally Structured Domain. Consequently, there are two essential roles carried out by this team: as an architect of solutions, the team must have a balanced combination of technical, social and business skills; as a supplier, the team must offer solutions based on any combination of technologies, processes, analysis, commercialization, internal organizational environment or consulting. Such solutions can be customized for your clients; or, it can provide solutions based on existing products or services. Regardless of the roles played by the CgI-SAP team, the core of its activity is the interaction with the elements of the triplet of Equation (2) and applying science and technology advances to take advantage of all the knowledge that exists around to achieve the Capitalization of Experience or Knowledge and to provide a CgI-S. It is undeniable that the above represents a complex situation [1, 2], but also an excellent opportunity for the CgI-SAP team. Cognitive Analysis (CgAn). The CgAn is a process of examining in detail a given ISD in order to understand it or explain it. Commonly, one or several strategies or processes are used that enable the knowing and formalizing of the existing relationship between certain types of functions, actions and concepts related to this domain. The main objectives of performing the CgAn in a given ISD are: (a) to obtain the best view of your own internal processes, e.g., in a business domain this could be how the market receives its products and services, customer preferences, how customer loyalty is generated or other key questions where precise answers are used to provide a company with a competitive advantage; and (b) to set up the cognitive architecture established by the semantic base and the components of the appropriate cognitive system. It is worth mentioning that the CgAn often focuses on the realization of a predictive analysis, where the extraction of data and other cognitive uses of the data can generate business and commercial predictions. Therefore, the practical problems surrounding such analyses involve the precise methods used to collect and store data in a special location, as well as the tools used to interpret this data in various ways. Solution Cognitive Architects & Providers can provide analysis services and other useful help, but in the end, the practical use of the analysis depends on the people who are part of the domain, where they not only need to know how to collect data but also how to use it correctly.
Cognitive Innovation Archetype for Smart Cities Applicaitons 99
3.5 Agile process of innovation The high dynamism and constant change of the world and its markets require that innovation is contained in an agile, continuous, cyclical and constant process of changes and adjustments where the CgI-S frees time from the process actors so that they focus on supervisory activities and that can agilely search for new products, services, internal processes or improvements, adaptations or updates to existing ones. Currently, a "complete study of x-ray + computed tomography + magnetic resonance imaging" of the client's environment or its ISD is required to identify areas of opportunity and map the process, know the products and services to clarify and be assertive in the client's vision and goals. From the beginning of the KMoS-REload that will implement the CgI-S, through the CgAn, this "complete study" starts and the client will become aware of the intangible good that will be obtained. The Cognitive Architecture, since it is being formed, is offering the client content and tentative activities to be carried out. It is necessary to highlight that at the beginning it is impossible to detail the components of the architecture to the minimum since a given ISD is unknown and, in the same way, the end of the process is relative since it depends on the client's satisfaction concerning its environment. As the environment is changing, the cognitive architecture and, therefore, the solution could/ must change, that is, innovate. The concept of agile innovation implies an organizational culture that is prepared with the necessary technological architecture to be at the forefront, but also with the appropriate mentality to assimilate the exhausting challenge of permanent change. That is to say, it is useless to have an environment full of cutting-edge technology, when the mentality of the organization remains anchored in past paradigms of work in silos, focused on particular objectives and leading profiles. Finally, the expertise of real solutions implementation indicates that innovation is implicitly presented—however marginally—and even more, it accelerates the cyclical process of innovation whose impact can occur as an improvement of Products, Services or Processes; or, the generation of new ones.
4. FLUTEC: A Client Study FLUTEC worldwide company—located on the US-Mexican Border (Juarez City)—designs, builds and sells Heating Ventilation and Air Conditioning (HVAC) modules tailored to meet the particular needs of its clients; that is, each module could be similar but not identical. In fact, a build-to-suit approach for every project makes for a high-cost project. To find greater benefits from a project requires the improvement of the process to carry out it. The HVAC project process starts when a client issues the basic specifications for its design and ends with the delivery of it. Therefore, it includes a mare magnum of aspects to take into account when carrying out a project and, consequently, an erroneous decision directly impacts the time of the general process and even its viability. In addition, the dynamism of the HVAC's singular market motivates the company to find greater benefits, and at the same time obliges it to continuously improve its processes, especially the delivery time of the project budget, the time and the quality of the design process. Is the implementation of the CgI-M model convenient? The company, and all its processes, are subjected to the HVAC's market to innovate continuously; otherwise, take a risk that compromises the survival of the company. Besides, all characteristics of the ISD, indicated in subsection §2.1, are present in FLUTEC environment. Therefore, the HVAC's domain specification can be listed as: • Seven main processes at least required to achieve it: Heating, Cooling, Humidifying, Dehumidifying, Cleaning, Ventilating, and Air Movement;
100 Innovative Applications in Smart Cities • Five complex tasks that include each one's activities: Basic specifications establishment, building characteristics analysing, Air circulation patterns analysing, Appropriate components selection, and Control system analysing; • Non-organized and incomplete data is present in it; • Determination of the criteria and decision making about the achievement of the project is carried out under the umbrella of an ahCN; and • The project has a unique design and solves or addresses a particular situation. To deal with the challenges of obtaining the knowledge requirements of an HVAC project, typified as belonging to an ISD, the company FLUTEC uses an empirical guide—DNA document— composed of general attributes that gather the necessary basic information for each project. This document should be a guide to obtain the knowledge requirements that would allow a good design of an HVAC module. However, being an empirical and, therefore, informal document, it was a very flimsy communication bridge between the ahCN within FLUTEC. In addition, there were often delays, reworkings and high-cost problems arising from this DNA document and the additional processes of the FLUTEC's processes related to the realization of a project. Characterization of the CgI-M model through the determination of the peculiar attributes and additional activities related to the HVAC project. Once the Flutec's environment relative to the HVAC's design process has been identified as an ISD domain, the Cognitive Architect starts the KMoS-REload process to characterize and, consequently, establish the CgI-M model: • Distributed Tacit Knowledge: Tacit distributed technical knowledge, heterogeneous, diverse degrees of specificity; • Incomplete data: Unorganized and Incomplete Data of all the processes related to the HVAC and that should be used in the development of the Cognitive Architecture Specification; • ad hoc Collaborative Network: composed of multiple specialists in the Flutec's domain, the CgI-SAP team, decision-makers; and • any other problems, in particular, must always be addressed when developing an HVAC; therefore, it is a unique project that requires a CgI-S. Results of the use of the CgI-M Model. In order to provide FLUTEC with an adequate cognitive solution (CgI-S), the CgI-SAP team identified the elements of the HVAC's process that needed to be improved and established a consistent model that would give it the corresponding support from the following: • Analysis of the DNA guide document. As mentioned above, the DNA guidance document is empirical and lacks the overall vision of the project. Therefore, the analysis should describe the significant assumptions and conceptual relationships of all FLUTEC knowledge. Consequently, the DMP confirmed that the DNA had the following deficiencies: − Disorganization; − Incomplete: Missing essential information for the proper development of the HVAC project; − Incorrect: Existence of fake attributes; − Irrelevant information: Informal descriptions had been recorded; − Ambiguous information. The initial and basic knowledge requirements were not well described; and − Time lost due to searching, as often as necessary, of missing or poorly recorded information. The analysis allowed the obtaining of knowledge, the formalization of empirical DNA domain, and its transformation into a new and formal solution.
Cognitive Innovation Archetype for Smart Cities Applicaitons 101
• Specialized Explicit Training. Before applying the KMoS-REload process, it was already difficult for FLUTEC engineers to understand the importance of a formal process of obtaining knowledge requirements and, consequently, there was great ignorance about certain elements or concepts belonging to the domain of the project. Once the process was used in the project, the FLUTEC's specialists were trained by the domain modelling phase and were able to assimilate (make tacit) new explicit knowledge, reduce their own ignorance and ambiguity and, as a result, improve the quality of work in ahCN learning that: − Knowledge-Requirements Elicitation could be carried out systematically; − The CgI-M Model transfers knowledge; and − FLUTEC has preconceived and tacit ideas or expectations of the project and when they turn explicit the redesigns are usually avoided on the project's post-delivery time. • Improvement of HVAC-DNA Process. The CgI-M Model was carried out through KMoS-REload process, as a result, the models will be established and, therefore, the HVAC-DNA process will be renewed: − HVAC project concepts, attributes, relationships between concepts and basic integrity restrictions were formalized, e.g., HVAC design and budget project properties. The externalization, transfer, and consensus are activities carried out within the ahCN with its knowledge in order to integrate a set of pieces of explicit knowledge that minimizes the symmetry of ignorance. Thus, the learning curve about the HVAC domain was reduced from a couple of months to a couple of weeks. In addition, the CgI-SAP team noticed that the DNA document was not useful during the project process, especially, since it requires a lot of time to be filled and does not meet the goal it is supposed to achieve. − The process view as a stream of decisions from the ahCN allowed the CgI-SAP team to obtain a Cognitive Architecture with support of the KMoS-REload process. − The set of knowledge-requirements are derived and integrated into the CgI-S's specification document. A CBR to support a fast delivery of proposals. The cognitive architecture from the knowledge, acquired and managed, also allowed to constitute as a part of the CgI-S: a robust Case-Base, textual files and all necessary to implement a CBR prototype in the jCOLIBRI tool [7, 8]. This tool provides a standard platform for developing CBR applications through specialized methods using several Information Retrieval or Information Extraction libraries as Apache Lucene, GATE, etc. Thus, an important goal achieved by the CBR was to demonstrate if FLUTEC could do a time reduction in obtaining the matches between the expectations of the clients with the HVAC project design blueprints and, consequently, a fast delivery of budget proposals. In summary, the establishment of an adequate cognitive architecture, using the KMoS-REload process, manages to capitalize the knowledge of the ahCN and its expertise, explicitly and formally, to allow: a clear understanding of the project's ISD; its assimilation by the CgI-SAP team; give a CgI-S; and characterize, as a whole, the CgI-M model—to a total customer satisfaction—whose remarkable products were a new DNA guide and the CBR prototype.
5. Discussion We live in a world that changes minute by minute, for better or for worse, due to advances in science and technology strongly framed in artificial intelligence, machine learning, and cognitive computing. The assimilation of the advances is not a trivial issue and, consequently, the companies, the individuals, and the society must find the way to survive at the great speed with which "the future and the present are amalgamated". This issue is not trivial because changes that happen too quickly can often produce a disconnection between a scientific or technological advance and
102 Innovative Applications in Smart Cities the understanding of its potential by the providers of technological solutions. Scientifically and technologically speaking, there are many examples throughout history of wrong judgments about what the future holds. For example, when business owners introduced electricity into their factories, they stayed with older models of how to organize the use of their machines and industrial processes and lost some of the productivity improvements that electricity enabled. In 1977, the president of Digital Equipment Corporation, the largest computer company of that time, saw no market for home computers. Thirty years later, in 2007, Steve Ballmer, CEO of Microsoft, predicted that the iPhone would not take off [9]. From these examples, it is possible to infer that there are essential reasons for justifying the investment of time, money and effort required to develop a successful bridge to knowledge and technology. In addition, the problems or needs that belong to an Informally Structured Domain must be solved by a solution that will come to innovate, either because it modifies the procedure to address the problem or by itself is a new solution. So, this innovative solution will be cognitive because, in order to obtain it, it must have extracted persistent knowledge; or, of the existing cognitive process or that the solution by itself is of the scope of the Artificial Intelligence, the Machine Learning or Cognitive Computation. There will be occasions when the problem or need can be addressed without further ado by a product or tool of Artificial Intelligence, Machine Learning or Cognitive Computing; or, in most cases, the cognitive solutions will be tailored to the situation of problem or need. Trivializing tasks of modeling and analyzing the problems derived from "time pressure" or the well-known phrase "we no longer have time" can translate into a real loss of time, making bad decisions and opportunities that are going away or that will never come. Who can identify the right type of solution for each real situation that arises? In Cognitive Architecture: Designing for How We Respond to the Built Environment, Ann Sussman and Justin B. Hollander review the novel trends (2014) in psychology and neuroscience to support architects and planners, of the current world of construction, to better understand the clients, like the sophisticated mammals they are, and in response to a constantly evolving environment. In particular, they describe four main principles that observe a relationship with the cognitive processes of the human being: people are a thigmotactic species, that is, they respond to the touch or surface of external contact; visual orientation; preference for bilateral symmetrical forms; and finally, narrative inclinations, unique to the human being. The authors emphasize that the more we understand human behaviour, the better we can design for it and suggest the obligation to carry out activities of analysis, "preparation of the cognitive scaffolding", before carrying out construction and anticipating the future experience of the client [10]. Similarly, Portillo-Pizaña et al. suggest the importance of considering four stages for the implementation of a process of conscious innovation in an organization: consciousness, choice, action, and evolution. This conscious process of corporate innovation initially implies that every human being who integrates an innovation effort within an organization understands that any process of change or transformation begins with a state of consciousness where the existing gaps between the current situation are identified and the desired situation under the perspective of a user; subsequently, a decision-making process must be faced that allows an agile iteration, in order to move on to a real commitment to innovation and entrepreneurship—with all the necessary characteristics required of a true entrepreneur—and conclude then with a disposition to agile evolution, without sticking to ideas that are not well received by the market [11]. Thus, to identify the right kind of cognitive solution in a real situation it is highly convenient to have a cognitive architect. A Cg.Ar is a role with multidisciplinary knowledge in areas that should be treated as they are: Artificial Intelligence, Machine Learning, Cognitive Computing, logic, cognitive processes, psychology, sociology, philosophy to mention a few.
Cognitive Innovation Archetype for Smart Cities Applicaitons 103
5.1 Why the need for a cognitive architect? In the Cognitive Industry, development projects usually require the knowledge and understanding of psychology, artificial intelligence, computer science, neuroscience, philosophy, linguistics, anthropology. These are important disciplines that deal with the process of creating the cognitive basements and structures on which the new schemes of how to go about thinking of innovative products, solutions and things are based. These disciplines have critical functions that are essential in a cognitive architecture job and they rely on one-another to accomplish a given cognitive task. Currently, many people believe that there is no difference between Computer's Science or Systems and Cg.Ar and, even though there are similarities in some subjects, there are clear differences between them. As a matter of fact, they have well-defined roles that make them distinguishable from each other. Both Computer's Science or Systems and Cg.Ar are involved in programming and designing apps. However, Cg.Ar focuses more on planning and designing cognitive structures, elements and collation of the development work and is more concerned with the knowledge elicitation, management, modeling and functionality of the design, making certain that those structures can support normal and extreme cognitive demands. Even though computer science or systems engineers are involved in the design process of software solutions, architects take the lead role in terms of the design of the structure. The Cg.Ar will initiate and create the design, including the Knowledge Requirements, Cognitive Modelling and processes of the development work, then computers science or systems engineers professionals, when Cognitive Solutions will be software, will analyse it to find ways to make the software design possible. The computer engineers could be responsible in finding suitable intelligent algorithms, suggesting modifications and adjustments and evaluating the structural integrity to transform the cognitive architect's vision into realization. To summarize, cognitive architecture's primary concern is making very good models from cognitive blueprints and designing the development work while the computer engineering's responsibility is ensuring that everything that is foreseen in the cognitive blueprints can be implemented functionally and reliably. Computer Engineers and Cognitive Architects may sometimes overlap each other's work but a good relationship between the two professions will make the cognitive software solution job more effective and successful. Today, computer engineers make a point to work harmoniously with Cg.Ar to ensure superior quality results and proper design implementations for all stakeholders because they understand that teamwork and cooperation are vital to the success of any cognitive project. Finally, to highlight the work of the Cg.Ar in three words it can be said he is an orchestrator of innovation (see subsection §3.5). 5.2 Why the need to establish a model to support a cognitive architecture? When the CgI-SAP team faces a problem and must design and implement a CgI-S, it must have to appropriate a reality, which, in initial conditions, overcomes their capacity and comprehension. Therefore, it is very convenient to have a model that simplifies this reality: CgI-M. This model will be as detailed as it is necessary to offer a global vision of the CgI-S environment. Thus CgI-M allows a better understanding of the ISD to which CgI-S belongs with the aim of: • visualize what is or how it should be a solution; • establishes a guide to specify the structure or behaviour of the domain in relation to an implementation of a possible solution; and • document the decisions and actions that are carried out. CgI-M is useful for solving problems with both simple and complex solutions. As a problem becomes more and more complex, it's possible solution will also be, therefore, the CgI-M model
104 Innovative Applications in Smart Cities becomes more important for a simple reason: it seems that when a model is proposed for complex situations, for example, to build the scaffolding of a cognitive architecture, it is due to our inability to deal with complexity in its entirety. In the meantime, CgI-M reduces the complexity of what is being addressed, focusing on only one aspect at a time. It is important to highlight that CgI-M intends to formalize, because if informality were allowed, it would generate products that could not clearly address the domain of the solution to be implemented. In conclusion, and in spite of some solution providers, the more complex the domain and the problem to be addressed, the more it is imperative to use a model. Solution providers are already faced with situations where they have developed simple CgI-S without starting from any model and, after a short time, notice that the domain grew in complexity nullifying the effectiveness of the solution to the detriment of the quality of its service and the loss of the client. Finally, the CgI-M model after being used in real cases that its components as a whole can, de facto, respond through a cognitive collision to situations that occur within the domains of informal structure. There is a lot of work to be done on the subject of obtaining and representing commonsense information; for existing frames of representation must evolve and be integrated with other frameworks in order to enhance representation and, consequently, reasoning with common sense information. In general, the results obtained by CgI-M suggest that the knowledge obtained from it is highly congruent with that expressed by ahCN when validated by the client and the results of the solutions provided by it. However, it is also clear that it is not possible to explain the complete cognitive process of ahCN exclusively in the current terms of the CgI-M model. Consequently, the model is open and dynamic for the improvement of its components and to better explain the harmonization and integration of different types of cognitive processes that are supposed to coexist in a perspective of heterogeneous representation, for which additional research and collaboration among those we approach is needed. In particular, in our opinion, such improvements should be oriented to (i) in which cases the components of the CgI-M model play a more relevant role in establishing the scaffolding necessary to develop a particular cognitive solution (ii) or cases where they are not at all evoked by a cognitive system, since the need to react in real time is more urgent and, therefore, (iii) accelerate the activities proposed by the model. Since there is no clear answer to such questioning, these aspects will imply, in our opinion and in congruence with [12], the future research agenda of cognitive psychology and the cognitive—artificial—systems research.
6. Conclusions and Future Challenges This paper communicates the convenience of walking in the direction of an archetype that characterizes the essential aspects of Cognitive Architecture, namely, what elements make up the ahCN and how they interact with each other, the cognitive architecture and the activities, tasks that lead out the Cognitive Architect. It was argued that, based on the results of client studies, these aspects should be addressed to formalize and accelerate the establishment of Cognitive Architecture with the limitations and challenges that require the daily tasks of a cognitive process. Such challenges, from a technological perspective, are crucial to being addressed in order to be able to operate cognitive solutions and make decisions in general scenarios exploiting a plethora of integrated reasoning mechanisms. Based on these assumptions, we confirm the convenience of integrating a model to deal, jointly, with the aspects mentioned above. Finally, there are already several crucial problems of real situations that have been addressed by our model, of which one of them was mentioned, where the cognitive processes are harmonized in the CgI-M, interacting with an ahCN, and reflected in a cognitive architecture that supports to the CgI-S implemented by the Cg.Ar. The results obtained suggest that, although the systematic
Cognitive Innovation Archetype for Smart Cities Applicaitons 105
process for knowledge management KMoS-REload provided by the CgI-M represents an adequate way to integrate different knowledge acquisition and representation mechanisms, it is still not clear if they are sufficient and robust. Therefore, it is still an open question of what and what kind of processes, techniques or elements should be part of a general architectural mechanism and if it is worth implementing them in the processes of the model to operate their conceptual structures. As mentioned above, answers to questions or efforts will require a joint research effort on the part of cognitive psychology and the community of cognitive models and processes, cognitive computation, machine learning, and artificial intelligence.
This chapter provides a summarized, critical and analytical point of view of the data-centric solutions that are currently applied for addressing urban problems in cities. These solutions lead to the use of urban computing techniques to address their daily life issues. Data-centric solutions have become popular due to the emergence of data science. The chapter describes and discusses the types of urban challenges and how data science in urban computing can face them. Current solutions address a spectrum that goes from data harvesting techniques to decision making support. Finally, the chapter also puts in perspective families of strategies developed in the state of the art for addressing urban problems and exhibits guidelines that can lead to a methodological understanding of these strategies.
1. Introduction The development of digital technologies in the different disciplines, in which cities operate, either directly or indirectly, is altering expectations among those in charge of the local administration. Every city is a complex ecosystem with subsystems to make it work such as work, food, clothes, residence, offices, entertainment, transport, water, energy, etc. With the growth of cities, there is more chaos and most decisions are politicized, there are no common standards and data is overwhelming. The intelligence is sometimes digital, often analogue, and almost inevitably human. University Grenoble Alpes, CNRS, Grenoble INP, LIG, France. Universidad Nacional Autónoma de México, Mexico. 3 Fundación Universidad de las Américas Puebla, Mexico. 4 University of Lyon, LIRIS, France. 5 French Mexican Laboratory of Informatics and Automatic Control. Emails: [email protected], [email protected] * Corresponding author: [email protected] 1 2
108 Innovative Applications in Smart Cities Urban computing [36] is a world initiative leading to better exploit resources in a city to offer higher-level services to people. It is related to sensing the city's status and acting in new intelligent ways at different levels: people, government, cars, transport, communications, energy, buildings, neighbourhoods, resource storage, etc. A vision of the city of the "future", or even the city of the present, rests on the integration of science and technology through information systems. Data-centric solutions are in the core of urban computing that aims at understanding events and phenomena emerging in urban territories, predict their behaviour and then use these insights and foresight to make decisions. Data analytics and exploitation techniques are applied in different conditions and using ad hoc methodologies using data collections of different types. Today important urban computing centres in metropolises, have proposed and applied these techniques in these cities for studying real state, tourism, transport, energy, air, happiness, security and wellbeing. The adopted strategies have to do with the type of context in which they work. This chapter provides a summarized, critical and analytical point of view of the data-centric solutions that are currently applied for addressing urban problems in cities leading the use of urban computing techniques to address their daily life issues. The chapter puts in perspective families of strategies developed in the state of the art for addressing given urban problems and exhibits guidelines that can lead to a methodological understanding of these strategies. Current solutions address a spectrum that goes from data harvesting techniques to decision making support. The chapter describes them and discusses their main characteristics. Accordingly, the chapter is organised as follows. Section 2 characterises urban data and introduces data harvesting techniques used for collecting urban data. Section 3 discusses approaches and strategies for indexing urban data. Section 4 describes urban data querying. Section 5 summarizes data and knowledge fusion techniques. Finally, Section 6 discusses the research and applied perspectives of urban computing.
2. Data Harvesting Techniques in Urban Computing Urban computing is an interdisciplinary field which concerns the study and application of computing technology in urban areas. A new research opportunity emerges in the database domain for providing methodologies, algorithms and systems to support data processing and analytics processes for dealing with urban computing. These processes involve harvesting data about the urban environment to help improve the quality of life for people in urban territories, like cities. In this context, academic and industrial contributions have proposed solutions for building networks of data, retrieving, analysing and visualizing them for fulfilling analytics requirements stemming from urban computing studies and projects. Urban data processing is done using: (i) continuously harvested observations of the geographical position of individuals (that accept sharing their position) over time; (ii) collections of images stemming from cameras observing specific "critical" urban areas, like terminals, airports, public places and government offices; (iii) data produced by social networks and applications like Twitter, Facebook, Waze and similar. Independently of the harvesting strategies and processing purposes, it is important to first characterise urban data. This is done in the next section. 2.1 Urban data Urban data can be characterized concerning three properties: time, space and objects (occupying urban territories). They are elementary properties that can guide the way urban data can be harvested and then processed for understanding urban phenomena. For urban data, time must be considered from two perspectives, as its mathematical definition as a continuous or discrete linearly ordered set consisting of time instants or time intervals, called time units [3]. But, also under a cyclic perspective to consider iterations of seasons, weeks and days. Regarding space, it can be represented by [3] different referencing models: coordinate-based models with tuples of numbers representing the
From Data Harvesting to Querying for Making Urban Territories Smart 109
distance to certain reference points or axes, division-based models using a geometric or semanticbased division of space, and linear models with relative positions along with linear reference elements, such as streets, rivers and trajectories. Finally, the third urban data property, object, refers to physical and abstract entities having a certain position in space (e.g., vehicles, persons and facilities), temporal properties, for objects existing in a certain period (i.e., event), and spatiotemporal properties, which are objects with a specific position in both space and time. Besides time, space and object properties, Yixian Zheng et al. [25] identify six types of data that can be harvested and represent the types of entities that can be observed within urban territories according to the urban context they refer to, i.e., human mobility, social network, geographical, environmental, health care and divers. Human mobility data enables the study of social and community dynamics based on different data sources like traffic, commuting media, mobile devices and geotagged social media data. Traffic data is produced by sensors installed in vehicles or specific spots around the city (e.g., loop sensors, cameras). These data can include vehicles' positions observed recurrently at given intervals. Using these points (positions), it is then possible to compute trajectories which are spatiotemporally time-stamped and can be associated with instant speed and heading directions. Traffic occupation inroads can be measured with loops that compute, within given time intervals, which vehicles travel across two consecutive loops. Using this information, it is possible to compute travel speed and traffic volume on roads. Ground truth traffic conditions are observed using surveillance cameras that generate a huge volume of images and videos. Extracting information such as traffic volume and flowrate from these images and videos is still challenging. Therefore, in general, these data only provide a way to monitor citywide traffic conditions manually. People's regular movement data are produced by personalized RFID transportation cards for buses or metro that they tap in station entries to enter/exit the public transportation system. This generates a huge amount of records of passenger trips, where each record includes an anonymous card ID, tap-in/out stops, time, fares for this trip and transportation type (i.e., bus or metro). Commuting data recording people's regular movement in cities can be used to improve public transportation and to analyze citywide human mobility patterns. Records of exchanges like phone calls, messages, internet, between mobile phones and cell stations collected by telecom operators are data that contain communication information, people's locations based on cell stations. These data offer unprecedented information to study human mobility. Social Networks Data. Social networks posts (e.g., blogs, tweets) are tagged with geo-information that can help to better understand people's activities, the relations among people and the social structure of specific communities. User-generated texts, photos and videos, contain rich information about people's interests and characteristics, that can be studied from a social perspective. For example, evolving public attention on topics and spreading of anomalous information. The major challenges with geo-tagged social network data lie in their sparsity and uncertainty. Finally, data refer to points of interest (POI) to depict information of facilities, such as restaurants, shopping malls, parks, airports, schools and hospitals in urban spaces. Each facility is usually described by a name, address, category and a set of geographical coordinates. Environmental data. Modern urbanization based on technology has led to environmental problems related to energy consumption and pollution. Data can be produced by monitoring systems observing the environment through different variables and observations (e.g., temperature, humidity, sunshine duration and weather conditions), air pollution data, water quality data and satellite remote sensing data, electricity and energy consumption, CO2 footprints, gas. These data can help to provide insight regarding consumption patterns, on correlations among actions and implications and foresight about the environment.
Energy City's energy consumption: obtained directly from sensors or inferred from data sources implicitly, e.g. from the GPS trajectory of a vehicle
Figure 1: Urban Data Types.
Urban data can be harvested from different sources and using different techniques. These aspects are discussed next. 2.2 Data harvesting techniques Data acquisition techniques can unobtrusively and continually collect data on a citywide scale. Data harvesting is a non-trivial problem, given the three aspects to consider: (i) energy consumption and privacy, (ii) loose-controlled and non-uniform distributed sensors, (iii) unstructured, implicit, and noisy data. Crowdsensing. The term "crowdsourcing" is defined as the practice of obtaining needed services or content by soliciting contributions from a large group of people. People play the role of urban data consumers, but also participate in the data analysis process through crowdsourcing. Techniques use explicit and implicit crowdsourcing for collecting data that contain information about the way people evolve in public and private places. These data collections can be used as input for learning crowd behaviour and simulating it more accurately and realistically. The advances of location-acquisition technologies like GPS and Wi-Fi have enabled people to record their location history with a sequence of time-stamped locations, called trajectories. Regarding non-obstructive data harvesting, work has been carried out using cellular networks for user tracking, profiting from call delivery that uses transitions between wireless cells. Geolife1 is a
From Data Harvesting to Querying for Making Urban Territories Smart 111
social networking service which aims to understand trajectories, locations and users, and mine the correlation between users and locations in terms of user-generated GPS trajectories. In [17] a new vision has been proposed regarding the smart cities' movement, under the hypothesis that there is the need to study how people psychologically perceive the urban environment, and to capture that quantitatively. Happy Maps uses crowdsourcing and geo-tagged pictures and the associated metadata to build alternative cartography of a city weighted for human emotions. People are more likely to take pictures of historical buildings, distinctive spots and pleasant streets instead of carinfested main roads. On top of that, Happy Maps adopts a routing algorithm that suggests a path between two locations that is the shortest route and maximizes the emotional gain. 2.3 Discussion and synthesis An important aspect to consider is that data is non-uniformly distributed in geographical and temporal spaces, and it is not always harvested homogeneously according to the technique and the conditions of the observed entities in an urban territory. Having the entire dataset may be always infeasible in an urban computing system. Some information is transferrable from the partial data to the entire dataset, for example, the travel speed of taxis on roads can be transferred to other vehicles that are also travelling on the same road segment. Some information cannot be transferable, for example, the traffic volume of taxis on a road may be different from private vehicles. In some locations, when crowdsensing is used, more data can be harvested as required and in other places fewer data than required. In the first case, a down-sampling method, e.g., compressive sensing, could be useful to reduce a system's communication loads. In the last case, in the context of crowdsensing, some incentives that can motivate users to contribute data should be considered. How to configure the incentive for different locations and periods to maximize the quality of the received data (e.g., the coverage or accuracy) for a specific application is yet to explore. Three types of strategies can be adopted for harvesting data. (i) Traditional sensing and measurement that implies installing sensors dedicated to some applications. (ii) Passive crowdsensing using wireless cellular networks built for mobile communication between individuals to sense city dynamics (e.g., predict traffic conditions and improve urban planning). We described how this technique can be specialised into three strategies: • Sensing City Dynamics with GPS-Equipped Vehicles: mobile sensors continually probing the traffic flow on road surfaces processed by infrastructures that produce data representing citywide human mobility patterns. • Ticketing Systems of Public Transportation (e.g., model the city-wide human mobility using transaction records of RFID-based cards swiping). • Wireless Communication Systems (e.g., call detailed records CDR). • Social Networking Services (e.g., geotagged posts/photos, posts on natural disasters analysed for detecting anomalous events and mobility patterns in the city). (iii) Participatory sensing where people obtain information around them and contribute to formulating collective knowledge to solve a problem (i.e., human as a sensor): • Human crowdsensing: users willingly sense information gathered from sensors embedded in their own devices (e.g., GPS data from a user's mobile phone used to estimate real-time bus arrivals). • Human crowdsourcing: users are proactively engaged in the act of generating data: reports on accidents, police traps, or any other road hazard (e.g., Waze), citizens turning into cartographers to create open maps of their cities.
112 Innovative Applications in Smart Cities
3. Managing and Indexing Urban Data The objective of managing and indexing urban data is to harness a variety of heterogeneous data to quickly answer users' instant queries, e.g., predicting traffic conditions and forecasting air pollution. Three problems are addressed in this context: stream and trajectory data management, graph data management and hybrid indexing structures. 3.1 Stream and trajectory data management Urban data, often collected recurrently or even continuously (velocity), can lead to huge volumes of data collections that should be archived, organized (indexed) and maintained on persistence supports with efficient associated read and write mechanisms. Indexing and compression techniques are often applied to deal with data velocity and volume properties. The continuous movement of an object is recorded in an approximate form as discrete samples of location points. A high sampling rate of location points generates accurate trajectories but will result in a massive amount of data, leading to enormous overhead in data storage, communications, and processing. Thus, it is necessary to design data reduction techniques that compress the size of a trajectory while maintaining the utility of the trajectory. There are two major types of data reduction techniques running in batch after the data is collected (e.g., Douglas-Peucker algorithm [7]) or in an online mode as the data is being collected (such as the sliding window algorithm [12,16]). Trajectory reduction techniques are evaluated concerning three metrics: processing time, compression rate, and error measure (i.e., the deviation of an approximate trajectory from its original presentation). Recent research [18], has proposed solutions to the trajectory reduction through a hybrid spatial compression algorithm and error-bounded temporal compression algorithm. Chen et al. [5] propose to simplify a trajectory by considering both the shape skeleton and the semantic meanings of the trajectory [31,32]. For example, when exploring a trajectory (e.g., travel route) shared by a user, the places where she stayed, took photos, changed moving directions significantly would be more significant than other points. Consequently, points with an important semantic meaning should be given a higher weight when choosing representative points for a simplified trajectory. 3.2 Graph data management Graphs are used to represent urban data, such as road networks, subway systems, social networks, and sensor networks. Graphs are usually associated with a spatial property, resulting in many spatial graphs [36]. For example, the node of a road network has a spatial coordinate and each edge denoting a road segment has a spatial length. Graphs also contain temporal information; for instance, the traffic volume traversing a road segment changes over time, and the travel time between two landmarks is time-dependent: spatio-temporal graphs [36]. Queries like "find the top-k tourist attractions around a user that are most popular in the past three months", can be asked on top of graphs. Hybrid Indexing Structures are intended to organize different data sources; for example, combining POIs, road networks, traffic, and human mobility data simultaneously. Hybrid structures can be used for indexing special regions; for instance, a city partitioned into grids by using a quadtree-based spatial index (see Figure 2) where each leaf node (grid) of the spatial index maintains two lists storing the POIs and road segments. Then, each road segment ID points to two sorted lists: a list of taxi IDs sorted by their arrival time ta at the road segment, and a list of drop-off and pick-up points of passengers sorted by the pick-up time (tp) and drop-off time (td). Different kinds of index structures have been proposed to manage different types of data individually. Hybrid indexes can simultaneously manage multiple types of data (e.g., spatial, temporal, and social media) and enable the efficient and effective learning of multiple heterogeneous data sources. In an urban computing scenario, it is usually necessary to harness a variety of data and integrate them into a data-mining model. This calls for hybrid indexing structures that can organize
From Data Harvesting to Querying for Making Urban Territories Smart 113
Figure 2: Hybrid index for organizing urban data [36].
different data sources, like hybrid indexing structure, which combines a spatial index, hash tables, sorted lists, and an adjacency list.
4. Querying Urban Data Querying the actual location of a moving object has been studied extensively in moving object databases using 3DR-Tree [19] and MR-Tree [28]. Yet, sometimes queries must explore historical trajectories satisfying certain criteria, for example, retrieving the trajectories of tourists passing a given region and within a period. This corresponds to a spatiotemporal range query [23,24]), for example, taxi trajectories that pass a crossroad (i.e., a point query), or the trajectories that are similar to a query trajectory [6,20] (i.e., a trajectory query). Dealing with the uncertainty of a trajectory refers to positioning moving objects while their locations can only be updated at discrete times. The location of a moving object between two updates is uncertain because the time interval between two updates can exceed several minutes or hours. This can, however, save energy consumption and communication bandwidth. Map matching is to infer the path that a moving object like a vehicle has traversed on a road network based on the sampled trajectory. Map-matching techniques dealing with high-samplingrate trajectories have already been commercialized in personal navigation devices, while those for low-sampling-rate trajectories [15] are still considered challenging. According to Yuan et al. [30], given a trajectory with a sampling rate around 2 minutes per point, the highest accuracy of a mapmatching algorithm is about 70%. When the time interval between consecutive sampling points becomes even longer, existing map-matching algorithms do not work very well any more [36]. Wei et al. [26] proposed to construct the most likely route passing a few sampled points based on many uncertain trajectories. Krumm et al. and Xue et al. [13,29] propose solutions to predict a user's destination based on partial trajectories. More generally, a user's and other people's historical trajectories as well as other information, such as the land use of a location, can be used in destination prediction models. Other important problems include observing a certain number of moving objects travelling a common sequence of locations in similar travel time where the locations in a travel sequence are not consecutive for finding sequential patterns from trajectories. Other approaches discover a group of objects that move together for a certain time period, under different patterns such as flock
114 Innovative Applications in Smart Cities [8,9], convoy [10,11], swarm [14], traveling companion [21,22], and gathering [34,35,36,25]. These "group patterns" can be distinguished based on how the "group" is defined and whether they require the periods to be consecutive. For example, a flock is a group of objects that travel together within a disc of some user-specified size for at least k consecutive timestamps [10]. Li et al. [14] relaxed strict requirements on consecutive periods and proposed the pattern swarm, which is a cluster of objects lasting for at least k (possibly non-consecutive) timestamps.
5. Data and Knowledge Fusion In urban computing scenarios, it is necessary to exploit a variety of heterogeneous data sources that need to be integrated. Then, it is necessary to fusion knowledge to explore and exploit datasets to extract insight and foresight of urban patterns and phenomena. Data fusion. There are three major ways to achieve this goal: • Fuse data sources at a feature level putting together the features extracted from different data sources into one feature vector. Beforehand, and given the heterogeneity of data sources, a certain kind of normalization technique should be applied to this feature vector before feeding it into a data analytics model. • Use different data at different stages. For instance, first partition an urban region, for example, a city, into disjoint regions by major roads and then use human mobility data to glean the problematic configuration of a city's transportation network [33]. • Feed different datasets into different parts of a model simultaneously given a deep understanding of the data sources and algorithms applied to analyse them. Building high-quality training datasets is one of the most difficult challenges of machine learning solutions in the real world. Disciplines like data mining, artificial intelligence and deep learning have contributed to building accurate models but, to do so, they require vastly larger volumes of training data. The traditional process for building a training dataset involves three tasks: data collection, data labelling and feature engineering. From the complexity standpoint, data collection is fundamentally trivial as most organizations understand what data sources they have. Feature engineering is getting to the point where it is 70%–80% automated using algorithms. The real effort is in the data labelling stage. New solutions are emerging for combining strong and weak supervision methods to address data labelling. Knowledge fusion. Data mining and machine-learning models dealing with a single data source have been well explored. However, the methodology that can learn mutually reinforced knowledge from multiple data sources is still missing. The fusion of knowledge does not mean simply putting together a collection of features extracted from different sources but also requires a deep understanding of each data source and the effective usage of different data sources in different parts of a computing framework. End-to-end urban computing scenarios call for the integration of algorithms of different domains. For instance, data management techniques with machine-learning algorithms must be combined to provide both efficient and effective knowledge discovery ability. Similarly, integrating spatio-temporal data management algorithms with optimization methods. Visualization techniques should be involved in a knowledge discovery process, working with machine-learning and datamining algorithms.
From Data Harvesting to Querying for Making Urban Territories Smart 115
6. Perspectives of the Role of Data Science for Making Urban Spaces Smart This chapter discussed and described issues regarding data for enabling urban computing tasks that can lead to the design of smart urban territories. Having a data-centred analysis of the problems and challenges introduced by urban computing exhibits the requirement to study data concerning different perspectives. First, the chapter characterised data produced within urban territories in terms of their mathematical properties (spatio-temporal), concerning the "semantics" of the entities composing urban territories (e.g., points of interest, roads, infrastructure) and also from the mobile entities that populate urban territories, like people, vehicles and the built environment. This variety of data is produced by producers with different characteristics, and approaches today use hardware, software and passive and active participation of people to generate phenomenological observations of urban territories. Finally, the chapter discusses how to create insight and foresight of important situations happening in urban territories, for example, computing trajectories of entities evolving in these territories observed in space and time, and other social foresight of behaviours like popular POIs, the population of regions, etc. The vision of urban computing—acquisition, integration, and analysis of big data to improve urban systems and life quality— is leading to smarter cities. Urban computing blurs the boundary between databases, machine learning, and visualization and even bridges the gap between different disciplines (e.g., computer sciences and civil engineering). To revolutionize urban sciences and progress, quite a few techniques still need to be explored, such as the hybrid indexing structure for multimode data, the knowledge fusion across heterogeneous data sources, exploratory visualization for urban data, the integration of algorithms of different domains, and intervention-based analysis.
Utilization of Detection Tools in a Human Avalanche that Occurred in a Rugby Stadium, Using Multi-Agent Systems Tomás Limones,1,* Carmen Reaiche2 and Alberto Ochoa-Zezzatti1
This article aims to make a simulation model of an avalanche that occurred at a Rugby football match due to the panic caused by riots between fanatical fans of the teams that were playing. To carry out this model, the specific Menge simulation tool is used, which helps us to evaluate the behavior of people who consciously or unconsciously affect the contingency procedures established at the place of the event, to define them preventively to reduce deaths and injuries. From the definition of the factors, an algorithm is developed from the combination of the Dijkstra tool and the simulation tool that allows us to find the route to the nearest emergency exit, as well as the number of people who could transit safely. Additionally, Voroni diagrams are used to define perimeter adjacency between people.
1. Introduction Thousands of deaths have happened in different parts of the world where football is like a religion. The very serious disturbances that occurred after a football match, avalanches caused by panic, riots between fanatical fans, landslides in poor condition, overcapacity, are just a few examples of events that generate deaths in the stadiums. The tragedies have been numerous, and the main causes occur when people enter a panic, which unfortunately causes an imbalance in their thinking, failing to have control over their actions and causing the agglomerations with catastrophic consequences. Some historical events with the greatest consequence in deaths during football games are described in Table 1: As can be seen on the Table 1, most of the eventualities presented here have their origin in the disturbances incited by the same fans, causing stampedes wherein, due to closed doors, people become pressed against bars or meshes causing human loss by severe blows and asphyxiation. This type of agglomeration is not exclusive to football. In the article "Innovative data visualization of collisions in a human stampede occurred in a religious event using multiagent systems" [1], the author analyzes this type of phenomenon, but focused on religious events, where large concentrations of people come together. In this example, an analysis is made about the tragedy that occurred in Mecca in 2015, where 2717 people died and 863 were injured as a result of the largest human stampede ever recorded. Universidad Autónoma de Ciudad Juárez, México. The University of Adelaide, Australia. * Corresponding author: [email protected] 1 2
Excess capacity causes avalanche. It did not meet the security requirements
7
Mateo Flores, Guatemala Stadium
1996
Over sale due fake tickets and closed stadium doors (The doors opened inwards)
200
83
8
Said Port Stadium
2012
Caused by fans who attacked players and fans with weapons
1000
74
96
Figure 1: Representative graph of injuries and deaths in the history of soccer.
In the case of this study, the simulation exercise will be carried out in the Rugby Stadium of the Australia Adelaide City, known as Oval Stadium. Its characteristics are described below: NAME
The Adelaide Oval
DESCRIPTION
It is a multipurpose stadium located in the city of Adelaide, Australia. It is mainly used for the practice of cricket and Australian rules football, as well as soccer and rugby
DIRECTION
War Memorial Dr, North Adelaide SA 5006
CAPACITY
53,500
OPERATION SINCE
1871
PROPERTY
Government of South Australia
The city of Adelaide is in southern Australia and is characterized as a peaceful city, where eventualities due to fighting or aggressions are unusual. Historically, there has been a fight raised on August 25, 2018, where two fans in a Rugby match between Port Adelaide and Essendon AFL started a fight. The fans themselves tried to intervene to avoid this quarrel. It is noted that the
Detection Tools in a Human Avalanche 119
actions of these two individuals was an isolated element among a crowd of more than 39,000 fans. The realization of this exercise will be carried out simulating an avalanche in the Oval stadium, provoked by the panic caused by riots among fanatical fans. The result will help us to define the best alternatives of preventive solutions to avoid possible catastrophes. The Figure 2, shows a distribution graph of the Oval Adelaide stadium.
Figure 2: Oval Adelaide stadium of rugby football.
The anthopometry Anthropometry is considered as the science in charge of studying the physical characteristics and functions of the human body, including linear dimensions, weight, volume, movements, etc., in order to establish differences between individuals, groups and races [2]. This science turns out to be a guideline in the design of the objects and spaces necessary for the environment of the human body and that, therefore, must be determined by their dimensions [3]. By knowing these data, the minimum spaces that human needs to function daily are known, which must be considered in the design of his environment. Some factors that define the physical complexion of the human being are race, sex, diet, age. The reference plane distributes 3 imaginary flat surfaces that cross the body parts and are used as a reference in taking body dimensions (See Figure 3). Sports fans have seen the evolution and development of professional players and how it has been shocking in recent years. A rugby defender of 80 kg, or 160 pounds, which was previously considered enough, now looks less heavy with not enough weight to take the job of a restorer. Dimensional standards and spatial requirements must be constantly adequate. The need to establish standards that guarantee the adaptation of the interior spaces for sports practices to the human dimension and the dynamics of people on the move constitutes, today, a potential threat to the safety of the participants. The lack of this kind of regulation not only involves a serious threat to
120 Innovative Applications in Smart Cities
Figure 3: Reference plane.
the physical integrity of the users, but also makes the client and the designer potentially legally liable in the event of an accident with injury or death. The inference of the human body-interior space not only influences the comfort of the first but also in public safety. The size of the body is the fundamental measurement reference for dimensioning the width of doors, corridors and stairs in any environment, whether public or private. Every precaution is little in the use and acceptance of existing methods or empirical rules to establish critical clearances without questioning their anthropometric validity, even for those likely to be part of affected codes and ordinances. In short, certain dimensions and clearances that guarantee public safety must be defined. Public spaces must be designed so as not to hinder their use for people outside a standard, such as children, small people, overweight people. The designs of the different attachments and accessories will also have the reach of these people; the stairs, seats, hallways, open spaces among others. Horizontal space Two measures are important to consider in a space of the people movement: (1) Body dimensions and (2) larger people. Slacks should be considered for both measures Figure 4 shows two fundamental projections of the human body, which include the critical dimensions of the 95th percentile. A tolerance of 7.6 cm (3 inches) has been included for width and depth. The final dimension with the tolerance included is 65.5 cm (28.8 inches); The critical anthropometric dimension to be used during a massive agglomeration is the body width. The diagram representing the body ellipse and the lower Table 2 have proven utility in the design of circulation spaces. The latter is an adaptation of a study of the movement and formation of pedestrian queues, prepared by Dr. John Fruin, whose purpose was to set the relative levels of service based on the density of pedestrians. The basic unit is the human body, which is associated with an elliptical shape or ellipse body of 45.6 x 61 cm (18 x 24 inches). The panic Panic attacks, also known as crisis of distress, are usually accompanied by various manifestations of somatic nature, such as tachycardia, sweating, tremor, choking sensation, chest tightness, nausea,
Detection Tools in a Human Avalanche 121
Figure 4: Two fundamental projections of the human figure. Table 2: Analysis of the circulation space for the human being "density of queues". Density Analysis In "Queues" Denomination
Description
A-Contact zone
Ratio
Surface
Inches
cm
Ft2
cm2
In this area of occupation, body contact is almost inevitable; circulation impossible, movement reduced when walking, shuffling; occupation like an elevator.
12
30.5
3
0.25
B-Non-contact zone
While it is not necessary to move, body contact can be avoided; possible movement in group form.
18
45.7
7
0.65
C-Personal zone
The depth of the body separates people; limited lateral circulation by passing people; This area is in the selected space occupation category, experiencing comfort standards.
21
53.3
10
0.95
D-Circulation zone
It is possible to move in «queue» without disturbing other people.
24
61
13
1.4
dizziness, fainting, hot flashes, feeling of unreality and loss of control [4]. This can happen when the person experiences the sensation of being near imminent death and has an imperative need to escape from a feared place or situation (aspect congruent with the emotion that the subject is feeling in the perceived imminent danger). The fact of not being able to physically escape the situation of extreme fear in which the affected person is greatly accentuates the symptoms of panic [5]. Taking this in consideration, the relationship with possible triggers of panic attacks can be classified as: • Unexpected. The beginning of the episode does not match manifest triggers.
122 Innovative Applications in Smart Cities • Situationally determined. Attacks occur in the presence or anticipation of a specific stimulus or situation. • Situationally predisposed. Episodes are more common in specific situations, although they are not completely associated with them. Panic attacks can originate from different situations, especially in those capable of generating a state of high physiological activation or in the event of a specific stress event. The panic attack is linked to agoraphobia, characterized by an intense anxiety response to situations in which it is difficult to escape or get help [6]. Factors that cause a human stampede Most of the human stampedes have occurred during religious, sporting and musical events since they are the ones that gather more people. The most common causes occur when people nervously react in moments of panic, whose detonator is fear. This fear can be caused by a fire, an explosion, fear of a terrorist attack, etc. When people want to escape, people from behind push those in front, not knowing that those in the front are being crushed. This stacking thrust force occurs in both forms, vertically and horizontally. The vast majority of deaths are caused by compression asphyxiation and rarely by trampling. The degree of physical strength is the main ally to cling to life. That is why, in most cases, children, elderly people, and women are the most affected ones. For the Honduran human behavior specialist, Teodosio Mejía (2017), one of the reasons is that when people are in a crowd, they "lose their condition of being rational". Mass men place their ego to the collective ego "and that is criminal" because when human beings are frustrated, they begin to despair, and this causes bad decisions to be made [7,10]. Multiple agent systems A complex system can be defined as a system that presents a large number of interactive components whose aggregate activity is not derivable from the sums of the activity of individual (non-linear) components and typically exhibits hierarchy of self-organization under selective constraints [8]. Multiple agent-based simulations (MABS) offer the possibility of creating an artificial universe in which experiments representing individuals, their behaviors and their interactions can be performed by modeling and simulating complex systems of multiple levels of abstraction. To conceive a MABS (Figure 5), here is the multi-view approach proposed by Michel [9], which distinguishes four main
Figure 5: Example of different aspects of a multilevel simulation.
Detection Tools in a Human Avalanche 123
aspects in a MABS: (i) Agent behavior that deals with agent modeling of the deliberative process (their minds). (ii) The environment that defines the different physical objects in the simulated world (the situated environment and the physical body of the agents) and the endogenous dynamics of the environment. (iii) The programming that deals with the modeling of the passage of time and the definition of programming policies is used to execute the behaviors of the agents. (iv) The interaction focuses on the result of the modeling of the actions and interactions between agents at a given time. Our approach broadens these different perspectives to integrate related multilevel aspects. In the next chart (Table 3) a comparison between some different methodologies for the use of multiagent systems with Mathematical models, are showed:
2. Review of the Problem Table 1 breaks down a total of 8 events that have represented a greater impact on deaths and injuries in the history of football themselves that passed from 1964 to 2012. The main cause was death due to suffocation for the pressure exerted by the masses, caused by human avalanches. According to the analysis of the table the main reasons for these avalanches were: • • • •
Fan riots. Excess of stadium capacity. Aggression with a weapon. Fire
In most of the events, a factor that influenced these outcomes was the closure of the accesses, since people could not leave because the doors were closed to prevent the entry of people who did not pay for a ticket or because the doors were opened in the opposite direction to the flow of people. Considering the reasons described on all the bibliography reviewed, the following variables that affect the results of a human avalanche can be defined: 1. Anthropometry considering the definition of the minimum horizontal space necessary to ensure integrity between people (4 different areas of circulation). 2. The population in the event (Considered 2 groups according to the sample of the number of people involved (734 people and 170 people). 3. The distribution of spaces in the stadium (Corridors and emergency exits). Considering the different combinations of these three variables, the movement of people can be simulated using the multiple agent system. Distribution and spaces in the oval stadium The architectural design of the Rugby Stadium contemplates a good security system. Four sets of photographs on the architectural distributions of the oval stadium are shown below The stadium's design includes spacious areas to avoid the crowding of people, as well as security systems, fire areas and areas for people with disabilities. Photographs 1 represents the open spaces in the stadium, both at the entrance and internally. Photographs 2 represents the access stairs to the middle and upper part of the stadium, these being spacious. Photographs 3 represents the exits of the stadium, external, internal and escalators. For the realization of this simulation model, the specific area of the exits on the internal central part of the stadium will be contemplated (Photographs 4). These exits are placed at the bottom of the stairs, almost at the level of the football field. A greater concentration of people is located in this area due to the seating arrangement. There are 23 rows arranged alphabetically. The probability that
Table 3: Some different methodologies for the use of multiagent systems.
124 Innovative Applications in Smart Cities
Detection Tools in a Human Avalanche 125
126 Innovative Applications in Smart Cities
Photographs 1: Open access spaces, Oval Stadium.
Photographs 2: Access stairs to the Oval stadium and back side of the open stadium.
these two points are a focus of attention for a possible crash problem or agglomeration of people is greater than in the rest of the exits.
3. Methodological Approximation For the development of this simulation exercise, two pedestrian equations developed in a study by Ochoa et al. (2019) "Innovative data visualization of collisions in a human stampede
Detection Tools in a Human Avalanche 127
occurred in a religious event using multiagent systems" will be used as a reference, where it is considered a catastrophic incident with critical levels of concentration of people at a religious event in Mecca. These equations will be used to simulate the movement of people within a stampede and determine the probability of their survival. The equation is based on the BDI methodology, which involves 3 fundamental factors that define the result: (1) Desires, (2) Beliefs, (3) Intentions. Equipmet Equipment description used during the simulation trials: Machine name: DESKTOP-G07PBE6 Operating System: Windows 10 Pro, 64-bits Language: Spanish (Regional Setting: Spanish) System Manufacturer: Lenovo System Model: ThinkPad Product ID: 8BF29E2C-5A1A-4CA2-92E8-BE228436613D Processor: Intel (R) Core (TM) i5-2520M CPU @2.50 GHz. Memory: 10.0 GB Available OS Memory: 9.89 GB usable RAM Disk Unit ID: ST320LT007-9ZV142 Hard Disk: 296 GB Page File: 243 GB used; 53.5 GB available Windows Dir: C:\WINDOWS DirectX Version: 10.0.17134.1 Software Menge: A framework for modular pedestrian simulation for research and development, free code. Unity: A multiplatform video game engine created by Unity Technologies; a free personal account was used. Sublime Text 3: A sophisticated text editor to encode, used a free trial evaluation. Git Bash: An emulator used to run Git from command line, a code software Layout definition Taking into consideration the exit door shown in Photo no. 4, where the width of the space of the tunnel that leads to the exit has a dimension of 2.4 meters, begins with the formalization of the layout. For the realization of the layout, the set of seats located on the left side and right side of the tunnel will be considered, making an initial group of 304 people who could leave through this exit door. The initial distribution is as follows: 1. Number of people located on the left side of the exit tunnel: 80 people. 2. Number of people located above the exit tunnel in three groups: 8, 32, 8 people. 3. Number of people located on the right side of the exit tunnel: 80 people. 4. Number of people located under the exit tunnel in two group: 48 people each. The distribution are 48 people on the left side and 48 people on the right side.
128 Innovative Applications in Smart Cities To make the distribution of the layout, the coordinates of the dimensions of the seats as well as stairs are considered, making a distribution of coordinates which have been handled in excel for it to define a preliminary space according to the following Figure 6:
Figure 6: Layout of the scenario exit taking in consideration 304 persons for this evacuation simulation.
Figure 7: Layout of the journey that is made by the 304 persons.
Detection Tools in a Human Avalanche 129
Performing the first run using the coordinates of the scenario defined, as well as using the pattern that people will follow during the evacuation, the image that is defined during the run of simulation [9] in menge shows, as a result, the following Figures 8 and 9.
Figure 8: First trial simulation on menge for the 304 people evacuation.
Figure 9: First trial simulation on menge for the 304 people evacuation (simulation advance).
Figure 9 shows how the agglomeration of agents causes a bottleneck at the entrance to the tunnel. This agglomeration is due to the narrow dimension of the roadway to the tunnel. The elements used during the development of this scenario run are shown in the following Table 4, achieving a total time of 762,994 seconds during the evacuation, a time considered very high for an evacuation process. Table 4: elements used during the first simulation evacuation trial. Common max_angle_vel= max_neighbors= obstacleSet= neighbor_dist= r = class= pref_speed= max_speed= max_accel= 90
10
1
5
0.19
1
1.34
2
50
Full frame
scene update
scene draw
buffer swap
simulation time
(avg): 976.929 ms in 762 laps
(avg): 927.179 ms in 763 laps
(avg): 47.0648 ms in 763 laps
(avg): 2.21781 ms in 763 lasp
762.994
4. Looking for Evacuation Time Improvement To improve the evacuation time defined in the first test run, an experiment should be carried out [11], making changes with the different elements that are used during the development of this run, using the menge simulation software, to optimize the time during the evacuation process. Table 5 shows the results from after the change elements, with the purpose to define the best condition during the simulation related to the elements that impact in the best time evacuation result.
Elements defined to be changed for simulation time improvement Performing experimentation tests by making changes to the elements of algorithms agents, like maximum speed angle, maximum neighbor, maximum neighbor distance, pre speed, maximum speed, and maximum acceleration, gives us, as a result, the definition of the best number to be considered during the performance of the simulation test using menge. Some algorithms agents perform better than others [12]. The elements that need to be changed in order to improve the simulation time are: 1. max_angle_ve = from 90 to 60 2. max_neighbors = from 10 to 2 3. neighbor_dist = from 5 to 3 4. pref_speed = from 1.34 to 1.5 5. max_speed = from 2 to 5 6. max_accel = from 50 to 80 7. the entrance to the tunnel dimension from 2.5 to 3.6 with the purpose of increasing the entrance of the exit tunnel. 8. increase the quantity of people to be involved during the evacuation simulation from 304 to 352 people. The new layout of the scenario for the evacuation simulation is considering improvements on the tunnel entrance, increasing the dimension related to this scenario. The new Layout is shown on the Figure 10. Performing the multiagent simulation trial with the new data (elements change), the simulation time result is improved, reaching 30.90 s (Table 6).
Detection Tools in a Human Avalanche 131
Figure 10: New improved layout considering the change in the tunnel entrance dimension. Table 6: Elements defined to be use for the best simulation time 30.9. Common max_angle_vel= max_neighbors= obstacleSet= neighbor_dist= r = class= pref_speed= max_speed= max_accel= 60
2
1
3
0.19
1
1.5
5
80
Full frame
scene update
scene draw
buffer swap
simulation time
(avg): 560.557 ms in 308 laps
(avg): 510.916 ms in 309 laps
(avg): 46.8925 ms in 309 laps
(avg): 2.29343 ms in 309 lasp
30,9001
Figure 11: Second trial simulation on menge for the 352 people evacuation.
Figure 12 shows the second run of simulation with the increase to 352 agents as well as the improvements included. It is possible to appreciate the increase in the dimension of the entrance of the tunnel. To facilitate and appreciate the movements of the agents, after including the improvements proposed during the run of the simulation, Figure 13 shows the agents that are separated in 7 groups and different colors, assigned to each one of them.
132 Innovative Applications in Smart Cities
Figure 12: Second trial simulation on menge for the 352 people evacuation (simulation advance).
Figure 13: Third trial simulation on menge for the 352 people evacuation defined on 7 groups.
Figure 14: Third trial simulation on menge for the 352 people evacuation for 7 groups (Tunnel view).
Figure 14 shows the increase in the number of agents, which rises from 304 to 352, separating into 7 groups and identifying them are different colors. This will allow us to see the improvement in terms of the decrease in the agglomeration of agents at the entrance of the tunnel. In Figure 15, can be appreciated and see the improvement in width dimension at the entrance of the tunnel, which greatly facilitates the exit of agents, avoiding collisions between them.
Detection Tools in a Human Avalanche 133
Figure 15: Third trial simulation on menge for the 352 people evacuation for 7 groups (simulation advance).
Figure 16: Third trial simulation on menge for the 352 people evacuation for 7 groups (Final simulation).
5. Conclusions The use of the menge tool for the development of this simulation exercise allows us to perform the simulation with different scenarios, considering the changes in the factors that impact on the outcome, facilitating the alternative evaluation, thereby seeking the preservation and safety of the agents involved. In this exercise, it was demonstrated that the changes of these elements during the development of the simulation allow us to have a clearer view of the potentially catastrophic results that may occur in a real eventuality. These results will give us indicators that can be determined for real decision making, which allows us to generate preventive actions. For future studies, it is necessary to continue with the development of runs considering changes in the elements that affect the behavior of the agents' travel, to improve the travel at higher speed, without agglomerations or generation of bottlenecks. The simulation must consider the improvements in the use of the software and databases currently available, such as UNITY, Visual Studio, among others, which will allow us to find and make proposals for the solution of potential problems of human avalanches.
Floods are natural disasters resulting from various factors, such as poor urban planning, deforestation and climate change, to provide just a couple of examples. The consequences of such disasters are often devastating and bring with them not only losses of millions of dollars but also of human lives. The purpose of this work is to offer a first approximation of people's reactions in the case of an evacuation due to a hypothetical flood in an area located in the Colonia Bellavista of Ciudad Juárez that is adjacent to the Río Bravo, the Acequia Madre, and the Díaz Ordaz viaduct, which are plausible to be subject to overflow or flooding after heavy torrential rains in a scenario where climate change has seriously affected the city's climate.
1. Introduction According to [1]"A flood is referred to when usually dry areas are invaded by water; there are two possible causes of why this type of disaster occurs, the first reason is related to natural phenomena such as torrential rains and rainy seasons, for the second cause there is talk of human actions that largely induce natural disasters; ...". Among the factors associated with human intervention are deforestation, elimination of wetlands, high CO2 emissions that cause climate variations [2,3], bad urban planning, etc. [1]. On the other hand, floods can be of two types according to [1]: sudden/ abrupt and progressive/slow. In addition, floods may occur in urban or rural areas. The environment of cities is greatly affected by climate change due to flooding [4]. The authors point out that, in general, public spaces do not adapt well to abrupt changes in the environment and that is why their design must be well worked out to avoid problems in the event of a disaster. One of the main problems affecting urban and rural populations is flooding. Table 1 shows the greatest floods in Europe during the 90's decade and their effects. The characteristics of Ciudad Juárez, as well as its climate, make it propitious to carry out a study referring to sudden floods, since these are characterized by the precipitation of a large volume of water in a short time, causing a rapid accumulation of water in conurbation areas; this because of the rupture of dams, torrential rains or overflowing of basins or rivers [1]. In addition, according to [2] an increase in torrential rainfall is expected that can cause the type of floods mentioned above is expected. In the case of Ciudad Juárez, this is characterized by the existence of the Río Bravo as well as a desert climate with torrential rains, which have caused severe flooding as in 2013 [6], in addition to the fact that the infrastructure and urban planning of the city are additional also factors Universidad Autónoma de Ciudad Juárez, Av. Hermanos Escobar, Omega, 32410 Cd Juárez, Chihuahua, México. * Corresponding author: [email protected]
136 Innovative Applications in Smart Cities Table 1: Heavy Floods in the EU and Neighboring Countries, 1991–2000, and Their Effects on the Population [5]. Region
that lead to flooding in the rainy season. Under this scheme, it is imperative to create scenarios of possible evacuations in case torrential rains can cause flooding in areas that are more prone to overflows and water stagnation. The objective of this work is to make a first approximation, by using a simulation of the behavior of people who live in an area susceptible to flooding as well as to analyze two possible scenarios where people may be located during the incident. All this assumes that climate change could alter the amount of water that falls in the rainy season and cause an overflow of the Río Bravo as well as floods in the Díaz Ordaz viaduct and the Acequia Madre.
2. Mathematical Models According to [7] there is a form to estimate the velocity of pedestrians by an Equation. "The pedestrian equation is based on the BDI methodology where the factors used are affected by the desires, beliefs, and intentions of the individuals" [7]. The velocity of an agent is dictated by Equation 1: Vi(t) = [v + h + nc)/(a + d]* f * imc * s Where: • xi is the velocity of agent i at time t. • To solve Vi(t) determines the position of an agent with respect to time. • v is the average pedestrian speed for all agents.
(1)
Humanitarian Logistics and the Problem of Floods in a Smart City 137
• • • • • •
h7] the criteria that can be used can be justify: • Density of people: If the density is higher the mobility decreases. • Level of consciousness: If the person is in a state of drunkenness or has just awakened his speed will not be the most, Additionally: The minimum space necessary to consider when a multitude is analyzed is equal to the vector Xi=[xivi]t, where xi and vi belong to R2 (radius between two agents) [8]. In this case, the pedestrian equation was implemented in a simulator that is named "Menge", where 2 types of people were taken, which have different size attributes but equal speeds, this because we assume there is a flood on the streets and the analysis is less complex.
3. Materials The following are the specifications of the equipment, software and materials used for the implementation of the simulations. Computer equipment to run the simulation System information Machine name: DESKTOP-FJF469O Operating System: Windows 10 Home Single Language 64-bit Language: Spanish (Regional Setting: Spanish) System Manufacturer: Dell Inc. System Model: Inspiron 15 7000 Gaming
4. Scenarios The stage is located as shown in the red polygon, as shown in Figure 1. These are houses adjacent to the Río Bravo in Ciudad Juarez's downtown neighborhood. The chosen scenario is interesting because the area is trapped between possible sources of flooding. The Río Bravo is located in the northeastern part while the Acequia Madre is in the southwestern, a natural stream; the Díaz Ordaz viaduct is located in the northwestern part. A satellite view of the stage can be seen in Figure 2. It is estimated that the people closest to the edges through which water flows, and therefore the first to experience flooding when there is torrential rain, will be the first to react to try to evacuate the area, while the more distant people will do so with some delay, which is why the model estimates that there will be an agglomeration of individuals trying to get out that will cause congestion by the exit routes.
Figure 1: View of the stage in the Colonia Bellavista located in the center of Ciudad Juárez [10]. It can be observed that the group of houses is located between the Río Bravo, the Díaz Ordaz Viaduct and the Acequia Madre.
Humanitarian Logistics and the Problem of Floods in a Smart City 139
Figure 2: Satellite view of the stage in the Colonia Bellavista located in the center of Ciudad Juárez. It can be observed that the group of houses is located between the Río Bravo, the Díaz Ordaz Viaduct and the Acequia Madre [10].
5. Simulation The simulation was performed using Menge software by modifying an open-source example called 4square.xml [9] and adapted to the conditions of the scenario as well as the objectives or goals to be achieved during the simulation of a flood evacuation in the borders surrounding the scenario. The stage was set by making a slight rotation of about 16º as it is shown in Figure 3 to be able to carry out the layout of the streets more easily, but for practical purposes, this is not an issue for the final results. In one of the scenarios, an evacuation of only people was contemplated, where 610 people intervened in groups of 10 people distributed in different locations of the scenario as shown in Figure 3. An estimate of 122 homes with families of approximately 5 people in equal conditions to mobilize during the disaster was estimated. Besides, it was estimated that the size of the people was the same, with a radius of 1 m for each of them.
140 Innovative Applications in Smart Cities
Figure 3: The image shows the distribution of people in groups of 10. Each red dot in the image represents an individual. Only one group of them, the one at the bottom of the stage, has 20 people.
In this part of the simulation, it was contemplated that the only objective of the people was to move from their initial location to a point located in the coordinate (250,0) as shown in Figure 4. The simulation considers an objective or goal, which must be reached by people. This objective is declared as a point from which displacement vectors which will be a reference of the speed direction to be maintained by the pedestrians are traced. This type of goal is simple, but in terms of simulation, it makes mobility difficult when the pedestrians encounter an obstacle that prevents their mobility in the direction of the displacement vector. That is why, during the simulation, they advance slowly in the "y" direction when there is an obstacle that prevents them from advancing in the "x" direction. Figure 5 shows the evacuation of the pedestrians towards the goal. The simulation takes place over a time of about 200 seconds (the program time is marked as 400 cycles) and is the maximum simulation time, so in order to facilitate the simulation, the movement speed of most pedestrians was chosen to be increased to 8 m/s, almost 5 times faster than the normal speed of a person who is moving freely [11]. According to actual dimensions and without considering obstacles in the path, a person running at an average speed of 1.5 m/s should complete the path indicated in Figure 6, with a length of 375 m, in 250 seconds or 4.16 min. However, it must be considered that the speed of the pedestrians will be affected by the environment, surely flooded, which would imply a reduction of their speed to
Figure 4: The blue dot represents the coordinates of the goal that people have to reach during the evacuation, which is located at the coordinate (250,0) according to the frame of reference.
Humanitarian Logistics and the Problem of Floods in a Smart City 141
Figure 5: Displacement of the pedestrians who congregate at the evacuation point at (250.0).
Figure 6: The segment marked in green has a length of 375 meters.
about 0.5 m/s, which would imply that by moving in a straight line it would take not about 4 minutes but a little more than 12 minutes to complete the journey. Another interesting situation to analyze is one in which people do not escape to an evacuation point, but gradually gather in the geometric center of the stage. Figure 7 shows this point in blue, which is located at the coordinate (41, -2). This situation is less realistic than the previous one because one would expect to escape from the flood sources located at the top, bottom and left of the scenario, however, it can be thought that, in the confusion, people decide to go in the opposite
Figure 7: The pedestrians move from their respective locations to the center of the stage at point (41, -2).
142 Innovative Applications in Smart Cities direction from the nearest flood source. However, the behavior of the people located to the right of the scenario would not have to move and agglomerate with all the others. 5.1 A more complex scheme Despite the results, it should be noted that these were obtained under the consideration that people evacuate the area on foot and do not move in vehicles. It is for this reason that a simulation that contemplates vehicles, with the capacity to move more people without them being exposed to rain or accidents, should also be considered. A vehicle was considered for each group of people, i.e., a total of 61 vehicles, each with an estimated radius of 2 m and moving at similar speeds to the people on foot. Figure 8 represents this new scheme, where the blue dots are the vehicles. The simulation reflects, in the case where evacuation to the point (250.0) is treated, how people and vehicles as a whole make movement slower as they become obstacles during travel due to their interactions. This behavior can be seen in Figure 9 and Figure 10. In this scenario, the vehicles have the same speed when they move, but in reality, their speed will be affected by the flooded environment, as well as by various obstacles. Also, the versatility of vehicles is much less than that of people, so evacuation times, in general, will be severely affected by the presence of vehicles. On the other hand, if one considers the case where people conglomerate in the center of the stage (41,-2), see Figure 11, it is observed that people and vehicles occupy more and more space and their interactions make movement difficult. Figure 12 shows this behavior.
Figure 8: Vehicles (blue dots) are located between people groups.
Figure 9: Evacuation simulation with vehicles and people. Vehicles are marked as blue dots while people are red dots.
Humanitarian Logistics and the Problem of Floods in a Smart City 143
Figure 10: Closer view of the simulation. Vehicles (blue dots) interact with people (red dots) and serve as obstacles during displacement.
Figure 11: The pedestrians move from their respective locations to the center of the stage at point (41, -2), blue dots represent vehicles while red dots represent people.
It
6. Conclusions and Future Work This simulation exercise should not be considered as a real predictive case of the behavior of a group of people in the middle of a disaster, such as a flood, but as a first approximation to estimate behavior and response times in the event of an evacuation. It should be noted that the real scenario has multiple quantitative variables that have not been considered in this paper, such as the age of people, sex, physical complexion, etc., due to the complexity of the mathematical model that would be required in order to address the behavior of pedestrians and vehicles involved in the simulation. On the other hand, there is no way to quantify the qualitative factors of people in a scenario like this and put them into an equation to estimate behavior. These factors can be the shock caused by the situation, fear of leaving home and possessions, and even disbelief at the consequences of the overflow of the Río Bravo and the imminent flooding in the area.
144 Innovative Applications in Smart Cities Besides the above, these types of models can help in urban planning at the time of building lots of houses, especially in areas susceptible to floods or overflows. In this sense, as in this hypothetical case, it could be observed that in the face of the flooding of the Río Bravo, the Acequia Madre and the flooding of the Díaz Ordaz viaduct, there are only a few evacuation routes for the people who live in the area. Also, the evacuation time will depend on several factors, but in the best-case scenario, as analyzed in Figure 6, it would take just over 4 minutes without considering severe obstacles or flooding that impede mobility, i.e., moving at a constant speed of approximately 1.5 m/ sec. However, this speed could be reduced by two-thirds due to terrain conditions and obstacles, so that time could easily be tripled to 12 minutes, which would put people's lives at risk. In the case of the more complex scheme, which involves the presence of vehicles, these play an important role in the mobility of individuals since they function as obstacles on the stage because they are less versatile than people, which will result in a much longer evacuation time than if it were only people. That is why this simulation represents a first step in the elaboration of protocols and evacuation routes in case this type of flooding occurs in the future. As future work, the idea would be to modify the simulation to establish possible emergency routes in case of floods, which would be used by people depending on their location in the area. Besides, the simulation can be improved by adapting menge to new platforms, such as Unity, which can be used to create more realistic scenarios and objects.
Simulating Crowds at a College School in Juarez, Mexico A Humanitarian Logistics Approach Dora Ivette Rivero-Caraveo,1,* Jaqueline Ortiz-Velez2 and Irving Bruno López-Santos2
1. Introduction Due to the frequency of natural disasters and political problems, interest in humanitarian logistics among academics and politicians has been increasing. In the literature, studies that analyze trends in humanitarian logistics were found to focus more on how to deal with the consequences of a disaster than on its prevention [1]. Simulation can be useful to pose different scenarios and be able to make decisions about strategies to help avoid stampedes when a natural or man-made disaster occurs. This helps to define a preventive strategy in the eventuality of a disaster. As a case study, we present a model to simulate crowds, based on a building of a college in Ciudad Juárez, Mexico: the Instituto Tecnológico de Ciudad Juárez (ITCJ). Ciudad Juárez is in the northern border area of Mexico. It is a city that has had a population growth due to a migratory process of great impact, receiving a significant number of people from the center and south of the country in search of better opportunities, which has resulted in many cases in the settlement of areas not appropriate for urban development, a situation that has been aggravated as the natural environment has changed negatively [2]. Recently, the migratory flow has also come from countries in Central and South America in the form of caravans seeking asylum in the United States. In 2016, the municipal government of Ciudad Juárez made an list of natural and anthropogenic risks. As for geological risks, the document mentions that in 2014 there were some earthquakes that measured up to 5.3 on the Richter scale, it is mentioned that: the province has a tectonic activity is an internally active zone and will have seismic activity sooner or later [2]. The ITCJ, founded on October 3rd 1964, is ranked number 11 among the National System of Technological Institutes [3]. The institution is located at 1340 Tecnológico Ave, in the northern part of the city. Figure 1 shows the satellite location obtained through Google Maps. Ciudad Juárez has a total of 27 higher education schools, and the ITCJ ranks third with a total of 6510 students enrolled [2]. To date, the institution offers 12 bachelor's degrees, three master's degrees, a doctorate and an open and distance education program [3].
Over the years, the institution has grown and new buildings have been built, with the Ramón Rivera Lara being the oldest and most emblematic. Figure 2 shows photographs of the building. To model the simulation, classrooms were taken into account in the aforementioned building, since it is the oldest building in the institute and it is where the majority of students are concentrated. This work presents a model and simulation based on the Ramón Rivera Lara building of the ITCJ. The objective is to evaluate the feasibility of using the Menge framework to simulate the evacuation of students and teachers in the event of a disaster. In the context of humanitarian logistics, simulations help to plan strategies before the occurrence of a natural or anthropogenic disaster. As future work, the aim is to model the whole institution and contrast the simulation against the simulacrums that are carried out in the school. Finally, to develop an informatic tool based on Menge and Unity so that decision-makers can evaluate different distributions in the classrooms and through the simulation, to be able to evaluate if it is possible to obtain a more efficient one that minimizes the risks in case of a disaster.
Simulating Crowds at a College School in Juarez, Mexico: A Humanitarian Logistics Approach 147
Figure 2: Ramón Rivera Lara Building [5].
2. Related Work This section presents a brief review of the literature from previous work related to the topic presented. It is divided into three subsections: humanitarian logistics, crowd simulation, and mathematical models. 2.1 Humanitarian logistics Humanitarian logistics is a process by which the flow and storage of goods and information is planned, implemented and controlled from a point of origin to the point where the emergency occurred [6]. Three phases are identified in the life cycle of a disaster: pre-disaster (preparedness phase), postdisaster (response phase), and finally, the recovery phase [7]. In the initial phase of the life cycle mentioned above, risk preparedness and prevention plans are established; in this regard, simulation can be a tool to evaluate prevention plans. In the pre-disaster or preparedness phase, it is important to identify different scenarios, specific risks, and complexity; simulations help to assess the risks in different scenarios [8]. The next section discusses crowd simulation, both literature, and tools. 2.2 Crowd simulation Crowd simulation is a fundamental problem of video games and artificial intelligence; recently, it has also been used for other serious applications, such as evacuation studies [9]. For the previous one, this sort of simulation can contribute to the planning phase in the humanitarian logistics life cycle, specifically to elaborate evacuation plans in case of a possible natural or man-made risk. These types of simulations apply to humanitarian logistics in four types of situations: trampling and crushing at religious events, trampling, crushing and sinking of ships, crushing at concerts and in bars, and contingency situations due to natural disasters, such as earthquakes, floods, fires, etc., that cause destruction to man-made structures [10]. For the simulation presented in this paper, we used Menge, which is an open platform based on C++. Menge is based on the needs of pedestrians and breaks down the problem into three subproblems: target selection (where people will move), computational plan and adaptation plan. This platform has the advantage that it does not require advanced knowledge of programming and multiagent systems for its use [11–13]. It provides documentation and examples in order to be able to adapt it to different contexts.
148 Innovative Applications in Smart Cities 2.3 Mathematical models According to [14], there is a way to estimate the velocity of pedestrians using an Equation. The velocity of an agent is dictated by Equation 1: Vi(t) = [(v + h+ nc)/(a + d]* f * imc* s
(1)
Where: • • • • • • • • •
xi is the velocity of agent i at time t. To solve Vi(t) determines the position of an agent with respect to time. v is the average pedestrian speed for all agents. h14] the criteria that can be used can justify: • Density of people: If the density is higher the mobility decreases. • Level of consciousness: If the person is in a state of drunkenness or has just awakened his speed will not be
3. Materials The following describes the hardware and software used to perform the simulation. 3.1 Hardware materials A Lenovo Laptop was used for the simulation and the characteristics of the device are shown in Figure 3.
Simulating Crowds at a College School in Juarez, Mexico: A Humanitarian Logistics Approach 149
4. Methodology To model the rooms and the section of the Ramón Rivera Lara building, first, the architectural plans of the building were analyzed. Figure 4 shows the upper view of the ground floor of the building. To establish the coordinates of the agents and the obstacles, measurements were taken of four halls adjacent to the ground floor. First, a single room was simulated and later, the simulation was made with the four rooms. To establish the speed of the pedestrians, the characteristics of the morning shift students were analyzed in the classes from 8:00 to 9:00 AM and Equation 1, which can be viewed in the previous section, was applied.
5. Simulation The simulation was divided into two stages. First, a single classroom was simulated using the methodology described in the previous section. Subsequently, four adjacent classrooms were used. 5.1 Simulation of a single classroom For this simulation, a project XML file was defined, which is shown in Figure 5. Inside the project folder, three XML files require Menge: scene, behavior, and view, as well as a file called graph.txt which contains the trajectories of multiple agents. Figure 6 shows the project folder with the four files mentioned.
150 Innovative Applications in Smart Cities
Figure 4: Ground floor of the Ramón Rivera Lara building.
Figure 5: Project XML File to simulate one classroom.
Figure 6: Project folder with scene, behavior and view XML files, as well as the graph file.
Simulating Crowds at a College School in Juarez, Mexico: A Humanitarian Logistics Approach 151
In the graph.txt file, the paths of the different agents were defined; Figure 7 shows some of the paths defined in that file. It is worth mentioning that the darkest blue agent represents the teacher, while the students are represented in light blue. One of the files that most require configuration is the scene file, since there are agents and obstacles declared. As the number of agents increases, this file grows proportionally. Figures 7 and 8 show some sections of this file.
Figure 7: Some of the trajectories towards the target of the agents.
To run the simulation, we must run the menge.exe and send as a parameter the project we want to run (XML file of the project) which in this case is Salon5.xlm. Figure 9 shows an example of how to run the simulation. Figure 10 shows the simulation in its initial, intermediate and final stages.
152 Innovative Applications in Smart Cities
Figure 8: Scene XML file segment 1.
Figure 9: Scene XML file segment 2.
Figure 10: Running simulation.
Simulating Crowds at a College School in Juarez, Mexico: A Humanitarian Logistics Approach 153
Figure 11: Simulation of a classroom.
Figure 12: Simulation of four classrooms.
154 Innovative Applications in Smart Cities 5.2 Simulation of four adjacent rooms To run this simulation, the procedure mentioned in Section 5.1 was used. The structure of the files is similar, simply scaled. The figure shows the simulation of the four adjacent rooms.
6. Conclusions and Future Work Crowd simulation can be a first approach to analyzing possible evacuation plans in an emergency. It can help to detect bottlenecks in the event of a mass evacuation, so the distribution of halls can be improved, minimizing such bottlenecks. One of the disadvantages is that this model does not take into account the stress and panic behaviors that a disaster situation can induce. As future research, the plan is to simulate the entire Ramón Rivera building, including the upper floor, as well as other ITCJ buildings. It is also planned to make a computer tool based on Menge and Unity for a more realistic simulation. This tool is intended to be used so that people who have no knowledge of the use of Menge can move some parameters in a way and change the simulation scenarios.
Perspectives of State Management in Smart Cities Zhang Jieqiong and Jesús García-Mancha*
1. Introduction The development of technology in public management will be the first step towards the transformation of large cities, the use of big data, technologies for industrial internet within the cloud, new accounting tools, budget management, and others. The transformation towards the development of the "Start cities" in countries like the Russian Federation, the People's Republic of China and Mexico and two societies in Africa as emerging powers of development is the first-level priority, they are usually commissioned to develop innovations in the areas of artificial intelligence, mass data processing, intranet, and computer security, that is why great efforts are being made in legislative matters by prioritizing laws whose main objective is the inclusion of a digital economy through the use of technologies, as it will cover absolutely everything in the development of trade, infrastructure, urban development, public transport, payment of taxes, etc. In the case of the Russian Federation, cities in full industrial development have a greater advantage over larger metropolitan cities such as Moscow and St. Petersburg, although these cities are large metropolises, their congestion, and limited growth space could hinder the use of new state management systems, for example in the use of new public transport systems and urban development, compared to emerging cities such as Kazan, Ekaterinburg, Rostov-on-Don or Sochi, which is the best-planned city in the federation, the "Intelligent transport" how introduction to the transport systems in these cities and future metropolises will be able to minimize the potential and future problems well in advance, in the particular case of the city of Kazan in the Republic of Tatarstan is contemplated the creation of a model of development of public transport, which will bring together public and private specialists from the sector of construction, insurance, civil protection, transport and communications and the automotive sector. As for the improvement of the quality of transport and road services, one vital point cannot be forgotten: road safety. Road accidents in these big cities are increasing year by year, mainly due to an imbalance between their infrastructure and the needs of citizens and the state; a strong investment is needed in the construction of new avenues, the maintenance of the existing ones, pedestrian crossings, ring roads around metropolitan areas to avoid traffic congestion [1]. In the implementation of such measures, a special role is played by the introduction of technical means of regulation with the use of electronic control systems, automation, telemetry, traffic control and television to control roads in a large area or throughout the city. The
156 Innovative Applications in Smart Cities construction of new communication and road transport centers is not enough in itself. The role of the intellectual component in the organization of the operation of the street and road networks is increasing. The concept of "Smart cities" in recent years in Mexico and Latin America has ceased to be considered a fantasy, due to the rise of interest in this topic, unlike the large metropolises of the Russian Federation mentioned above, the large cities in Mexico do have vital space to develop at an even faster level to increase the quality of life of its citizens, however, the challenge in Mexican cities is not the budget or government development plans, is the absence of political projects and a marked lack of automation strategies and legislation. In Mexico City, the main problem is transportation, since mobility in a city of more than 20 million inhabitants, in addition to the people who travel every day from the State of Mexico, is an alarming priority, as can be seen in Figure 1.
Figure 1: Data based on portal Cities in Motion Índex of the Escuela de Negocios IESE [2].
Mexico City is ranked 118th in the world in the use of technology in state management, and it is clear that there has been little development in the area of mobility and transport on a par with the use of technology and little legislation in this area. However, the combination of public and private initiatives is increasing day by day, and the population spends an average of 45 days a year using transportation [3]. Querétaro on the other hand already has a legislation developed since 2013 focused mainly on online tools, all public service information will be managed and connected to the internet 100% in the city by 2021, will be used in services such as garbage collection, payment of electricity, gas and water services, transportation services, traffic reports, while in the industrial sector will promote the use of sustainable development. In these times of innovation, humanity has entered an urban era, never in the whole history of mankind, half the population of the planet lives in cities, life is more connected than ever, connectivity is not measured in distance, it is measured in data consumption, data that is used as big data, cloud storage, etc. The functioning of government institutions in terms of information management improves performance and its development allowing later regional and municipal governance. There is much discussion about how and how much information should be collected from citizens. Intelligent cities are now an experiment for new public management to ensure the proper use of data, the quality of life of citizens and their rights, seeking a rapprochement between the citizen and the state. Other possible risks derived from the management of information and data collected in the intelligent city would be, on the one hand, the generation of information bubbles that thanks to big data and algorithms deform reality and only show us information according to our respective preferences and, on the other hand, the consolidation of the phenomenon of the so-called "post-truth", which consists of the possibility of lying in the public debate, especially on the Internet, without relevant consequences and without the supporters of those who have lied reacting even though the lie is even judicially proven. The "posttruth", in short, is built on a "certain indifference to the facts" [4]. In countries where corruption indexes are high, the panorama of misuse of citizens' information is one of the biggest challenges for the conception of smart cities, that is why emphasis is placed on the development of general public law, that is why transparency systems must have legal tools in the data collection and storage sets
Perspectives of State Management in Smart Cities 157
opened by state administrations, over time there will be a serious problem of duplication of data and many other issues associated with the digitization of documents that are currently written on paper, which will be very complicated to develop the digital transformation of state institutions, it will be necessary the participation of the general population and organizations to solve this problem, citizens will have to take responsibility for uploading their data legally if they want to be part of the process in question. For example, when a child is born, its parents are obliged to register it at the local civil registry office to obtain a birth certificate, the secretary of foreign affairs is obliged to obtain a passport, the institute of social security is obliged to obtain a certificate of insurance, and the institute of education is obliged to obtain a certificate of preschool education, at the time of the child's birth all his information will be captured in state databases in a digital data "Cloud", where the data of the citizens will be stored, the interaction between the citizens and the state will change forever, the state will not only be limited to provide basic services but also to manage all these data and more complex but at the same time more dynamic and fast life situations, by using tools and algorithms developed with high quality. The direction and priority of the development of these tools is directly focused on the improvement of bureaucratic, economic and social processes and the improvement of the quality of life of the cities, all the state information managed from a centralized system as an objective to improve the governance system also giving rise to new public services via electronic forming an intelligent system of "self-government". Not only is the use of citizen data proposed in the "Intelligent state management", but the systems will also contain information on the population, territories, properties, urban planning, social development, public space, budgeting, etc. This will mean a considerable increase in state income in a faster and more reliable way, for this reason, the following conceptual diagram is proposed, which explains in a general way the potential of digitalization in the management of public information in the framework of state management. Figure 2.
Figure 2: The evolution in the state management to optimize a Smart City.
The evolution in the state management has many questions before the automation of the processes, as it is already known the fear to the disappearance of traditional jobs is not well seen by anybody, it is not necessary to think about the disappearance of state jobs but of the evolution of these, will job disappear? Of course, they will, but at the same time, new ones will also arise, as has always happened throughout the history of the world. Rational management of resources
158 Innovative Applications in Smart Cities and automation in public management will seek to eliminate excessive costs, diversion of public resources, duplication of jobs, money laundering, etc. It will free up personnel resources that can be used in the improvement of other services to make the bureaucratic system more efficient. Constant monitoring 24 hours a day 7 days a week throughout the year will provide an audit of the resources to detect corrupt processes by detecting possible irregularities thanks to the implementation of new algorithms when creating public contracts, state concessions, recruitment of personnel to avoid conflicts of interest. 1.1 Public health and education The new management tools in the field of health will seek to adapt programs to provide better services by replacing obsolete, slow, cumbersome information management systems to make them less bureaucratic. With the popularization of applications and their use in the bureaucratic process, technologies for voice interaction, computer vision and cognitive computing, and cloud computing have gradually matured. Several applications of artificial intelligence have become possible in the field of medicine. Saving the time doctors spend on disease diagnosis, long waiting hours and patient mobility, improving diagnosis and greatly reducing care times and costs for the state. For another example, images are an important diagnostic foundation, but how to obtain high-quality medical imaging data is a difficult problem in the medical field, and deep learning can improve the quality of medical images, and deep learning technology using artificial intelligence can be extracted from medical images. Useful information to help physicians make accurate judgments. 1.2 Urban planning When we think about cities we think about buildings, streets, buildings, noise but we should think about people, cities are fundamentally for people, the buildings in the city and public spaces promoting more community among people, the Smart Cities is just a concept in which we add the management of aspects related to the environment such as water, electricity, transport, etc., with aspects linked to social programs of education, health and aspects associated with the administration, good governance when it comes to being efficient, obviously it is a model that depends on the use of new technologies so what is happening is that several words are always on everyone's lips when we talk about these smart cities, concepts such as sustainability, efficiency, effectiveness, innovation, and investment. Because this is a business with a lot of money in the development model of smart cities, why? Because of course there are many interests, we are not talking about money invested just because we want to solve a problem of resource efficiency, these resources have much to do with the capabilities with the development in the future thanks to the objectives of the smart cities, not only should be a mechanism to meet our challenges to make our economy more competitive and more efficient in the future, ICT in the thread of an intelligent city are from public administration to energy consumption to the use of urban transport all this has to do with the management of the Big data, the information they give us from the state archives, the crossing of information allows us to generate new solutions and applications to live in a much more efficient city, we are in a process of transformation not so much of the model of city but of the economic model, new opportunities for our industry and technology, to look for a different model of the city, an intelligent city cannot exist without an intelligent government, it is necessary to develop a model of education, creativity and innovation that are the motors to find a way so much if the city is a development of ICT as if the city is a development of the common good and the search of procedures that make us the life more "intelligent" the debate of the model of participation is needed, because at the end the great question is that most of the people want to be participant of all this project. The extended city model is considered an obsolete model, little dense and peripheral generates high costs, for this reason, it is important to study it since it generates an inequality in the quality with which the government provides public goods to the citizens, that is why the unplanned human settlements generate that the cost of the public services rises. Public policies and programs should be implemented to encourage
Perspectives of State Management in Smart Cities 159
the concentration of larger population centers, improving public services in rural areas, with urban planning complaints. This will promote the cooperation of different levels of government and the participation of civil society in the organization of the city, taking care of the economic and environmental order, Through the construction of buildings and urban settlements near major centers to generate jobs and avoid unnecessary expenses, each time you generate a construction of industrial, housing or any other is important to take into account the environmental factor according to the contact in which it is counted to generate a more sustainable building in this way it is possible to plan to organize and use the resources for each space or time.
2. Automation of State Systems in China and use of AI In China, over time, artificial intelligence is gradually being commercialized and is undergoing profound development in different fields. The project continues to be favored by major AI organizations. In the last five years, investment in artificial intelligence in China has grown exponentially. Since the first years of AI development, in 2015, the total investment reached 45 billion yuan in researching its development and new uses in state management alone, and it continued to increase in frequency in 2016 and 2017. In the first half of 2019, China's artificial intelligence sector raised a total of more than 47.8 billion yuan, and achieved remarkable results [5]. With the advent of the "AI+" era, innovations have been unleashed in hardware and software services such as mobile phones, the Internet of Things, cars and chips, and features such as face recognition and virtual reality have continued to expand. With the deep understanding of artificial intelligence technology by the business and investment communities, investment in artificial intelligence is becoming more rational. While the amount of investment in human wages and energy is decreasing, the amount of investment is increasing year by year. For example, the Shanghai government has provided tax incentives, capital subsidies, talent introduction and has optimized government processes to optimize the business environment, attracting a large amount of investment and funding for its public administration, artificial intelligence companies and talents, its scientific research force is outstanding. Promote the scale effect of upstream and downstream enterprises in the chain of artificial intelligence industry, and increase the strength of urban artificial intelligence industry. The top-level cities represented by Shanghai and Beijing have long been at the top of the ladder in terms of number of talents, number of enterprises, capital environment and scientific research capabilities. The number of artificial intelligence enterprises in Shanghai and Beijing cities has exceeded 600 all through private funding and state control. Among them, Shanghai has established business laboratories with technology giants Tencent, Microsoft and unicorn artificial intelligence merchant Tang and Squirrel AI who are currently working to develop AI uses in smart cities through public and private support and funding. Artificial intelligence empowers the financial industry to build a high-performance ecosystem with a broader range of capabilities, improves the financial efficiency of financial firms and transforms the entire process of internal company operations. Traditional financial institutions and technology companies have jointly promoted the deep penetration of artificial intelligence in the financial industry and state bureaucracy, restructured service architecture, improved service efficiency and provided customized services to long-distance customers, while reducing financial risks. Among the types of application of artificial intelligence technology in the field of education, adaptive artificial intelligence learning is the most widely used in all aspects of learning. In addition, due to China's large population, scarce educational resources and favorable factors such as the importance of education, it is expected that adaptive intelligent learning systems will be applied in recent years and will be able to reach even the most remote rural areas, to ensure that the entire population has access to education publicly, free of charge and universally through the use of distance education based on new educational models, for example the first Chinese textbook on artificial intelligence, aimed at rural secondary school students, was published earlier this year [6]. The construction of digital government affairs depends mainly on
160 Innovative Applications in Smart Cities top-down promotion so it is very important that the state is the first one to make use of the new tools that are available in terms of technologies since the main beneficiaries will be the citizens who in turn will provide large corporations with better qualified human capital to be able to understand and make optimal use of new technologies regardless of their age, the objective of the digitalization of government affairs is to accelerate the intelligent transformation of government. The requirements of building digital government in different places can become very diversified, so companies must provide customized solutions in view of the country's cultural diversity. The technology requirements in the country's major metropolises will not be the same as those in rural areas or in small or developing cities in the west of the nation. Barriers to entry in the field of public safety have been lifted. The automotive industry, dominated by driverless technology, will mark the beginning of innovation in the industrial chain. The production, channeling and sales models of traditional automotive companies will be replaced by emerging business models. The boundaries of the industry between emerging driverless technology companies and traditional automotive companies will be broken. With the rise of the car-sharing concept. Driverless technology carpools will replace the traditional concept of private cars. With the development of specifications and standards for the unmanned industry, emerging industries such as the safer and faster cars and at the same time will be able to solve 2 of the most serious problems of large cities in China and the world, the problem of traffic due to excessive vehicle fleet and pollution emitted by them significantly lowering travel times and carbon dioxide rates in cities, in addition to reducing health problems caused by pollution that in the end also represent a high cost to the state. For this reason, the potential for the application of artificial intelligence in the field of intelligent car manufacturing in large cities should not be underestimated. At present, the decrease in costs is greater than ever, and it is therefore possible to invest in this area as it is a guarantee of success for the future, even though high-quality data resources are not fully available or fully develop Through the use of algorithms that allow communications to connect their devices to an internet network, decision support systems can be made, to process large amounts of data for user support, control systems that also process data and allow "to manage" in real time such as intelligent lighting for energy saving or the traffic light network for full traffic flow to eliminate traffic problems in addition to obtaining real time data. The development and use of intelligent vehicle traffic management is an obligatory aspect in Smart Cities, which is not only limited to vehicle data, but also by using data obtained from the infrastructure to connect to the internet and process this data. The most used is the use of video cameras and different types of sensors, magnetic, infrared, radar, acoustic and of course the devices that travel inside the vehicles that are circulating. Through the simulation in real time to be able to predict the traffic at a certain time but the accuracy of the data will depend on the quality of the tools and their use, with the simulators it is possible to learn and understand the traffic in the Smart Cities the accomplishment of maintenance to the public roads, the pedestrianization, intelligent traffic lights etc. As previously mentioned, logistics companies will benefit and increase due to the demand for these intelligent systems. In the area of vehicle safety, the possibility of issuing fines in real time for the violation of traffic laws, such as ignoring traffic signs, parking in prohibited places will be detected by video surveillance systems that allow the identification of the vehicle by recording the license plates or the proportion of emergency vehicles in case they are necessary in the event of a breakdown of a vehicle that could compromise the flow of traffic.
3. The Learning of Government through the Entry of AI The analysis of AI investment trends is mainly divided into the following points: - Investors are looking for readily available AI application scenarios. In recent years, investment and financing data show that corporate services, robotics, medical and healthcare, industrial solutions, building blocks, and financial sectors are higher in investment frequency and amount
Perspectives of State Management in Smart Cities 161
of financing than other industries. From a company perspective, the world's most important equipment, financial strength, and technology genes are more likely to be favored by investors in the secondary market. From the industry perspective, the new retail, driverless, medical, and adaptive education that is easy to land indicates more opportunities, so companies in these areas have more investment opportunities. The investment market has begun to favor the underlying new technology companies. Unlike the previous investment preferences of applied artificial intelligence companies, the investment market has gradually started to focus on start-ups with underlying artificial intelligence technologies. The underlying technology is more popular, and due to the high ceiling, these companies are more competitive in the market. The development of the underlying artificial intelligence technology in China continues to lag behind that of the United States, and the underlying technology is an important support for the development of artificial intelligence, with the further development of artificial intelligence in China, investment in the underlying technology will continue to grow. - The proportion of companies that have won rounds A and B remains the highest, and strategic investments have gradually increased. Currently, more than 1,300 AI companies across the country have received venture capital investments. Among them, the proportion of A-round investment frequency started to gradually decrease. Investors continue to be very enthusiastic about round A, and it is currently the most frequent round of investment. Strategic investments started to explode in 2017. With the gradual maturity of the artificial intelligence market segment, leading companies, mainly the Internet giants, have turned their attention to strategic investments that seek long-term cooperation and development. This also indicates that strategic cooperation between the artificial intelligence industry and the capital level industry has started to increase. The giants are investing in artificial intelligence upstream and downstream of business-related industries. At the height of the development of artificial intelligence, Internet giants with a keen sense of smell have also initiated their strategic design. Technology giants like Alibaba, Tencent, Baidu, and JD.com have invested in various sectors of artificial intelligence, supported by technology investment funds backed by the Ministry of Science and Technology, the National Science Holding of the Chinese Academy of Sciences, the Local Finance Bureau and the Economic and Information Commission. In terms of fields, the projects in which investment institutions decide to invest are all before and after their future strategic industrial design, and these investment projects also promote the implementation of national strategies for the development of artificial intelligence. For example, Alibaba's investment is mainly focused on security and basic components. Representative companies that have won investments include Shangtang, MegTV, and Cambrian Technology. Tencent's investment focus is mainly in the areas of health, education and intelligent cars. Representative companies include Weilai Automobile and Carbon Cloud Smart. Baidu's investment focus is primarily in the areas of automotive, retail and smart homes. JD.com's investment focus is on areas such as automotive, finance and smart homes. The customer transformation and market strategy of the new retail platform Tmall.com, an online sales platform operated by the Alibaba Group. In the age of the internet, as traditional retail modes are concerned with the difficulties of finding sustainability, artificial intelligence technologies have been gaining popularity in the Chinese retail market. In addition to unmanned stores, new emerging innovations such as unmanned delivery vehicles and artificial intelligence customer support have also been launched or planned in China. The National Science Department, which is based on the Chinese Academy of Sciences system, is involved in artificial intelligence technologies and applications such as chips, medical treatment, and education. With the transformation and integration of digitization in various industries, artificial intelligence will become a necessity for giants in many fields such as automotive, medical and health care, education, finance, and intelligent manufacturing.
162 Innovative Applications in Smart Cities
4. Government and Smart Security The main purpose of intelligent security is: by transforming the unstructured image information in video surveillance into structured data that can be understood by computers, with the help of data processing, "mass video data" is transformed into "effective intelligence" to perform the security industry's intelligent upgrade from "seeing clearly" to "understanding and analyzing". Intelligent security needs to use machine learning to implement feature extraction, target recognition and power, organize into text information that can be understood by computers and people according to standard video content. This can drive significant improvements in image recognition and rating accuracy. The intelligent security industry chain includes primarily construction and maintenance engineering secondly, hardware and systems manufacturers, representing companies such as the listed intelligent security companies Hikvision, Dahua, etc. Thirdly, the software and algorithm companies in the artificial intelligence start-up companies, four large companies that need facial recognition (Shangtang Technology, Queng Technology, Yuncong Technology, Yitu Technology) and other technologies such as facial recognition, online identity verification, intelligent monitoring, and image recognition [7]. Because of the above mentioned, China already has a highly developed monitoring system unlike other countries like Mexico, which could take that leap and implement their intelligent security systems, the improvement of security processes will ensure the safety of the Smart Cities, public security agencies will constantly improve their working mechanisms by adopting various measures for the standardization of law enforcement and the protection of human rights, which in turn is one of the factors in carrying out security procedures. Inevitably, all municipal entities will at some point employ smart city technologies to optimize resource management and improve the lives of people living in the community. However, how they manage security will be the determining factor in the success of their efforts. This allows the analysis of moving crowds in urban areas, airports, train stations, shopping centers, and sports stadiums, among others. It is also used for forensic analysis, due to the capacity of intensive search of subjects in video recordings, for the location of suspects or automatic classification. However, the advancement of technology and the application of scenarios are a gradual process that should be taken as a priority in emerging countries, the delay in its transformation will pose a great threat to the economy and the efficiency of state processes, it is advisable to make a strong investment by the government to keep up with the technology that is currently evolving at a rapid pace as it is constantly changing day by day.
5. Conclusions and Future Research Living in an intelligent city will be an inspiration to everyone in the future, for governments, investors, citizens in general, to see groups of all ages use technology for the common good, the fact that everyone lives together as a united society that is connected beyond their creed, color or origin is wonderful, so will the cities of the future, cities that educate the talents and innovators of the future a home of advanced invention, a city where technologies coexist in an urban center, a city where people work together to create a new kind of city, the city of the 21st century, an intelligent city that works for everyone, a city that is not afraid of the challenges of the future. In addition, as a result, 12 societies in Africa will exceed 100 million inhabitants in this century and their capitals will become the next metropolis around the world, a perfect laboratory to implement the efficient management and process optimization of a future Smart City, as is shown in Figure 3.
This chapter explores the relationship among different routing policies for order picking and the features of the problem (describing both warehouse layout and orders), the results obtained by simulation show that some policies are especially sensitive to the presence of certain conditions that are likely to be present in real-world cases. Moreover, the routing policies are represented—for the first time in the literature as far as our knowledge—on structured algorithms. This contribution can facilitate their implementation because the features of the policies are modeled by formal mathematical structures, laying the foundations to standardize the way they operate.
1. Introduction A warehouse is a fundamental part of a company, and its performance can impact the entire supply chain [1]. Order picking is a problem that is present in all companies. It has received special focus from research areas related to planning and logistics. This fact is a consequence of several studies that identify order picking as the activity that demands more resources inside the warehouses, reaching up to 55% of the operational cost of the entire warehouse [2]. This activity has a strong impact on production lines, so companies with complex warehouses have areas dedicated to improving their product collection processes.
166 Innovative Applications in Smart Cities There are optimization models to support the resolution of this problem in the generation of product-picking routes; however, being considered as an NP-complete problem is not feasible to solve the models when working at medium and large scale due to the high cost that this represents. Thus, it is possible to apply some heuristics to get an approximate solution in real cases. Although several studies in the literature (e.g., [3]) show that these procedures are far from finding solutions close to the optimal one; these heuristics are still applied to real problems due to the simplicity and the way they relax the problem, granting a good balance between the quality of the solution and the ease of implementation. The picking routes are dependent on the structure of the warehouse and the properties of the orders, so studies have stated [4] that the input elements and the performance of the routes obtained are highly related. For example, a greater number of cross aisles facilitates movements inside the warehouse, so that the distance of the route tends to decrease. Throughout this chapter, we are going to define the algorithms for five of these heuristics and deepen on the study of which of them are more sensible to the characteristics describing the layout of a warehouse.
2. Background Warehouses are an important part of supply chain in a factory, and the main activities inside of it, like reception (receive and collect all product data), storage (move products to their locations), pick up (pick products from their storage location), packing (prepare for being transported), and shipping (place product in the transport medium). On this last step, the warehouse operation ends. 2.1 Warehouse Layout The chief features of the warehouse are: the central depot, where the picker starts its route and finish it, also usually is where the picker gets the order. The picker has to walk by the—the second element— picking aisles. A picking aisle is the space between two racks and facilitates the picker to pick product from the shelf, and the aisles have the following characteristics: length (distance between front aisle and rear aisle), and distance between aisles (distance that exists from the center of one aisle to center of next aisle); based on the distance, we can classify the aisles as short or long, in this case, we are going to use short aisles, which means that we can get the products from the shelves without making lateral displacements over the aisle. Cross aisles are perpendicular to picking aisles and are used to travel from one aisle to another, picking node is the location in the shelf where you can get the product, shelf is the space in the rack where products are stored, block is the area between the cross aisle and picking aisle, and sub-aisle is the area between the picking aisles inside the block if the blocks have picking aisles. Finally, picker is the person, tool or machine that picks up the products. Figure 1 represents an example of a layout of a warehouse with five cross aisles, four blocks, six picking aisles, and a central depot. Also, we are going to clarify the key concepts and briefly explain how the order picking routing policies work. 2.2 Steiner STP In the literature, it is verified that TSP (Travel Salesman Problem) is on the classification of NP-hard problems [5]; likewise, TSP and Order Picking have a close relationship. Unfortunately, the optimal solution may require intolerable run times for high-scale problems. Considering an approximate solution might be more convenient as this provides an favourable relationship between time and cost. Optimization algorithms based on local searches focus on achieving a good quality solution in a reasonable time to get a minimum or maximum value and avoid being stuck in a local optimum.
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 167
Figure 1: Warehouse layout elements.
It is necessary to start from a solution and, by applying operators, calculate solutions better than the initial solution. Normally, this strategy is applied to NP-hard problems, where heuristic functions are used to eliminate non-promising routes [6]. The solution for this project is represented as SPRP (Single-Picker Routing Problem), which consists of finding the shortest path that includes all the products to pick up [7]. This problem could be represented as a special case for TSP and can be applied as such to solve the initial SPRP problem. The objective is minimizing the distance and the time travel of the picker, either a human or machine, so it becomes a TSP. TSP consists of a salesman and a set of cities. The salesman must visit each of the cities, starting from a specific location (for example, native city), and come back to the same city. The challenge of this problem is that the salesman wishes to minimize the total duration of his/her trip. SPRP could be modeled as TSP, where the vertices of the correspondent graph are defined by the location of the available products inside of the warehouse and the location of the depot, as presented in Figure 2. This graph shows all the vertices and not only the picking ones, so the SPRP was modelled as a Steiner TSP (which is a variant of the classical TSP) that is defined as follows: Let G = (V, E) be a graph with a set of vertices V and a set of edges E. Let P be a subset of V. The elements of V \P are Steiner points. On a Steiner route, each vertex of P is visited only once. Steiner points should not be visited multiple times. However, a Steiner route could travel through
168 Innovative Applications in Smart Cities
Figure 2: Example of a warehouse structure as a set of V vertices.
some vertices and edges more than one time. In conclusion, the TSP of Steiner consists of finding a route of Steiner with the minimum distance [8]. Figure 3 shows an example of a warehouse layout with different parameters and variables.
Figure 3: Example of a warehouse structure with different parameters.
In Figure 4, the black-filled vertices are the picking vertices and the initial vertex, also known as the depot. This set of vertices is the set P, a subset of all the vertices V. This subset will form a Steiner graph, and the vertices formed at the intersections of the cross aisles and the picking aisles we will call Steiner points. Once the graph is obtained, the objective is to find a Hamiltonian circuit with the minimum cost. The initial and finish point of this circuit will always be the depot. Also, it is important to know that there are six different ways to travel through a picking aisle [9]. Figure 5 describes each one over one example of a unique block, one front cross aisle, and one rear cross aisle. Picker enters by the front cross aisle, crosses it completely, picks up all required products, and finishes leaving the aisle by the rear cross aisle.
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 169
Figure 4: Example of a Steiner graph.
Figure 5: Six ways to travel edges through the picking aisles.
Picker enters by the rear cross aisle, crosses it completely, picks up all required products, and finishes leaving the aisle by the front cross aisle. Picker enters and leaves twice through the aisle, enters once through the front cross aisle and once more through the rear cross aisle, picker enters and leaves by the same place. The picker will make its return defined by the largest gap, which is the largest distance between two adjacent picking vertices or the picking vertex and cross aisle. Picker enters through the front cross aisle, and its return point is the picking vertex farthest from the front aisle. Picker enters through the rear cross aisle, and its return point is the picking vertex farthest from the rear aisle. The picker doesn't need to travel through the aisle because there are no picking vertices inside.
170 Innovative Applications in Smart Cities These ways to travel are combined, generating different routing policies, which are highly popular in practice.
3. Routing Policies The routing policies determine the collection sequence of the SKUs (Stock-keeping unit) [10]. The objective of the routing policies is minimizing the distance traveled by the picker using simple heuristics [3]. To achieve this, it is necessary to consider the following features of the warehouse layout and the product orders, which can influence the final performance of each policy: quantity of products in the order, picker capacity, aisle length, and the number of aisles. Five of these heuristics are described below. 3.1 S-Shape The picker must start by entirely crossing the aisle (with at least one product) that is at the left or right end (depending on which is the closest one to the depot) until reaching the rear cross aisle of the warehouse. Then, the sub-aisles that belong to the farthest block of the depot are visited one by one until they end up at the opposite end of the warehouse. The only case where it is not necessary to cross a sub-aisle completely is when it is the last one in the block. In this case, after picking up the last product, the picker returns to the cross aisle from where it entered the sub-aisle. When changing blocks, the picker visits the closest sub-aisle to the last visited sub-aisle of the previous block. After picking up all the products, the picker must return to the depot [11]. Figure 6 shows an example.
Figure 6: An example of a route applying S Shape.
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 171
3.2 Largest gap This routing policy consists of identifying which is the largest gap in each sub-aisle and then avoiding crossing it [12]. This gap can be either between the rear cross aisle and the first product to be picked, between adjacent products or between the last product of the sub-aisle to the front cross aisle. All products that are before this gap will be picked up from the rear cross aisle and afterwards the picker must return to the cross aisle where it departed and do the same until all the sub-aisles of the block are explored, then, pick up all the remaining products from the front cross aisle. The sub-aisles that are completely crossed are those belonging to the first visited picking aisle and the last one of each block (this way, it passes from one cross aisle to another). When the picker picks up all the products, it must return to the depot. Figure 7 shows an example.
Figure 7: An example of a route applying Largest Gap
3.3 Midpoint This routing policy is similar to Largest Gap; the main difference is that the picker identifies the product to pick closest to the center of each sub-aisle, which is considered the travel limit [13]; at first, the products that are in the upper half of the sub-aisle are picked up from the rear cross aisle, then, after picking up all the upper half products of the entire block, continues picking up the remaining products from the front cross aisle. If the product is exactly in the center, the picker takes it from either of the two cross aisles. In the end, the picker must return to the depot. An example is represented in Figure 8.
3.4 Return When applying this routing policy, the picker enters and leaves the sub-aisle from the same cross aisle; this means that, after picking up the last product of the sub-aisle, the picker must return to the cross aisle [14].
172 Innovative Applications in Smart Cities
Figure 8: An example of a route applying Midpoint.
In the case that the warehouse configuration contains more than one block, the picker visits all the sub-aisles of the two adjacent blocks to the cross aisle alternately. After that, the picker moves to the next cross aisle that is adjacent to two unexplored blocks. The picker must return to the depot once all the products have been picked. This route is shown in Figure 9.
Figure 9: An example of route applying Return.
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 173
3.5 Combined This routing policy is considered a combination of S Shape and Largest Gap policies. When all products are picked up completely, the picker must decide between (1) continuing through the front cross aisle, or (2) returning to the rear cross aisle [14]. This decision is made according to the shortest distance to the next product to pick up. An example of this route is shown in Figure 10.
Figure 10: An example of a route applying Combined.
4. Development of Routing Policies The necessary elements for the development and application of the algorithms for each heuristic will be defined below. 4.1 S-Shape A fundamental part of the implementation of this heuristic is to define the order in which the picker will visit the sub-aisles, an example of the correct order according to its characteristics is shown in Figure 11. Also, to help obtain the order, it is necessary to assign "auxiliary coordinates" to each sub-aisle. Figure 12 represents an example.
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 175
Once the order is defined, the next step is to obtain the direction in which the picker will pick products from each sub-aisle. Algorithm 1 describes the procedure used to obtain the final path. Algorithm 1. S Shape in s in ascending order to C 6 Else if s-1 was explored in an ascending direction 7 Add products in s in descending order to C 8 Else if s-1 was explored in a descending direction 9 Add products in s in ascending order to C 10 If the current block was explored completely 11 Add products in s in descending order to C 12 While end 13 Return C 14 End
Where all the sub-aisles that are part of the first picking aisle are explored and added to the final path ascendingly (lines 4–5); then, the direction must alternate between each sub-block (lines 6–9) until the picker has visited the block completely; in that case, the first sub-block the picker visits in the new block is always traversed in a descending direction (lines 10–11). 4.2 Largest gap The order in which the picker will visit the sub-aisles in this heuristic is different in comparison to S-Shape. In Largest Gap, each explored block starts and ends from the same sub-aisle. Figure 13 shows an example.
Figure 13: An example of sub-aisles visit order generated by Largest Gap.
176 Innovative Applications in Smart Cities The application of this equation only depends on whether the sub-block is on the first picking aisle; in any other case, the blocks are explored from left to right. The following equation represents this: y f ( x, y) = ymax + (xmax −1)( ymax − ( y +1)) + x − 1
if x = 0, otherwise.
(2)
Once the order in which the picker will visit the sub-blocks is defined, the generation of the final route begins. This method is described in Algorithm 2. Algorithm 2. Largest Gap to C in an ascending direction 6 Else 7 Valuate elements in s 8 Calculate the distance between the current element and the next one 9 If the current distance is higher than the current limit 10 The new limit is the current element. 11 Traverse elements in s 12 If the current element is higher than the limit 13 Add the current element to C 14 Else 15 Add the current element to pending 16 If the current block was explored completely 17 Add elements in pending to C. 18 While end 19 Return C 20 End
Where the direction in which the picker traverses the sub-aisles that are in the first picking aisle is ascending, adding the elements to the final path (lines 4–5). Then, the largest gap of each sub-aisle is calculated by the distance between each of the elements that it contains; the new limit is defined by the element that detects the highest travel distance (lines 8–10), the next step is to add the elements that are above this limit to the final path (lines 11–12) and the elements that are below are stored in a stack (14–15); once all the elements of the block—that are above this limit—have already been added to the final path, lines 16–17 insert the pending elements in LIFO order (Last In, First Out).
4.3 Midpoint It is a heuristic similar to Largest Gap. They share the order in which the picker traverses the sub-aisles (Figure 13). Hereunder, the algorithm designed for the generation of the final path for Midpoint:
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 177 Algorithm 3. Midpoint Input: Sub-aisles s with at least one product Define the Midpoint value of the current block: LP-(LS/2) 7 Traverse elements on s 8 If the current element is higher or equal to Midpoint 9 Add element to C 10 Else 11 Add element to pending 12 If the current block was explored completely 13 Add elements on pending to C 14 Update Midpoint value of the next block: Midpoint-LS 15 While end 16 Return C 17 End
Where the sub-aisles that are in the first picking aisle must be traversed in an ascending direction while adding all the elements to the final path. (Lines 4–5). After, to obtain the midpoint and take it as a limit (line 6), starting from the farthest block to the depot, this value is obtained as follows: Mp = LA − Where:
LS 2
(3)
Mp is the midpoint, LA represents the locations per picking aisle, and LS represents the locations per sub-aisle. This value is defined as a limit until there is a block change (lines 12 and 14), where it must be updated as follows: Mp = Mp – LS
(4)
If the element is above the midpoint, it will be added directly to the final path, storing the rest on the pending elements (lines 8–11). The elements of the last sub-block of the block must be added completely in a descending way; so, it passes to the lower cross aisle and then starts adding the pending elements to the final path (line 13). 4.4 Return The first step in the implementation of this routing policy is to define the order in which the picker will visit the sub-aisles; Figure 14 shows an example of the correct order according to the previously described properties.
In this policy, the patterns of the order of travel are more complex than those seen before, so it was necessary to define a series of equations which consist of a function g(x, y) on which you can obtain direct results and functions h1(x, y) and h2(x, y) where it is necessary to call them recursively to reach the desired result. For the application of these equations, it is necessary to use the auxiliary coordinates exemplified in Figure 12. Once having the sub-aisles visit order, the process to generate the final path is shown in Algorithm 4:
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 179 Algorithm 4. Return Input: Sub-aisles s with at least products If the quantity of the blocks is an even number 7 If s is part of a block with an even coordinate 8 Add elements in s to C in an ascending direction 9 Else 10 Add elements in s to C in a descending direction 11 Else 12 If s is part of a block with a even coordinate 13 Add elements in s to C in a descending direction 14 Else 15 Add elements in s to C in an ascending direction 16 If s is part of the block on y=0 17 Add elements on s to C in an ascending direction 18 If the current block was explored completely 19 Add elements in s+1 to C in an ascending direction 20 While end 21 Return C 22 End
The elements found in the first picking aisle are added to the final path in an ascending direction (lines 4–5). Subsequently, it is important to define whether the number of blocks in the warehouse is even or odd (line 6). This fact influences because the characteristics of the routing policy states that the picker must alternate between the sub-blocks of two blocks, this implies that, in cases where the warehouse configuration has an odd number of blocks, the sub-blocks belonging to the last block to explore (closest to the depot) must be explored continuously (without alternating) (line 16). When the total number of blocks is even, all sub-aisles belonging to a block are crossed ascendingly, and in the odd ones in a descending direction (lines 6–10). In the opposite case (warehouses with an odd number of blocks), the sub-aisles that belong to the odd-number blocks are traversed in an ascending direction, and the sub-aisles that belong to the even-number blocks are traversed in a descending direction (lines 11–15). On the last block to be explored, all sub-aisles are visited ascendingly (line 16). In both cases, at the end of every block, the elements of the first sub-block of the next block (lines 18–19) are added to the final path in a descending direction. 4.5 Combined The order of how to traverse the sub-aisles is similar to S Shape (represented in Figure 11), but there are cases where it can vary because of the characteristics of this routing policy; when picking up the last element of every block, it is important to define which is the sub-aisle of the next block with product that has the smaller distance, if this sub-aisle is on the left end, the path of the block turns in the direction from left to right; otherwise (if the sub-block with the nearest product is in the extreme right sub-aisle), the opposite direction must be taken. The following equation represents the above behavior: if x = 0, y f (x, = y ) ymax + (xmax −1)( ymax − ( y +1)) + x − 1 if d1 < d 2 , y + (x −1)( y − ( y +1)) + (x − (x +1)) otherwise. max max max max
(8)
180 Innovative Applications in Smart Cities Where d1 is the distance between the last element of the block and the sub-block with the leftmost product and d2 is the distance between the last element of the block and the sub-block with the product on the rightest location. Once having the order in which the picker will visit the subaisles, the final route is generated as presented in Algorithm 5. Algorithm 5. Combined Input: Sub-aisles s with at least one product to pick up, visit order of sub-aisles. Output: Final path C 1 Start 2 While there is unexplored s with product Do 3 Select next s according to the visit order 4 If s is part of the first picking aisle 6 Add elements on s to C in an ascending direction 7 Capture the last element on s 8 Calculate d1: the difference between the last element on s and the first one on s+1 from the rear cross aisle. 9 Calculate d2: the difference between the last element on s and the first one on s+1 from the front cross aisle. 10 If d2 is greater than d1 11 Add elements on s to C in an ascending direction. 12 Else 13 Add elements on s to C in a descending direction. 14 If the current block was completely explored 15 Add elements in s+1 to C in an ascending direction 16 While end 17 Return C 18 End
Where the elements that belong to the first picking aisle are added to the final path in an ascending direction (lines 4–5), from this point, the distance from the last element of each subaisle to the first element of the next sub-aisle accessing from the front and rear cross aisle must be evaluated (lines 7–9). In the case that the distance to the first element from rear cross aisle is slower, the elements on the next sub-aisle have to be added in a descending direction to the final path; in the opposite case, the elements on the next sub-aisle are added in an ascending direction (lines 10–13). 4.6 Efficiency measurement To measure the effectiveness of the result, it is necessary to calculate the distance matrix between all product nodes and artificial nodes by applying Dijkstra. Subsequently, a sum of all the distances between consecutive nodes on C is made, represented by the following equation: p−1
d (C) = ∑ D( xi , xi +1 ) i =1
where: C = is the sequence of elements to evaluate, p is the number of vertices that forms a circuit, and D is the distance matrix.
5. Experimental Results In this section, the results obtained by this project are shown and interpreted. A total of 125 different warehouse layouts were processed and defined according to the combination of the following values:
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 181
Number of products capacities per sub-block: 18, 30, 42, 54 and 60 Number of picking aisles: 3, 6, 9, 12 and 15 Number of cross aisles: 1, 3, 5, 7 and 9 In each layout, five routing policies were processed and applied to five different orders with 4, 7, 10, 12, and 15 percent of the full capacity of the products in the warehouse. The results give a total of 3125 different instances (625 results per routing policy). This information was processed by SAS 9.4 software to generate a correlation analysis. The rank is one of the different variables to consider, it represents the position obtained by each routing policy in comparison to the other ones. The first position is assigned to the one that gets the lower distance and the fifth the highest. The locations per sub-block with or without product, the quantity of picking aisles, the number of cross aisles, and the percent of total locations of the warehouse that contains products to pick up were other variables used. 5.1 Insights Let us remember that, as the correlation gets closer to 1 or –1, the correlation is greater. Because this is a minimization problem, a correlation with the performance is better if it is negative. So, the main insights are: • S Shape tends to be sensitive to the number of locations by sub-aisles (negative correlation of 0.41824) and the number of products in the warehouse (negative correlation of 0.24840) (Figures 15a and 16a). • For Largest Gap, the correlation coefficient that stands out is the number of picking aisles, obtaining a positive 0.37988 (Figures 15b and 16b). Figure 15b demonstrated the tendency generated by this result, where the level of efficiency compared with the other policies decreases as this value becomes greatest. • For Midpoint, the two most relevant variables are the number of aisles and the number of products to pick up, both with a positive correlation of 0.33772 and 0.26812, respectively (Figures 15c and 16c). • The variable with the most effect over the performance of Return is the number of locations by aisle, where it gets a positive correlation of 0.58660. The more locations by aisle, the more competitiveness Return obtains. Figures 15d and 16d show that this policy gets better results as the number of aisles increases. • Regarding the results of Combined, the number of locations by sub-aisle seems to be an important feature, with a negative coefficient of 0.48762, considerably higher compared to the other variables (Figures 15e and 16e show). • Combined has better results in warehouses where the number of locations by aisle is the greatest variable, while the most degraded is Return. • A high number of aisles tends to affect the performance of the five policies, but the policy with the most unfavorable results is Largest Gap. Being S-Shape and Combined the least affected. • In the case of cross aisles, there was no improvement in the performance of any studied policy. The policies where its effectiveness decreases are Return and Combined. While in S Shape is just a little bit affected. • S-Shape is the most benefited policy in warehouses where the number of products to be picked up increases. To be able to develop this project, it was necessary to process orders, get five different results, and compare them over different methods. A benchmark of instances was synthetically created, and the performance in this wide range of different conditions was measured.
182 Innovative Applications in Smart Cities a) S shape
b) Largest Gap
c) Midpoint
d) Return
e) Combined
Figure 15: Pearson correlation coefficient.
The main purpose of this project is to explain and get knowledge on policies and to reduce traveled distances in order picking processes in warehouses, offering an encouraging panorama for the construction of more complex routing policies.
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 183
6. Conclusions and Directions for Future Work In this chapter, we studied and developed five routing policies to apply them in order to multiple scenarios and situations and generate enough information to find trends and understand the behavior of these heuristics. These experiments provide evidence that the performance of these policies can be highly sensitive to the different characteristics of the warehouse; for example, in the case of the S Shape routing policy, its performance is mainly affected by (1) the number of locations per sub-block, and (2) the size of orders of products to be picked. In the case of Largest Gap, the most marked trend is defined by the number of aisles, affecting its performance. Similarly, although not as marked as in Largest Gap, in Midpoint, the number of aisles tends to reduce its efficiency. The Return policy seems to be highly sensitive to the number of locations per sub-aisle, with performance deteriorating in a highly marked way. On the contrary, Combined improves as the number of locations increases. These insights can be used in the future to design a heuristic enriched with these key elements about routing policies. In this way, such a heuristic can foresee aspects of the test instances that can affect its performance and mitigate the consequences.
On the Order Picking Policies in Warehouses: Algorithms and their Behavior 185
1. Introduction A fish tank can be installed in various spaces, from the living room of a home, a consulting
room, a restaurant, an aquarium or a hotel. There are more than 400 ornamental species that have commercial relevance: zebrafish, angel, Japanese, molly or sword. So, the possibilities of this agrobusiness, whose demand is growing in the Mexican market, are numerous. Commissioner Mario Aguilar Sánchez, during the closing of the First National Watercolor Expo in the Federal District [1], noted that 60 million organisms are produced each year, worth 4.5 billion MXN, from about 700 productive units. He affirmed that the national production is developed in 23 entities, where 160 multi-species species are cultivated, such as koi carp, guppy, molly, angelfish, platy, danio zebra, tetra, cichlid, betta, gurami, sword, nun, oscar, plecos, catfish, shark, sumatra, dragon and red seal. The national production of ornamental fish is a business with prospects of social and economic growth that is developed in 23 entities, where 160 species and varieties are cultivated by aquarists, said the national commissioner of Aquaculture and Fisheries, Mario Aguilar Sánchez. However, according to various groups of breeders and authorities of the Federal Government, the great challenge for this segment to take off and generate wealth at the local level, consists of strengthening the breeding, sale, and distribution of fish, since it is now possible to satisfy the demand only by importing animals in large volumes. Among the existing varieties, one of the most popular is the so-called Goldfish or Japanese fish. The conservation and breeding of cold-water fish are not new concepts, since from ancient times in the Asian continent, particularly in distant China, people began to select beautiful specimens. In this research we focus on colorful Koi carp species, were often introduced in small outdoor ponds or ceramic pots. These animals were not only raised for ornamental purposes but also had a practical purpose since their conservation in captivity facilitated the ability to eat fresh fish at any time without the difficulty of capture in the wild. With the passage of time and the boom aquariofilia
Color, Value, Price and type Koi Variant Species for the Aquaculture Industry 187
acquires, it gives way to the selective breeding of specimens, producing a great variety of fish, both in colors and in certain peculiar characteristics of their phenotype. The state of Morelos is an ideal setting for the breeding of Japanese ornamental fish. Many of the businesses, most are familiar and small. When you launch into this world of aquiculture do it without any system of decision making that allows them to go a good way to obtain greater profits and in the shortest possible time [2]. But other states could have several problems implementing fish breeding which could benefit the economy of several places, that is why the cultivation of ornamental species had a considerable increase in Mexico. One of the states where this economic activity is emerging is Chihuahua, which, despite not having the most appropriate weather conditions although it has the physiological spaces for it. Thinking in the costs of implementing a tank for different states and the required technologies we proposed with this research to determine which is the ideal model to optimize breeding and development processes of the different species of Koi Fish. Determining the ideal and optimal value of a tank of koi fish is essential to specify the marginal gain of this type of project in aquaculture but the adaptation of the thank depends on several factors like the size of the carps in the tank and its quantity. Problem Statement A project is a temporary effort, with a variety of resources that seeks to satisfy several specific objectives in a given time. Innovation is the creation and use of new ideas that give value to the client or business. Proper planning of a Japanese fish breeding project depends on many factors. Detailed planning should be done, foreseeing risks that may arise. Technological Innovation Projects Scheduling Problems (TI-PSP) are a variant of Project Scheduling Problems (PSP). The PSP is a generic name that gives to a whole class of problems in which the best form, time, resources and costs for the programming of projects are necessary. The problem studied in the present investigation corresponds to a PSP theme because it involves variables of resource allocation to tasks and processes (computation). 1.1 Justification and purpose of our investigation At present, there are no mathematical models to optimize resources of Japanese fish breeding projects, so the present research is a milestone in the subject seeking a mathematical solution. Additionally, although it will be tested on Japanese fish farms projects, research will be the basis for any fish breeding problem that requires cost optimization. In addition, it will serve to centralize the use of economic resources in a more real way. The main social contribution of the research will be to expand the coverage of the budget presented by the breeders and increase the possibility of earning profits in less time. It will be achieved as the budget of each project is determined in real form, to avoid overcharging and underestimating them. Therefore, we propose to implement a method based on a generated dataset for identifying the size in the carps underwater using Artificial Neural Networks (ANNs) so that only carps over 10 cm size be detected with cameras helping to maintain the correct quantity of liters in the tanks, for the proposed economic model. We selected ANNs because there are good regression models as mentioned and they have been used before for smart farming, monitoring of the water quality environments and aquaculture operations, like shown in Figures [4–6].
2. Cluster Analysis and Artificial Neural Networks (ANNs) Cluster analysis [3,4] is an unsupervised learning technique that aims to divide a dataset into groups or clusters. It belongs to, like other typologies and that discriminant analysis, to the set of techniques that aims to classify individuals. The fundamental difference between cluster and discriminant analysis is that in cluster analysis the groups are unknown a priori and are precisely what we want to
188 Innovative Applications in Smart Cities determine. While in discriminant analysis, the groups are known and what we want to know is the extent to which the available variables discriminate against these groups and can help us to classify or assign the individuals to the given groups. Observations in the same group are similar (in a sense) to each other and are different in the other groups. Clustering methods can be divided into two basic types: hierarchical and partitioned grouping. Hierarchical clustering can be achieved with the agglomerated algorithm. This algorithm starts with disjoint clusters (n objects in a single group) and gradually proceeds to merge objects or clusters of more similar objects into a cluster. Algorithm 1. Basic Algorithm of Hierarchical Cluster Clusters Get the proximity matrix Repeat Merge the two closest clusters Update the proximity matrix that reflects the proximity between the new cluster and the new cluster Until there is a single cluster.
For the decision-making, a cluster analysis using hierarchical clustering and the agglomerated algorithm was employed. It was used as the input variables' initial budget and space to mount the Japanese fish farm (m2). The hierarchical cluster generates a dendrogram, which is a tree diagram frequently used to illustrate the arrangement of the clusters produced. A dendrogram is a diagram showing the attribute distances between each pair of merged classes in a sequential fashion. To avoid crossing lines, the diagram is graphically exposed such that the members of each pair of classes that merge are close elements. Dendrograms are often used in computational biology to illustrate the grouping of genes or samples, sometimes on top of heatmaps. After obtaining the dendrogram, a mathematical model will be applied to each element of the dendrogram. This will reveal the optimal values for the implementation of the project according to the values of entry of budget and quantity of square meters. ANNs are mathematical for representing biological neurons that were proposed in the 1950s , their applications cover several areas including regression models [4]. The activation functions are the hyperbolic tangent sigmoid for hidden layers, and the output layer is the linear activation function, which builds a good approximator for functions with finite discontinuities [7]. In this research, we train the neural network with the Scaled Conjugate Gradient (SCG) backpropagation and the Mean Square Error (MSE) as the expectation function in the equation (3). The SCG Backpropagation is a modified backpropagation proposed by Moller in 1993 that calculates the gradient in specific conjugate directions increasing convergence speed. With as the number of samples, as the target, and as the computed output of the FNNs.
3. Mathematical Model In this section, the mathematical model that was used to optimize the costs of the initial investments for breeding goldfish is addressed. The model analyzes the main elements necessary to start a business of Japanese fish farming. Each investment project includes material resources that are divided into infrastructure elements, equipment needed for cultivation, cost of young small fish (approximately 2 months old) and cost of fish feed. Most of the costs were obtained from the website Mercado Libre [5], except for costs for the construction of tanks [6]. The Objective Function corresponds to the budget necessary for the cultivation of Japanese fish and is formulated as follows: Rb = Af * Cf + Cff + I
Quantity of liters needed by a fish of 10 or more centimeters (constant value)
Cf
Cost of small fish (3-4 cm)
Cff
Cost of food for all fish
I
Infrastructure cost
Tc
Tank cost of 3 m x 2 m x 0.70 m
Ce
Cost per equipment
Nt
Number of tanks
Mt
Square meters of a tank (constant value 3m x 2m)
Where Af corresponds to the quantity of fish, Cff is the general cost of food for the fish, I corresponds to the estimated cost of infrastructure for the crop and Cf is the cost per Japanese fish where the average value is 10 MXN per fish. Model Restrictions: The budget Rb cannot exceed the initial budget that is denoted by Ib.
Ib ≥ Rb
(2)
Similarly, the restriction is made that the number of square meters used Nms cannot exceed the available Nma:
Nt refers to the number of tanks, Mt to the area each tank occupies in m2 and 2 is the value that is due to the space that must be left by each tank (2 meters wide). Af =
Nt * 3000 Lf
(7)
Where Af refers to the quantity of fish to be bought according to the quantity of tanks of 3m x 2m x 0.5m (3000 liters) and Lf is the quantity of liters required by Japanese fish of size greater than or equal to 10 cm,
In this, Tc is the cost per tank, where l is the length, a is the width and h the height of the tank, in this case l = 3 m, a = 2 m and h = 0.7 m. See table # 2. Cff =
Af * 0.408 * 170 1.5
(10)
Table 2: Description of the elements for the construction of the tanks. Concept
Costs
Annealed wall of 5.5 cm thick, common finish in flat areas
208.41
Polished planar with wood plane in walls, with mortar, cement-sand in proportion 1:6 of 2.0 cm of thickness, includes the repelled
125.61
Where Cff is the cost per food according to Af that is the quantity of fish, in a period of 4 months, period in which they must have 10 or more centimeters. See table # 3. For this computation [5,6] were used as reference where in the research the best results were obtained feeding the goldfish twice a day, where the amount is calculated by 2% of body mass. Ce = Mph + Wt + Ctf + Wp
(11)
Ctf = Clf * 16000 + Csf * 4800
(12)
Table 3: Measuring Elements for Goldfish Food. Concept
Amount
1.5 kg food package cost
170 MXN
Amount of food per fish in 4 months
0.408 kg
Ce is the cost for equipment. See table # 4. Table 4: Equipment and costs. Acronym
Concept
Amount
Mph
PH meter
225 MXN
Wt
Water Thermometer
Ctf
Cost per filters
Clf
Number of large filters de 45000 liters
Nt * 3000
Csf
Number of small filters of 6000 liters
Nt * 3000 – 45000 * Ci 6000
Wp
Water Pump, (1/2) Hp Siemens Centrifuga
80 MXN Ctf * 16000 + Csf * 48000 45000
1297.60 MXN
The following equation is the one that is responsible for calculating the cost of infrastructure (I). I = Ce + Tc * Nt
(13)
Color, Value, Price and type Koi Variant Species for the Aquaculture Industry 191
Model for determining the location of a koi carp to determine its size underwater Based on the method described in [8], suppose an observer at the edge of a pool perceiving an object immersed in water located at a distance and at a depth of incidence of the light on the water, which are determined using the law of refraction. The apparent position of an object seen by the observer is in the direction of the refracted ray and the object is located at the origin at a depth of but the observer perceives it at h-associated with incidence of the light on the water. A ray (red color) starts from the object and forms an angle of incidence θi. The refracted ray forms an angle with the normal one (Figure 1). According to the law of refraction, and visualized in Figure 1: nsinθi = sinθr (14) where n = 1.33 is the water refractive coefficient and the angle of refraction θr is greater than the incident θi.
Figure 1: Apparent position of koi carp under water based on refraction incidence.
The direction of the refracted ray and its extension passes through the point (xs,h) and its slope is 1/tanθr. Knowing that xs=htanθi. The equation for this line is y–h=
x – xs
tan θr
From the object, a ray (blue) forms an angle of incidence θ'i. The refracted ray forms an angle θ'r with the normal one. The direction of the refracted ray and its extension passes through the point (x's,h) and its slope is 1/tanθ'r. Knowing that x's=htanθ'i. The equation for this line is y–h=
x – x's
tan θ'r
The extensions of the refracted rays are cut at the point indicated in blue coordinates xa = ya =
tan θi tan θ'r – tan θr tan θ'i
(
tan θ'r– tan θr
h
)
tan θ'i – tan θ'i h tan θ'r– tan θr
This is the apparent position (xa, already) of the object as seen by an observer in the direction of the refracted beam. Where θ'i = θi+δ, where δ is a small angle increment. We represent the apparent position (xa, ya) of an object located at the origin, for various angles of incidence.
192 Innovative Applications in Smart Cities The depth of the object is h = 1 m, the angle increment δ = 0.01 degrees. The arrows indicate the direction of the refracted beam, the direction of observation of the object, as is possible see in Figure 2.
Figure 2: Apparent positions of koi carp under water as observed position varies.
The position of the observer is fixed In this section we will describe the apparent position of objects immersed in a pool in relation to an observer on the edge (Figure 3). The Y-axis coincides with the position of the swimmer and the X-axis with the surface of the water. The position of the swimmer's eyes is (0,y0) and the position of the submerged object (red dot) in the water is (xb, yb), the apparent position of the object (blue dot) is (xa, already). An incident beam from the object is refracted at a point on the surface separating the two media at xs, forming an angle θi with the normal. The refracted ray reaches the eyes of the observer forming an angle θr with the normal.
Figure 3: Perceived position of koi carp under water by observer in the edge.
Color, Value, Price and type Koi Variant Species for the Aquaculture Industry 193
Knowing the position of the object (xb, yb) and that of the swimmer's eyes (0,y0), from the law of refraction, nsinθi = sinθr, we will calculate the xs position, where the ray of light coming from the object is refracted, and the angles of incidence θi and the refracted ray θr. In the figure, we see that tan θi =
We solve this transcendent equation, to calculate θr, then xs and θi The equation of the direction of the refracted beam and its extension is y=–
x – xs tanθr
xb – x = y tan θr – yb tan θi To determine the apparent position of the object, we need to draw one more ray of refraction angle θ'r= θr+δ, (being δ a very small angle, infinitesimal) and find the intersection of the extension of the two refracted rays, as shown exaggeratedly in the figure. The equations of the red and blue lines are, respectively
{
xb – x = y tan θr – yb tan θi xb – x = y tan θ'r – yb tan θ'i
We clear the apparent position xa as it is the point of intersection of the two lines
Color, Value, Price and type Koi Variant Species for the Aquaculture Industry 195
Calculation of the apparent position of a submerged object as different Koi carps issues inner diverse distances form the water tank in Figure 4. Let y0=1.5 be the height of the swimmer's eyes, the position of the object (xb, yb) is (5,-2)m We solve the transcendent equation to calculate the angle of the refracted beam θr and the xs position on the water surface where the incident beam is refracted xb – y0
{
sin θr
√1 – sin2θr
+ yb
sin θr
√n2 – sin2θr
=0
and then the apparent position (xa, ya) of an object in the position (xb, yb) xa = xb + yb ya = yb
(n2 – 1) sin3 θr (n2 – sin2 θr)3/2
n2 cos3 θr (n2 – sin2 θr)3/2
We trace the incident beam, the refracted beam and its extension to the apparent position of the object (Koi carp) in Figure 6. After obtaining last equation, we get a mathematical model for determining the apparent position of the koi carps, but we perceive images of the apparent position through the camera, therefore, we could fix the equations for obtaining the real position of the carp with the apparent position but in the real life identifying is a difficult task. Alternatively, we propose to generate a dataset using a simulation with the equation (28) and varying its parameters, then train an ANN with the structure in section 2 for determining the real position with the input of the apparent position of carps, then that information will be used for determining the size of a carp.
4. Results and Discussion Design of experiment We built dataset with 100,701 points for training the ANN and generating the model, it was generated by varying in 0.01 m the real position of the carps (xb, yb) in ranges 0 ≤ xb ≤ 5 and –2 ≤ yb ≤ 0, considering a 2 m deep thank and 5 meters as maximum distance for perceiving the carp. All those positions are show in Figure 4.
Figure 4: Variations of position of koi carp underwater for generating the dataset.
196 Innovative Applications in Smart Cities We made a script to draw the appearance of a circular object of radius 0.5. After training the ANNs with the different architectures we used 10-fold cross-validation obtaining the cross-validated errors in Table 5, that allow us to define the best architecture for the neural network as 2 hidden layers and 8 neurons per layer, as is shown in Figure 5.
Figure 5: 10-fold cross-validated architecture for the ANN model.
Figure 6: Training Performance for ANN model.
Finally, we drew the apparent shape of the bottom of a pool as seen by the swimmer on the edge. The real shape is described by the function
yb =
{
–0.9 –
0.5xb 15.5
–1.6xb + 21.02 2.7 –3
0 ≤ xb < 15.5 15.5 ≤ xb < 18.2 18.2 ≤ xb < 25
The histogram comparison the performance in the training which also support the suppression of overfitting is shown in Figure 7.
Color, Value, Price and type Koi Variant Species for the Aquaculture Industry 197
Figure 7: Histogram Performance for ANN model.
Features of this tank water with Koi carps. The regression responses for the outputs xb and yb showing a correct response based on the values specified during the training are shown in Figure 8.
Figure 8: Regression Performance for ANN model.
198 Innovative Applications in Smart Cities The Aquaculturist does not perceive the bottom of the fresh water tank from a certain distance xb of about 6 m, for which it is already almost zero (see in Figure 9).
Determining the value of each issue is a complicated task, mainly because many of the issues are subspecies of other species and the valuation models are different for each species, as shown in Figure 10.
Figure 10: Different species analyzed in our study.
Color, Value, Price and type Koi Variant Species for the Aquaculture Industry 199
5. Experimentation In order to be able to simulate the most efficient arrangement of individuals in a social network, we developed an atmosphere able to store the data of each one of them representing individuals of each society, this with the purpose of distributing in an optimal form to each one of the evaluated societies. One of the most interesting characteristics observed in this experiment was the diversity of the cultural patterns established by each community. After identifying the best architecture, we trained the neural with the 80% for training dividing it in 70% for training, 15% for cross-validation in the training and 15% for testing, finally with the trained model we test the performance in the 20% reserved at the beginning as test set. The training results comparing the training performance, cross-validated performance, and test performance are shown in Figure 6, showing that there is not overfitting in the ratio between train, validation, and test responses. The generated configurations can be metaphorically related to the knowledge of the behavior of the community with respect to an optimization problem (to select Aquaculture societies, without being of the same quadrant [3]). The main experiment consisted of detailing each one of the 21 koi carp's variant. This allowed us to generate the best selection of each Quadrant and their possible location in a Koi Fish Pond, which was obtained after comparing the different cultural and social similarities from each community, and to evaluate each one of them with Multiple Matching Model. Using ANNs we determine the correct species, size and relatively the possible weight of an issue as is possible to see correctly in Figure 11.
Figure 11: Intelligent application to determine the correct parameters associated with the final price of an issue of Koi Fish species determined.
The developed tool classified each one of the societies pertaining to each quadrant, with..
200 Innovative Applications in Smart Cities The design of the experiment consists of an orthogonal array test, with the interactions between the variables: socialization, the temperature required, adult size, cost of food, maintenance in a freshwater tank, growing time, fertility rate, valuation in a sale. These variables are studied in a range of colors (1 to 64). The orthogonal array is L-N (2**8), in other words, 8 factors in N executions, N Table A: Orthogonal array associated with our research. Factors No.
Color, Value, Price and type Koi Variant Species for the Aquaculture Industry 201
is defined by the combination of possible values of the 8 variables and the possible range of color (see Table A and the importance of each issue in Table B, socialization attribute is a Lickert model with better socialization 7 and poor socialization 1). Considering features of porosity in a Fresh Water tank with many koi fishes and when weather conditions affect these.
Conclusions and Future Research Using ANNs, we improved the understanding substantially to obtain the change of "best paradigm", because we appropriately classified the agent communities basing to us on an approach to the relationships that keep their attributes, this allowed us to understand that the concept of "negotiation" exists with base in the determination of the function of acceptance on the part of the rest from the communities to the propose location for the rest of the same ones. ANNs offers a powerful alternative to optimization problems and redistribution of the clustering technique. For that reason, this technique provides a quite comprehensible panorama with a model which implies maintaining the size of the carps upper 10 cm, making that if is used a digital camera outside of the water detecting positions of the carps will be affected by the refraction of water and phenomenon represented [7]. This technique allows including the possibility of generating experimental knowledge created by the ANNs for a novel dominion of application. The analysis of the level and degree of cognitive knowledge of each community is an aspect that is desired to evaluate for future work. The answer can reside between the similarity that exists in the communication between two different cultures and as these are perceived [9]. On the other hand, to understand the true similarities that have different societies with base in the characteristics that make them contributor of a cluster and it as well allows him to keep his own identity, demonstrates that the small variations go beyond phenotypes characteristics and are mainly associate to tastes and similar characteristics developed through the time [6] to diverse variant of Koi carps.
Future Research With. Deep Learning offers a powerful alternative for object recognition. On the other hand, with the sizes identified in the tank it is possible to send koi carps to different tanks went its size is not over 10 cm. The general description of future research is shown in Figure 11.
Evaluation of a Theoretical Model for the Measurement of Technological Competencies in the Industry 4.0 Norma Candolfi-Arballo,1,4 Bernabé Rodríguez-Tapia,1,4,* Patricia Avitia-Carlos,1,4 Yuridia Vega1,3 and Alfredo Hualde-Alfaro2
This chapter presents the design and validation of a measuring instrument using the digital questionnaire evaluation technique, oriented to the self-perception of business leaders, to diagnose the current state of company work dynamics regarding the use, incorporation, learning, and technological appropriation. From the study carried out, a theoretical model capable of measuring the technological competencies of business leaders is obtained to diagnose the current state of companies' work dynamics regarding use, incorporation, learning, and technological appropriation.
1. Introduction Industry 4.0 was defined due to the growing trends in the use of ICT for industrial production, based on three main components: the internet of things (IoT), cyber-physical systems (CPS) and smart factories [1]. Industry 4.0 undoubtedly generates numerous new opportunities for companies, but several automation and digitalization challenges arise simultaneously [2]. Therefore, management, as well as employees, must not only acquire specific technical skills but appropriate them [3]. Multiple studies have been developed around the growth of industries generated by the technological factor [4,5,6,7,8,9,10,11,12,13,14,15], pointing to the challenges faced by underdeveloped countries to achieve high levels of competitiveness, industrial scaling and similar scopes to those registered by developed countries, and these refer to a vision regarding the proposals of technological appropriation in the productive processes on which the context is prioritized [16,17]. The research highlights the importance of clear top-down governance to succeed in the appropriate use of technologies since an "uncoordinated bottom-up series" would block the path to Industry 4.0. The following chapter shows the design and validation of a measuring instrument, using the digital questionnaire evaluation technique, oriented to the self-perception of business leaders to diagnose the current state of the companies work dynamics regarding the use, incorporation, learning, and technological appropriation.
204 Innovative Applications in Smart Cities The measuring instrument is composed of five dimensions called: technological competences, environment and internal communication, environment and external communication, training and updating, and innovation factors in the company; these were constructed from a theoretical review of the literature on technological competencies, considerations of the technological competences concept in the industry, international considerations, such as the European Framework of Competencies, Emerging Markets and national considerations regarding current technical and technological knowledge in the industry. The instrument was validated through expert judgment, by experts on social studies topics in the industry, and with experience in anthropological studies, management leadership, market momentum, digital marketing, economics, innovation management, and human capital in organizations. The reliability of the instrument is also performed by calculating Cronbach's alpha, which is an internal consistency indicator that measures the degree to which the items are correlated; that is, the items' homogeneity in the measurement instrument.
2. Industry Technological Adoption The formation of human capital is one of the focuses of attention priorities, specifically regarding to the updating of knowledge about technological equipment, development policies and everything implied from the planning of projects in the field of information technology, communication and collaboration to the evaluation of results, which should be related to the increase in the level of competitiveness, industrial scaling and improvement opportunities for companies [18,19], achieving active participation in the global field [17]. In that sense, analyzing the human capital of the industry allows the development of proposals aimed to strengthen their labor competencies and the evaluation of the advantage taken from the technological equipment used, considering it a continuous improvement activity [12,13,20,21,22]. The author Hobday (1995) in [16] describes the concept of Technology as a resource that combines physical capital and human skills, representing the dynamic capacity to create and increase the skills and knowledge in the industry, so it will allow a company to improve its skills, permitting a specific production process or development of new products to be integrated and adapted [16]. An essential part of Technology, the Information and Communication Technologies (ICT) or, in a more extended concept, Information, Communication and Collaboration Technologies (ICCT), as described in [23], refers to the possibilities to develop collaborative experiences by breaking time and space barriers by modifying industrial processes, forcing the need for changes in organizational structures and allowing new mechanisms of interaction and communication between company members and even between companies, promoting national and international cooperation. ICTs are currently a relevant issue in multiple areas of impact from the educational, productive and governmental sectors. Emphasizing the productive sector reveals that industry plays a key role in the development and implementation of ICT and vice versa since they are directly linked to production processes, innovation, solutions, and transformation of the goods and services that the market requires. Nowadays, the ICT industry is developing faster, driven by Asian countries such as China and Japan. The growth in the ICT industry has not only resulted in the development and production of equipment, but also in services that transform the global distribution of software production [11]. In [9], some advantages of incorporating ICT are highlighted: • Increase the efficiency of industrial and business processes, through updated information, historical information, indicators comparative, collaboration among employees, generation, and dissemination of knowledge, as well as the monitoring of profits and investments. • Communication with suppliers, minimizing delivery times and accelerating operations for the acquisition of required in time inputs.
Measurement of Industry 4.0 Technological Competencies
205
• Digital integration of a client portfolio, a historical selection of products, reports generation, administration, and selection of media. In this sense, it is considered necessary to analyze the adoption of technology in the industry to establish objectives that lead to an increase in the incorporation and appropriation of ICT, considering studies that are aware of the context in which companies are developing within their country. Thus, it is also necessary to analyze human capital concerning technological competences to face the changes in the already defined processes. To establish and promote a proposal for the inclusion of the ICT within a certain productive sector, it is considered necessary to explore and characterize the various profiles, obtaining indicators about competencies and ICT vision. That is, describe that business or leader profile and its relationship with ICT. There is a need for a structured methodology that describes the steps to build an ideal profile in terms of technology for an industrial leader that allows him to favor the inside of his company with the incursion of technology.
3. Technological Competencies Studies on Industry from Global to Regional Perspective The evaluation of technological competencies in the industry has a history of study and implementation. In Europe, programs are developed under the European Framework of e-competencies/European e-Competence Framework e-CF [24], a program composed of forty competencies in information and communication technologies for the industry. In the e-CF initiative, competencies are listed in five levels, aimed to cover technological needs. The actions are oriented to raise awareness, certification, and training of human capital, participation in innovative teaching-learning programs, mobility, and practices to attract young people to join technological careers at universities and to increase awareness about the importance of technological skills. Continuing with the international scenario, another interesting setup is the structure known as Information Technology Infrastructure Library (ITIL), which promotes the information technologies services within companies while considering quality standards. To assure this quality, ITIL proposes a series of standards for the development of infrastructure, appropriation, and operations based on information technologies. To date, there are several versions of ITIL, in the latest version of ITIL, the topics are classified into service strategies, service design, service transition, services operation and continuous improvement of services [25]. In Latin America, other case studies have been developed with quantitative, qualitative, or mixed evaluations. The proposed dimensions can be directed to perception, social activity, interactivity, use of the content, updating of practices, among others [26,27,28,29,30,31,32,33,34]. In the case of Mexico, [35] describes the assessment of the need for competencies to cover profiles in administrative positions where the attributes of advanced technology skills and computational knowledge are relevant indicators. In [36], an analysis of the learning of technological competencies in the maquiladora industry is shown. The analysis characterizes the export maquiladora industry and its industrial, technological and productive scaling. The results indicate that technological and productive scaling present weaknesses. The author proposes growth strategies, such as supplier development programs, public investment in innovation and development links between universities and companies, as well as the need for clarity in the objectives and nature of the industry in the electronics sector Mexico's northern border. In Baja California, studies analyzing various sectors of the industry have been conducted in the productive sector, mainly oriented to the Software, Electronics and Aerospace industries, studying their components, development and future vision under a socio-economic and cultural
206 Innovative Applications in Smart Cities analysis approach. In [11], human capital skills and labor abilities related to a particular sector are identified; in [12] diagnosis of the Aerospace industry is made and in [10] the policies for business development in the state are reviewed. On the other hand, at the Autonomous University of Baja California, while [5] have worked on models of competitiveness based on the information and communication technologies knowledge; [6], applied the matrix of technological capabilities into the industry of Baja California; and [37] described a systematic review of the literature on the concept of technological competence in the industry, rethinking the meaning of the term in knowledge areas seldom explored.
4. Evaluation Delimitations The study is aimed at the Renewable Energy Sector industry in the state of Baja California, to which belongs a group of companies identified as a new and promising national investment strategy. The population of interest is professionals within the Renewable Energy Sector, named for this project business leader, who is currently in a managing responsibility position in Small and Medium Enterprises registered in the state. The purpose of the evaluation focuses on the description of the behavior of an industrial sector in the state, regarding the technological competencies that their leaders demonstrate, without emphasizing the particularities of each company, that is to say, it is not intended to evaluate each leader individually and point differences in performance and levels of knowledge among the companies analyzed, on the contrary, the proposal corresponds to a comprehensive evaluation of the sector, demonstrating as final evidence a situational analysis and behavior in a diagnostic manner. 4.1 Methodological perspective The evaluation is structured under a methodological perspective of the ethnography of digital culture, where literacy and digital awareness processes are explored, as well as the ethnography of innovation, dedicated to social participation in technological innovation processes. An analysis is determined from complexity, inquiring about business leaders and its relationship with higher education institutions, government agencies, and business clusters. Likewise, the social impact and the conditions of the industry in a border area are reviewed, relating the variables to innovation factors and/or company development. The focus of the research is oriented to a quantitative study under a methodological analysis of the organizations [38]. The analysis in the industrial sector considers indicators attached to administrative, economic, and engineering sciences that provide concepts for intervention, attention, and follow up of studies in organizations. 4.2 Dimensions definition and construction of evaluation variables The dimensions and evaluation variables are structured based on a theoretical literature review of the evaluation of technological competencies [26,28,31,32,33,34,39,40,41]; the conceptualization of the term technological competencies in the industry [37]; international considerations, such as the European Framework of e-Competences (e-CF); the SFIA Reference Model (Skills Framework for the Information Age); emerging markets; Information Technology Infrastructure Library— ITIL [24,25,28,42,43,44,45,46,47,48,49,50]; and national considerations regarding technical and technological knowledge in the current Renewable Energy Sector industry [4,5,6,8,9,13,35,36,51, 52,53,54,55,56,57,58,59,60,61,62]. From the theoretical review, five evaluation dimensions are defined. The dimensions are the technological knowledge dimension, oriented to measure the use and mastery of electronic devices, the manipulation of specialized software and Web 2.0 applications; the dimension of environment
Measurement of Industry 4.0 Technological Competencies
207
and internal communication, where the internal structure of the company is evaluated, human relations and learning in terms of communication, knowledge acquisition, department analysis, training and promotion of human capital, professional degrees, and the impulse and technological vision of the company leaders are analyzed; the environment dimension and external communication, linkage, means and devices that make communication effective are assessed, collaboration between companies or sectors, working in networks for the company growth and recognition; the training and updating dimension, based on the follow-up of the leaders on issues to update technological knowledge, the preferred training modality, the training institutions, the periodicity of the training and the immediate application of the knowledge acquired in the training; and the company innovation factors' dimension, analyzed in terms of company innovation, from the products construction proposals, as their impact in the global market, patent registration, certifications acquisition, the administrative flexibility of the company's structure that allows the incorporation of an innovation and development group or department for the creation of new products, or, if such a group already exists, analyze conditions and context, as well as its impact on the company objectives. In Figure 1 the theoretical model of the evaluation is represented graphically. Regarding the measurement variables defined in the dimension of technological knowledge, in the software and hardware update, the version control in the programs and equipment is reviewed, as well as the constant revision of the market proposals; on interoperability and security, integrity and transfer of information; in collaboration and mobile applications, the technological tools that are used. In the environment and internal communication dimension, in the internal structure organization variable, it is reviewed what is related to the strategic planning of the company and its administrative conditions based on technological elements; the technological culture refers to the behavior of the leader within his work community, his relationship with technology and the diffusion in the use of it; the digital resilience considers the possibilities that the company shows to face computer problems, solve and quickly reorganize them, without affecting the processes and projects in execution. In the environment and external communication dimension, client tracking analyzes the structure defined for the acquisition and growth of the client portfolio; in the distribution strategies, the communication and collaboration methods with distributors or possible distributors are reviewed; in digital marketing, the promotion of the company is analyzed through social media and the market strategies used; in group participation, the collaboration and contributions of the leader in governmental, academic and industrial groups are investigated. In the training and updating dimension, the questions in training strategies are oriented to know how much the company leader promotes and receives updates; in innovative training practices, the modality in which the update courses are promoted and if mixed programs are considered is reviewed. Finally, in the company innovation factors dimension, the patents and new products/services variable is defined to review the results of the company on the design and registration of new products, models and/ or services; in innovation and development, the organizational conditions for the creation of spaces for development; in certification and regulation, the attention of human capital to the validation of knowledge by means of certifications and their knowledge about pre-established mechanisms in the sector for the regulation of processes. The quantitative measurement approach is proposed, and an evaluation instrument oriented towards self-perception of leaders is developed, using the digital questionnaire evaluation technique, with a nominal response scale. The evaluation instrument is constructed to diagnose the work dynamics current state on the state Small and Medium Enterprises companies of the Renewable Energy Sector, regarding the use, incorporation, learning, and technological appropriation. Table 1 shows the structure of the measuring instrument.
Figure 1: Theoretical evaluation model.
208 Innovative Applications in Smart Cities
Measurement of Industry 4.0 Technological Competencies
209
4.3 Variable operationalization The first version of the evaluation instrument was composed of 15 variables and 80 indicators divided into five dimensions, which were distributed as follows: Table 1: Evaluation instrument structure. Evaluation dimension
In addition to the mentioned items, the company identification data section is integrated into the instrument and the personal data of the business leader, with seven and four items, respectively. In Figure 1, the theoretical model of the evaluation is presented graphically, and Table 1 shows the indicators associated with each variable and evaluation dimension.
5. Validation of the Evaluation 5.1 Variable operationalization The Content Validity by Expert Judgment is applied [44,63,64], through a group of six judges with expertise in social studies topics in the industry and with anthropological studies experience, management leadership, market impulse, digital marketing, economics, innovation management and human capital in organizations. A table is used for the review, integrated by the dimension of the evaluation, the item number, the item, the item relevance (essential, useful and useless), clarity of the item redaction (legible or illegible) and general observations. Once validated, the Content Validity Reason (CVR) and the Content Validity Index (CVI) are calculated using the Microsoft Office Excel program. The CVI of the instrument is obtained by validating each item to determine if it is acceptable or unacceptable, indexes higher than 0.5823 are
210 Innovative Applications in Smart Cities expected, otherwise, the item must be removed from the instrument [64]. The operation variables are shown in Equations 1 and 2. ne CVR = — (1) N Equation (1): ne = Number of expert judges who agreed – essential, useful and useless; and N = Total number of judges. ∑ Mi–1 CVRi CVI = (2) M Equation (2): CVRi = Content validity ratio of acceptable items; and M = Total acceptable items on the instrument. 5.2 Item reliability Once the content validity test was made, a second reliability analysis of the instrument was carried out, through a pilot test, by non-probabilistic convenience sampling to 29% of the managers and/or executives from the 46 companies detected as potential for the analysis. The reliability of the instrument was carried out in the statistical software "Statistical Package for the Social Sciences (SPSS)", version 24. Through Cronbach's Alpha calculation, which is an indicator of internal consistency that measures the degree to which the items are correlated, that is the homogeneity of the items in the measuring instrument [65]. Its value ranges from 0 to 1, where the closer to zero the higher the percentage of the error in the measurement, while the instrument reliability is greater when closer to one [66]. An alpha greater than 0.7 is considered acceptable, greater than 0.8 is good and greater than 0.9 is excellent [67].
6. Results 6.1 Validation of the measuring instrument results The CVI global value was calculated at 0.93, based on Tristán's proposal [64], the result is catalogued as acceptable. Results show that 48 indicators from the total 80 obtained a CVR value of 1 the maximum scale score. Table 2 shows the CVR average per dimension on the evaluation instrument validation by the six expert judges; likewise, the items that were suggested to modify due to a lack of legibility are indicated. Table 2: Evaluation instrument validation. Average validation calculation (CVR)
Item illegibility
Technological knowledge
Dimension
0.89
1,2,4,5,7,8,9
Environment and internal communication
0.928
12,17,25,29,33
Environment and external communication
0.916
38,47,48,51,57
Training and updating
0.910
71
Company' innovation factors
0.937
73,79
Global CVI =
0.93
6.2 Instrument reliability results As can be seen in Table 3, a global Cronbach's Alpha of 0.972 was obtained in the instrument, which is excellent according to the acceptance values, and the individual values for each dimension are presented in it, showing an excellent result in the Environment and internal communication dimension as well as in the External communication dimension (0.968 and 0.913), good result for
Measurement of Industry 4.0 Technological Competencies
211
Table 3: Internal consistency analysis by dimension. Dimension
Cronbach's Alpha
Technological competencies
0.896
Environment and internal communication
0.968
Environment and external communication
0.913
Training and updating
0.868
Company' innovation factors
0.745
Instrument Total
0.973
Technological competences and Training and updating dimensions (0.896 and 0.868) and acceptable result in the company's innovation factors dimension (0.745). For this last case, the authors point out that results below 0.8 require reviewing the items writing [68], since it may not be understandable to the respondent. 6.3 Items reliability The review of the item's reliability is another relevant issue for the design of the instrument since, through this, it allows us to analyze if the items are consistent with the measurement of the instrument. For this test, the total item correlation indicator was used, which ranges from –1 to 1, and it measures the correlation of the items with all the others. Three criteria are considered if the correlation is close to zero, the question does not contribute to the scale if the value is negative, it is a question that is wrongly formulated or ambiguous, and if it is positive it is well related to the instrument; the closer to one, the greater the strength [69]. In Table 4, each item's correlations are shown, as well as the Cronbach's alpha corrected for the case of eliminated items. Therefore, question 40 is highlighted to be eliminated and, thus, increase Cronbach's alpha to 0.975. Besides, questions 69 and 70 were reviewed for being close to zero.
7. Discussion and Conclusions From the study carried out, a theoretical model capable of measuring technological competencies of business leaders is obtained, to diagnose the current state of the company's work dynamics regarding use, incorporation, learning and technological appropriation. The main dimensions identified and validated are the technological knowledge dimension, oriented to the measurement of the use and mastery of electronic devices, the manipulation of specialized software and Web 2.0 applications; the environment and internal communication dimension, where the internal company structure is evaluated, human relations and learning in terms of communication, acquisition of knowledge, analysis of departments, training and promotion of human capital, academic degrees, and the impulse and technological vision of the leaders in the company are analyzed; the environment and external communication dimension, the linkage is assessed, the means and devices that make communication effective, collaboration between companies or sectors, work in networks for the growth and recognition of the company; the training and updating dimension, based on the followup of the leaders on topics of technological knowledge updating, the preferred training modality, the training institutions, the periodicity of the training and the immediate application of the knowledge acquired in the training; and the factors of innovation in the company dimension, are analyzed in terms of company innovation, from the construction of proposals and products, as the impact of the same in the global market, the registration of patents, the acquisition of certifications, the administrative flexibility of the structure the company that allows the incorporation of a group or department innovation and development for the creation of new products, or, if that group already exists, analyze the conditions and the context, as well as the impact of the same on the objectives
0.863 0.677 0.666 0.481 0.627 0.732 0.600 0.436 0.579
P19
P20
P21
P22
P23
P24
P25
P26
0.506
P12
P18
0.869
P11
0.805
0.324
P10
0.808
0.758
P9
P17
0.532
P8
P16
0.679
P7
0.575
0.821
P6
P15
0.801
P5
0.676
0.734
P4
0.767
0.518
P3
P13
0.589
P14
0.193
P2
Total items correlation
P1
ItemsCronbach's alpha if the element is removed
P52
P51
P50
P49
P48
P47
P46
P45
P44
P43
P42
P41
P40
P39
P38
P37
P36
P35
P34
P33
P32
P31
P30
P29
P28
P27
Items
0.554
0.294
0.755
0.464
0.553
0.694
0.680
0.724
0.397
0.689
0.719
0.837
-0.310
0.601
0.231
0.455
0.639
0.794
0.863
0.648
0.831
0.799
0.756
0.873
0.649
0.822
Total items correlation97572
0.973
0.973
Cronbach's alpha if the element is removed
P79
P78
P77
P76
P75
P74
P73
P72
P71
P70
P69
P68
P67
P66
P65
P64
P63
P62
P61
P60
P59
P57
P56
P55
P54
P53
Items
Table 4: Total items correlation and Cronbach's alpha corrected.
0.312
0.427
0.485
0.487
0.206
0.802
0.530
0.291
0.632
0.062
0.062
0.282
0.252
0.282
0.589
0.752
0.469
0.687
0.341
0.726
0.429
0.652
0.662
0.195
0.886
0.527
Total items correlation
0.973
0.974
0.973
0.973
0.974
0.973
0.973
0.973
0.973
0.974
0.974
0.974
0.974
0.9740.973
0.973
Cronbach's alpha if the element is removed
212 Innovative Applications in Smart Cities
Measurement of Industry 4.0 Technological Competencies
213
of the company. These dimensions have operational variables, such as software and hardware update, interoperability and security, collaboration and mobile applications, internal structure organization, technological culture, digital resilience, customer tracking, distribution strategies, digital marketing, group participation, strategies for training, innovative training practices, patents and new products/services, innovation and technological development and certification and regulation. The final product has a measuring instrument, using the 79-item digital questionnaire evaluation technique. The theoretical conceptual model shown in Figure 1 was duly validated by the expert judgment employing the Content Validity Reason (CVR) and the Content Validity Index (CVI) obtaining the value of 1 for CVR and 0.93 for CVI. The measuring instrument was corrected in the reagents that had a lack of readability and then validated by Cronbach's alpha. It is important to note that Cronbach's global calculation of alpha was 0.972, which according to the acceptable values is considered excellent. As future work, it is necessary to validate the model by applying the measurement instrument to a considerable sample and thus be able to define a useful organizational diagnostic methodology to define a technological profile of a business leader concerning the incorporation, learning, and technological appropriation.
The technological progress, particularly in the implementation of biosignal acquisition systems, big data, and artificial intelligence algorithms, has enabled the gradual increase in the use of myoelectric signals. Its applications range from monitoring and diagnosing neuromuscular diseases to myoelectric control to assist the disabled. This chapter describes the proper treatment of EMG signals such as detection, processing, characteristics extraction techniques and classification algorithms.
1. Introduction The technological progress has made it possible for intelligent devices such as smartphones, tablets, and phablets to use sensors (like accelerometer, triaxial, gyroscope, magnetometer, and altimeter) to give the consumer a very intuitive sense of the virtual environment [1], but beyond the implementation of sensors in different devices, it has started the digitization of health. The Health and Healthcare in the Fourth Industrial Revolution article, published recently by the World Economic Forum, highlights that social networks, internet of things (IoT), wearables, sensors, big data, artificial intelligence (AI), augmented reality (AR), nanotechnology, and 3D printing are about to drastically transform society and health systems. Prominent leaders in health sciences and informatics have stated that AI could have an important role in solving many of the challenges in the medical sector. [2] mentions that almost all clinicians, from specialized physicians to paramedics, will use artificial intelligence technology in the future, especially for in-depth learning. A significant niche of this technological advance is related to the development of portable systems that allow the monitoring of biosignals and devices that can assist disabled people. Biosignals have been used in healthcare and medical domains for more than 100 years, among the most studied ones are electroencephalography (EEG) and electrocardiography (ECG), however, due to the development of commercial technologies for myoelectrical (EMG) signal acquisition, data storage, and management, monitoring and control based on EMG signals has increased [3]. Real-time evaluation of these signals may be essential for musculoskeletal rehabilitation or for preventing muscle injury. On the other hand, muscle activation monitoring is useful for the diagnosis of neuromuscular disorders.
Myoelectric Systems in the Era of Artificial Intelligence and Big Data 217
Recently, the HMI (Human-Machin-Interface) and IT communities have started using these signals for a wide variety of applications, such as muscle-computer interfaces. Sensors on human body extremities enable the use of exoskeletons, electric chair control, prosthesis control, myoelectric armbands, handwriting identification, and silent voice interpretation. Characteristics of EMG signals The EMG signal is known as the electrical manifestation of the neuromuscular activation associated with a contracting muscle. [4] defines it as "the current produced by the ionic flow through the membrane of muscle fibers that spread across intermediate tissues reaching the surface for the detection of an electrode", therefore, it is a signal which is affected by the anatomical and physiological properties of muscles, the control scheme of the nervous system, as well as the characteristics of the instrumentation that is used to detect and register it. The EMG signal consists of the action potentials of muscle fiber groups organized into functional units called motor units (MUs), this signal can be connected with sensors placed on the surface of the skin or through needle or wire sensors into muscle tissue. A graph of surface EMG signal decomposition at its motor unit action potentials is displayed in Figure 1. It is often desirable to review the data gathered at the time of individual motor unit discharges, in order to assess the degree of dysfunction in diseases such as cerebral palsy, Parkinson's disease, amyotrophic lateral sclerosis (ALS), stroke, and other diseases. Nonetheless, from a practical perspective, it is desirable to obtain such data from the signal detected from a single sensor that is as subtle as possible and that detects EMG signals high in MU rather than multiple sensors that detect EMG signals low in MU [5].
Figure 1: EMG signal and MUAP decomposition [5].
Electrical characteristic of EMG signals According to several authors' research, EMG signals can vary in amplitude from 0 to 10 mV and their energy density can be scaled from 0 to 500 HZ. An experiment performed by [11] identifies that the highest quantity of energy of an EMG signal ranges in a frequency scale from 20 to 150 Hz, making it very vulnerable to noise and interference. EMG signal contaminants [7] identifies two major issues of concern that influence signal accuracy. The first is the signal-tonoise ratio. In other words, the ratio of the energy in the EMG signals to the energy in the noise signal. The second concern is signal distortion, which means that the relative contribution of any frequency component in the EMG signal must not be altered.
On the other hand, [12] points out that the identity of a real EMG signal originated in the muscle is lost due to two main effects: the attributes of the EMG signal that rely on the individual's internal structure, including individual skin formation, bloodstream speed, skin temperature, tissue structure (muscle, fat, etc.), and external contaminants in EMG recordings, including inherent electrode noise, device motion, electric line interference, analog-to-digital conversion cutoff, quantization error, amplifier saturation, and electrocardiographic (ECG) interference. The major external contaminants are displayed in Table 2. Table 2: External contaminants of EMG signal. Contaminants
Due to the inherent characteristics of EMG signals, proper processing is necessary for its correct interpretation. An overall system based on pattern recognition consists of three stages: (1) Processing stage: The signal is collected with electrodes and preprocessed with amplifiers and filters and then converted into digital data. A raw signal is output as segments. (2) Extraction and characteristics reduction stage. It involves transforming the raw signal into a characteristics vector in order to highlight important data. At its output, there is a reduced vector of characteristics. (3) Classification stage. Classification algorithms are used to distinguish different categories between the reduced vector of characteristics. The categories obtained will be used for stages such as control commands or diagnostics. The following sections describe the main considerations for signal processing at each stage.
2. EMG Signal Processing: Signal Acquisition and Segmentation EMG signal processing consists of a series of stages that enable the information generated by muscle contractions to be processed and interpreted properly. The block diagram in Figure 2 clearly illustrates this required transformation.
Myoelectric Systems in the Era of Artificial Intelligence and Big Data 219
Figure 2: Block diagram of signal acquisition and processing.
2.1 Signal acquisition 2.1.1 Detection stage Two main techniques are used for EMG signal detection: Non-invasive, using surface electrodes (on the skin), and invasive, by inserting electrodes directly into the muscle (wire or needle type). Electrodes are normally used individually or in pairs, these configurations are called monopolar and bipolar, respectively [4]. The electrodes chosen for muscle and nerve registration will vary depending on the purpose of the research or the numbers of fibers to be analyzed. Superficial technique There are two categories of surface electrodes: passive and active. The passive electrode consists of a conductive surface (usually metal) that detects the current in the skin through its skinelectrode interface. The active electrodes contain a high input impedance electronic amplifier. This arrangement makes it less sensitive to impedance. The current tendency is towards active electrodes.
Figure 3: Different types of surface electrodes.
The disadvantages of surface electrodes lie in their restriction to surface muscles and that they cannot be used to selectively detect signals from small muscles or adjacent muscles. However, they are useful in myoelectric control for the physically disabled population, studies of motor behavior, when the activation time and magnitude of the signal contains the required information, or in studies with children or other people opposed to the insertion of needles [4]. Intramuscular technique The most common electrode is the needle type, like the "concentric" electrode used by clinicians. This monopolar configuration contains an isolated wire and a bare tip to detect the signal. The bipolar configuration contains a second wire and provides a second surface detection. [4] mentions that the needle electrode has two distinct advantages. One is that it allows the electrode to detect individual MUAPs during relatively low force contractions. The other is that the electrodes can be conveniently re-positioned within the muscle.
2.1.2 Sensor characteristics The experimental protocol for EMG signal detection plays an important role in giving greater reliability to the signal taken by the electrode, that is why it is necessary to take care of the properties of the sensor, skin preparation technique, sensor placement in the muscles, and electrodes fixing. Properties of the sensor In 1996, the Surface Electromyography for Noninvasive Assessment of Muscles (SENIAM) association was created with the objective of developing recommendations on key elements to allow a more useful exchange of data obtained by sEMG. After analyzing 144 studies, [16] points out the most used criteria regarding the configuration used, material, shape, size and distance between electrodes. Table 3 summarizes the desirable characteristics of the analyzed authors. Table 3: Users' desirable characteristics in sensor properties. Configuration
Bipolar
Material
Ag/AgCl
Shape and size
Round from 8 to 10 mm
Electrode distance
20 mm
Source: Author's own compilation
Placement procedure The most commonly used skin preparation techniques include: shaving, cleansing the skin with alcohol, ethanol or acetone, and gel application [16]. Sensor placement Three strategies can be identified for placement of a pair of electrodes [17]. ● In the center or most prominent lump of the belly muscle ● Someplace between the innervation zone and the distal tendon ● At the motor point The reference electrode is placed over inactive tissue (tendons or osseous areas), often at a certain distance from active muscles. The "popular" locations for placing the reference electrode have been the wrist, waist, tibia, sternum, and spinal process [16]. Fixing of electrodes The way the sensor is connected to the body is known as "fixation", this facilitates good and steady contact between the electrode and the skin, a limited risk of the sensor moving over the skin and a
Myoelectric Systems in the Era of Artificial Intelligence and Big Data 221
minimal risk of pulling the wires. Some methods may include adhesive tape (double-sided) or collar, elastic bands, and keeping the sensor in the desired placement by hand [16]. 2.1.3 Amplifier stage The quality of an EMG signal taken from an electrode will rely on the properties of the amplifiers, due to its nature, the amplitude of the EMG signals is weak and the amplifier gain must be in the range of 1000 to 10000. The consideration of incorporating amplifiers makes it necessary to have a high common-mode rejection ratio (CMRR), a high input impedance, a short distance to the source signal, and a strong direct current signal suppression [18]. 2.1.4 Filtering stage Analog filtering, usually bandpass, is applied to the raw signal before it is digitized. Bandpass filtering eliminates low and high frequencies from the signal. The low-frequency cut-off of the bandpass filter eliminates interference associated with motion, transpiration, etc., and any direct current (DC) compensation. Typical values for low-frequency cut are 5 to 20 Hz. The highfrequency cut-off of the bandpass filter eliminates high-frequency noise and prevents an alias from occurring in the sampled signal. The high-frequency cut-off must be high enough for the rapid on and off bursts of the EMG to remain easily identifiable. Typical values are 200 Hz–1 KHz [19]. The recommendations made by SENIAM for surface EMG are high pass with 10–20 Hz cut and low pass near 500 Hz [16]. The recommendation given by the International Society of Electrophysiology and Kinesiology (ISEK) for surface EMG are high pass 5 Hz and low pass with 500 Hz cut [20]. 2.1.5 A/D converter stage In the computer processing of EMG signals, the unprocessed EMG (after amplification and bandpass filtering) should be stored in the computer for digital processing. The minimum acceptable sampling is at least twice the highest frequency cut of the bandpass filter. For instance, if a 10–500 Hz bandpass filter was used, the minimum rate used to store the signal in the computer should be at least 1000 Hz, as indicated by the Nyquist sampling theorem, and preferably higher to improve accuracy and resolution, besides, the number of bits, model, and manufacturer of the A/D card used to display data in the computer should be provided [20]. It is desirable that as much information as possible be available to facilitate the interpretation of muscle contraction, however, the higher the sampling frequency, the more data will be collected in units of time, and this translates into more strict requirements for hardware equipment. Consequently, the cost can increase significantly, hence, appropriate reduction of sampling frequency is a highly desirable option [11]. Due to physical, processing, data transmission, and power consumption limitations, portable acquisition systems often sample EMG signals at a lower frequency than clinically performed (e.g., 200 Hz for the MYO armband or 250 Hz for the OpenBCI Cyton). In this sense, [21] developed a research to test the effects of frequency in the classification of basic movements of the hand and fingers of healthy subjects. Specifically, the study compared the effects of precision classification between 1000 Hz, frequency used in clinical acquisition systems, and 200 Hz, used in portable systems, finding that if the sampling frequency is lower than the one specified by the Nyquist theorem, there is an effect on the precision classification, however, it can be considered to work on the segmentation of data by analyzing small windows to analyze the data. 2.1.6 Amplitude analysis stage The EMG signal has a variation of amplitude in time, if a correct analysis in time is desired, the average of this signal will not provide useful information, since it presents variations above and below the zero value, that is why different methods are used for the correct analysis of amplitude.
222 Innovative Applications in Smart Cities Rectification: The rectification process is carried out before any relevant analysis method is performed. It entails the concept of rendering only positive deviations of the signal, it is achieved by eliminating negative values (half-wave rectification) or reversing negative values (full-wave rectification), the latter is the preferable procedure as it preserves all the energy of the signal [4]. Root mean square average (rms): An alternative to capture the envelope is calculating the value of the root mean square (rms) within a window that "slides" through the signal [19]. This approach is mathematically different from the rectification and filtering approach. [4] points out that, due to the parameters of the mathematical operation, the rms value provides the most rigorous measure of the information content of the signal, because it measures the energy of the signal. 2.2 Signal segmentation A segment is a subset of samples from a signal in which characteristics are extracted, and these characteristics are provided to a pattern classifier. The analysis window should have the following two considerations: the window time, considering the processing times of a classifier in real time, and the segmentation techniques, which can be adjacent or overlapping. Windows size. Due to real-time limitations, an adjacent segment length and the processing time to generate classified control commands must be equal to or less than 300 ms. In addition, the length of a segment must be appropriately large, as the bias and variance of characteristics increase as the length of the segment decreases, and consequently degrade the performance of the classification [22]. The same author notes that, due to real-time computing and high-speed microprocessors, processing time is usually less than 50 ms, and segment length can vary between 32 and 250 ms. Adjacent window. Disjointed adjacent segments with a predefined length are used for characteristic extraction; a classified movement emerges after some delay in processing. It is considered the easiest approach (used in the original description of the continuous classifier).
Figure 5: Adjacent window technique for an EMG signal channel. The data windows (W1, 12 and W3) are adjacent and disjointed. For each data window a classification decision is made (D1, D2 and D3) in time t , the processing time required of a classifier [23].
Overlapping windows. In this technique, the new segment slides over the existing segment, with a shorter increase time than the segment length.
Myoelectric Systems in the Era of Artificial Intelligence and Big Data 223
According to research performed by [23] and [24] regarding the effect of both techniques, they conclude that: overlapping segmentation increases processing time and produces no significant improvement, the segmentation of adjacent windows seems to achieve an increase in sorting performance. In this technique a smaller segment increase produces a denser but semi-redundant class decision flow that could improve response time and accuracy. [24] observed that a window of less than 125 ms produces high variation in frequency domain characteristics.
3. Extraction Methods for EMG Characteristics In the interpretation of EMG signals, characteristic extraction methods aim to transform the recorded and preprocessed signal, better known as "raw signal", and transform it into a relevant data structure, known as "characteristics vector"; in addition to reducing data dimensionality, these methods eliminate redundant data [3]. In this sense, the selection or extraction of highly effective characteristics is one of the most critical stages to improve sorting efficiency by [22]. According to [25], there are three sets of characteristics: (a) time domain, (b) frequency domain, and (c) time-frequency domain. Time domain characteristics are often calculated rapidly because they do not need a transformation. Frequency domain characteristics are based on the estimated power spectral density (PSD) of the signal, calculated by periodograms or parametric methods; these require more calculations, and time to be calculated. The characteristics in the time-frequency domain can locate the signal energy both in time and frequency, allowing a precise description of the physical phenomenon, generally requiring a transformation that could be computationally heavy. 3.1 Time domain This group of characteristics is widely used for pattern recognition in the detection of muscle contraction and muscle activity [26]. Due to their computational simplicity, time domain characteristics, also known as line techniques, are the most popular ones for EMG signal pattern recognition. They can all be done in real time and electronically, and their implementation is simple [27]. Characteristics in time domain are displayed in Table 4.
HEMG divides the elements in EMG signal into b equally spaced segments and returns the number of elements in each segment
Source: [3]
3.2 Frequency domain Frequency domain characteristics are mostly used for detection of muscle fatigue and neuronal anomalies [26]. They are based on power spectral density (PSD) and are calculated by periodogram or parametric methods [27]. These characteristics require more calculations and time to be calculated in comparison to the characteristics in time domain. The main methods are described in Table 5.
3.3 Time-frequency domain The characteristics in the time-frequency domain can locate the signal energy both in time and frequency, allowing a precise description of the physical phenomenon, generally requiring a transformation that could be computationally heavy. The primary methods are shown in Table 6. Tablet 6: Characteristics in the time-frequency domain. Short-term Fourier transform (STFT) Wavelet transform (WT)
Wavelet packet transform (WPT)
∫
STFTx(t, ω) = W* (τ – t) x (τ) e–jωτ dτ Wx(a, b) = x(t)
( √a1 ) Ψ* ( t –a b ) dt
WPT is a generalized version of the continuous wavelet transform and the discrete wavelet transform. The basis for the WPT is chosen using an entropy-based cost function.
Source: Author's own compilation
The main difference between STFT, WT and WPT is how each one divides the timefrequency plane. The STFT has a static pattern, each cell has an identical aspect ratio; while the WT has a variable pattern and the cell aspect ratio varies in a way that the frequency resolution is proportional to the center frequency. Lastly, the WPT has an adaptive pattern, which offers several tilt alternatives [28].
226 Innovative Applications in Smart Cities 3.4 Dimensionality reduction Dimensionality reduction is fundamental to increase the performance in the classification stage. In this process, the characteristics that best describe the behavior of the signal are preserved while the number of dimensions is reduced. There are two main strategies for dimensionality reduction [29]. Characteristics projection: this strategy consists of identifying the better combination of the original characteristics to form the new set of characteristics, usually smaller than the original, the principal component analysis (PCA) can be used as a characteristic's projection technique [25]. PCA produces a set of uncorrelated characteristics by projecting the data into the vectors of the covariance matrix [30]. Characteristics selection: This strategy selects the better subset of the original characteristics vector according to certain criteria to assess whether one subset is better than another. The ideal criteria for classification should be to minimize the probability of misclassification, although simpler criteria based on class separability are generally selected [25].
4. Classification Algorithms Once the characteristics of a recorded EMG signal have been retrieved and the dimensionality has been reduced, some classification algorithms must be implemented. [22] advises that, due to the nature of the myoelectric signal, it is reasonable to expect a wide variation in the value of a particular characteristic. In addition, there are external factors such as changes in electrode position, fatigue, or sweating that cause changes in a signal pattern over time. However, a classifier should be able to cope optimally with such variable patterns; it must be fast enough to comply with restrictions in real-time. There are several classifier approaches such as neural network, Bayes classifier, fuzzy logic, linear discriminant analysis, support vector machine, hidden Markov model and k-nearest neighbors [3]. The summary of the main classification algorithms is displayed in Figure 8. Examples of the uses of the different classifiers are shown in Table 7.
5. Conclusion The development of technology and portable system applications to monitor and control through myoelectric signals is possible thanks to acquisition systems, real-time processing, and classification algorithms, associated with the analysis of large amounts of data. This has made it possible to detect, process, analyze and control signals as small and complex as those generated by any muscle contraction. Knowing each stage in the processing of these signals allows us to identify criteria for the design of new human-computer interfaces, more efficient and useful for the user. No doubt there is a need for proper detection and ergonomic systems, despite the efforts of communities such as SENIAM and ISEK, the mapping for the location of sensors are still being studied. On the other hand, portable acquisition systems must be developed with adequate characteristics in the sampling frequency, in order to decrease computational costs in processing and time, but without losing vital frequency spectra for the correct monitoring and interpretation of patterns. The development of statistical algorithms, data analysis, and artificial intelligence are making possible the optimization of relevant characteristics in the interpretation of patterns, allowing the reduction of the raw data dimensionality of the sampled signals to facilitate their interpretation by the different classification algorithms. The difficulty of systems that can interpret patterns through myoelectric signals lies in the diversity of the anatomical set of users, the placement of sensors, and the relevant characteristics, which is why the algorithms of machine learning or deep learning can allow greater progress in dealing with each of these variables.
Implementation of an Intelligent Model based on Big Data and Decision Making using Fuzzy Logic Type-2 for the Car Assembly Industry in an Industrial Estate in Northern Mexico José Luis Peinado Portillo,1,* Alberto Ochoa-Zezzatti,1 Sara Paiva2 and Darwing Young3
These days, we are living in the epitome of Industry 4.0, where each component is intelligent and suitable for Smart Manufacturing users, which is why the specific use of Big Data is proposed to determine the continuous improvement of the competitiveness of a car assembling industry. The Boston Consulting Group [1] has identified nine pillars of I4.0, which are: (i) Big Data and Analytics, (ii) Autonomous Robots, (iii) Simulation, (iv) Vertical and Horizontal Integration of Systems, (v) Industrial Internet of Things (IoT for its acronym in English), (vi) Cybersecurity, (vii) Cloud or Cloud, (viii) Additive Manufacturing including 3D printing, and (ix) Augmented Reality. These pillars are components of the Industry 4.0 that can be implemented as models of continuous competitiveness. In Industry 4.0, the Industrial IoT is a fundamental component and its penetration in the market is growing. Car manufacturers, such as General Motors or Ford, expect that by 2020 there will be 50 billion (trillion in English) connected devices, Ericsson Inc. estimates 18 billion. These estimated quantities of connected devices will be due to the increase in technological development, development in telecommunications and adoption of digital devices, and this will invariably lead to the increase in the generation of data and digital transactions, which leads to the mandatory increase in regulations, for security, privacy and informed consent in the integration of these diverse entities that will be connected and interacting among themselves and with the users. Finally, the use of Fuzzy Logic type 2 is proposed to adopt the correct decision making and achieve the reduction of uncertainty in the car assembly industry in the Northeast of Mexico.
1. Introduction Today, technology is an important part of everyday life, from the way we communicate to the different types of technologies that allow us to carry out many types of processes in different industries. Universidad Autónoma de Ciudad Juárez, Av. Hermanos Escobar, Omega, 32410 Cd Juárez, Chihuahua, México. Universidad de Portugal. 3 Centro CONACYT * Corresponding author: [email protected] 1 2
230 Innovative Applications in Smart Cities The Mexican industry, particularly the automotive industry, is not exempt from these technological advances, which are part of industry 4.0 (I4.0), and has an endless number of technologies that make it competitive in the market. However, these technologies are not effective enough to meet the demands of today's world, therefore, this chapter will show a literature review of the concepts that will be the basis for the proposal of a new intelligent model that is able to combine cutting-edge technologies and optimize processes and resources within the automotive industry in northern Mexico.
2. Literature Review This section shows the main concepts of this article and how they have been generating and evolving throughout history. This section gives us an idea of what exists with respect to the technologies mentioned as Industry 4.0, Big Data, Fuzzy Logic Type-2. 2.1 Industry 4.0 Industry 4.0 (I4.0) is the latest standard for data and computation-oriented advanced manufacturing [2]. The term "Industry 4.0" originated from a project initiated by High-tech strategy of the German government to promote the computerization of manufacturing. Industry 4.0 is considered as the next phase in the digitization of the manufacturing sector, and it is driven by four characteristics: the amount of data produced, the increasing requirements of computational power, the usage of artificial intelligence techniques, and connectivity to high-speed networks [3]. The I4.0 was named thusly because it is the fourth industrial revolution, the first one (I1.0) refers to the first revolution which occurred in the 1800s, where the most important change was mechanical manufacturing, then in the 1900s the second revolution heralded the arrival of the assembly line, leading to an increase in mass production, and the third revolution occurred around 1970 with the introduction of robots that improved production efficiency. All this information is presented in the next table. Table 1: Technology evolution from Industry 1.0 to Industry 4.0 [2]. Time
Evolution Transition
Defining technology
1800s
Industry 1.0
Mechanical Manufacturing
1900s
Industry 2.0
Assembly Line (mass production)
1970
Industry 3.0
Robotic Manufacturing (Flexible Manufacturing)
2010
Industry 3.5
Cyber Physical Systems
2012 onward
Industry 4.0
Virtual Manufacturing
As mentioned before, the I4.0 is based on nine pillars; this was written by [1] and the pillars are: 1. 2. 3. 4. 5. 6. 7. 8. 9.
2.2 Big data One of the most important parts of I4.0 is Big Data and Analytics, which normally is associated with the result of the use of internet, sensors, management systems, but big data isn't about a big group of data, it is a model named "Model of 3 v's", i.e., Volume, Velocity, Variety [4]. Then this model was increased with a new "v", variability [5] for the "Model 4 v's", the next suggested for the "Model 5v's" was value, and over time this model has been increasing to the last model named "3v2 Model" and is mentioned by Wu et al. [6], and they show us the next Venn Diagram:
Figure 1: 32 v's Model for Big Data.
Some of the authors like Zhang et al. [7] talk about the use of Big Data in the automobile industry. They propose that the use of big data helps determine the characteristics that a user searches for in a car, in addition to predicting how sales will be in the coming months. Otherwise, Kambatla et al. [8] talk about the future to big data. They give us an idea of what the use of big data implies, from the type of hardware that is needed to apply this technology, be it the use of memory or the hierarchy of memory that this implies, to the types of network and systems distributed that enable the application of big data for companies. Furthermore, Philip Chen and Zhang [9] mention that in order to be competent, the use of big data is a big part of innovation, competition, and production for any company and that the use of big data should include the use of cloud computing, quantum computation and biological computation, besides that, the development of tools is an important part of the use of these technologies. 2.3 Fuzzy logic type-2 Fuzzy logic has attracted the attention of researchers for the last couple of decades. It has opened new horizons in both the academia and industry sectors, although, conventional fuzzy systems
232 Innovative Applications in Smart Cities (FSs), or so-called type-1 FSs, are capable of handling input uncertainties, they are not adequate to handle all types of uncertainties associated with knowledge-based systems [10]. The type-2 provide additional design degrees of freedom fuzzy logic systems, which can be very useful when such systems are used in situations where lots of uncertainties are present. The resulting type-2 fuzzy logic systems (T2 FLS) have the potential to provide better performance than a type-1 (T1) FLS [11]. A type-2 fuzzy set is characterized by a fuzzy membership function, i.e., the membership value (or membership grade) for each element of this set is a fuzzy set in [0,1], unlike a type-1 fuzzy set where the membership grade is a crisp number in [0,1] [12]. Membership functions of type-1 fuzzy sets are two-dimensional, whereas membership functions of type-2 fuzzy sets are three-dimensional. It is the new third-dimension of type-2 fuzzy sets that provides additional degrees of freedom that make it possible to directly model uncertainties [11].
Figure 2: Diagram of a fuzzy logic controller.
3. Discussion The automobile assembly industry today has multiple options for the assembly, from different models of cars, different types between these models, even the color of these is an important factor for decisions within companies. On the other hand, currently, companies use different mathematical models as a solution for decision making, which, although useful and functional, only present between 60% and 65% of success in them, showing a little less than half of the failure within the decisions for the company. Consider, a car is assembled in 7 stages and this passes through 4 work stations, only the assembly of this car has as result 28 critical points, now if 3 different models are made at the same time, and what happens if 4 cars are made of each model, the number of variables and critical points of the process grow significantly (Figure 3), so the mathematical and stochastic models are not being practical enough for this type of companies, representing 40% of losses or inefficiencies in the production of final products.
Intelligent Model for the Car Assembly Industry 233
Figure 3: A multiple production of cars with multiple variables produce multiple critical points within the company.
4. Proposed Methodology The proposal to help the way to optimize resources in the supply chain of a company is the realization of an intelligent model based on Big Data, which will be the technology responsible for generating the best options to optimize the use of materials in the warehouse of a car assembly industry in north-eastern Mexico (Figure 4), as well as a great help in making decisions for the company. Once the analysis through Big Data and the best options generated are available, Fuzzy Logic Type 2 technology will be integrated to determine the best way to use the company's resources or the best decision for the company. The combination of these cutting-edge technologies would represent an improvement for many of the warehouses within the assembly industry within Mexico; this model can even be adaptable to other industries and government agencies or any business that has a warehouse and involves decision making in it since the goal of this intelligent model is to increase the optimization of resources and the effectiveness of decisions made by the company by up to 85%.
Figure 4: Use of big data for sorting and generation of options.
234 Innovative Applications in Smart Cities
Figure 5: Integration of Fuzzy Logic Type-2 for the choice of the best option.
5. Conclusion and Future Research There are many scientific articles that enable the research and development of the intelligent model to continue; it is worth mentioning that, although there are articles related to Big Data, other Fuzzy Logic Type-2, there is not much about the combination of both technologies, so it is thought that the development of a hybrid intelligent model could be a great revolution in the management of decisions and warehouses within the industry.
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form Manuel R. Piña-Monarrez1,* and Paulo Sampaio2
In this chapter, a practical and dynamic method to determine the reliability of a process (or product) is presented. The novelty of the proposed method is that it let us to use the Weibull distribution to determine the reliability index, by using only the quadratic form of the analyzed process (or product) as an input. So, since this polynomial can be fitted by using, e.g., simulation, mathematical and/ or physical modeling, empirical experimentation and/or any optimization algorithm, the proposed method can easily be implemented in several fields of the smart manufacturing environment. For example, in the industry 4.0 framework, the proposed method can be used to determine, in dynamic form, the reliability of the analyzed product, and to give instantaneous feedback to the process. Therefore, to show the efficiency of the proposed method to determine the reliability in several fields, it is applied to the design, the quality and the monitoring product phases as well as to the fatigue (wearout and aging) phase. In order to let readers adapt the given theory to their fields and/or research projects, a detailed step by step method to determine the Weibull parameters directly from the addressed quadratic form is given for each one of the presented fields.
1. Introduction Nowadays smart manufacturing (SM) is empowering businesses and achieving significant value by leveraging the industrial internet of things. Therefore, because process and products are now more complex and multifunctional, more accurate, flexible and dynamic analysis' tools are needed in the SM environment. For example, these technical tools are now being implemented into the industry 4.0 framework to evaluate and to make instantaneous feedback in the SM environment. Therefore, in this chapter a method to determine and/or to design a product or process with high reliability (R(t)) is presented. More importantly, since the proposed method is based on the Weibull distribution [1], then based on its Weibull shape parameter (β), the proposed method allows us to evaluate the reliability of the process or product in either of their principal phases; to know the design phase, which occurs for β < 1, the production phase which occurs for β = 1, and the wearout and aging phase which occurs for β > 1 [2]. Hence, due to the flexibility given by the β parameter, the proposed method can be used in the SM environment to evaluate in dynamic form the reliability of any SM process for which we know the optimal function. Universidad Autónoma de Ciudad Juárez, Av. Hermanos Escobar, Omega, 32410 Cd Juárez, Chihuahua, México. Salvador University (UNIFACS), Brazil.. * Corresponding author: [email protected] 1 2
236 Innovative Applications in Smart Cities The novelty of the proposed method is that it lets us to determine the Weibull parameters directly from the quadratic form elements of the optimal polynomial function used to represent the analyzed process (or product). Thus, since in the proposed reliability method, its input are only the elements of the quadratic form, its integration into the industry 4.0 paradigm is direct, and it will leave it to the decision maker managers to continuously determine the reliability that their processes present. On the other hand, it is important to highlight that, because the proposed method can be applied based only on the quadratic form of any optimization function, then, since this optimization function can be determined by several mathematical, physical, statistical, empirical, and simulations tools, as they can be a genetic algorithm, mathematical and physical modeling, empirical experimentation [3] and [4], finite element analysis, and so on, readers easily will be able to adapt the given method to determine the reliability in their field and/or their projects. Therefore, with the objective that everyone can adapt the given method, in Sections 2 and 3 the theorical bases on which the proposed method was formulated, the references, where a detailed explanation of the technical formulations can be found, as well as the formula to determine the Weibull scale value which let us to determine the mean and the standard deviation of the input data, are all given. And to show how the proposed method works in several different fields, its application is presented in Section 4 to the mechanical stress design field [5]. In Section 5, it is applied to the quality field analysis [6]. In Section 6, it is applied to the multivariate statistical process control field [7]. In Section 7, it is applied to the physical field by designing and performing a random vibration test analysis for both the normal and the accelerated conditions. Finally, in Section 8, it is applied to the Fatigue (wear and aging) field. Additionally, to facilitate its application to the fields or projects of the readers, in each one of the above mentioned field applications, a detailed step by step formulation to fit the Weibull parameters which represent (1) the random behavior of the applied stress, and (2) the Weibull-q parameters from which We can validate that the estimated Weibull stress distribution accurately represents the random behavior of the applied stress, are both derived. And its validation is made by demonstrating that by using the expected stress values given by the Weibull-q parameters, we can accurately derive both the mean and the standard deviation values of the Q elements from which the Weibull parameters were determined.
2. Weibull Generalities This section has the objective of presenting the characteristics of the Weibull distribution that we can use to determine its parameters directly from an observed set of lifetime data or the known log-mean and log-standard deviation of the analyzed process. The main motivations to do this are (1) the Weibull distribution is very flexible to model all life phases of a products and processes, (2) in either phase of a process (or product), as can be among others, design, analysis, improvement, forecasting or optimization, both the region which contains the optimum (minimum or maximum) and the variable levels (values) at which the process represents the optimum, must be considered, and 3) because it is always possible to model the optimal region by using a homogeneous second order polynomial model of the form Yˆ = b + b X + b X + b X X + b X 2 + b X 2 (2.1) 0
1
1
2
2
12
1
2
11
1
22
2
Therefore, because from Equation (2.1), the optimum of the analyzed process is determined from the quadratic form of the fitted optimal polynomial, we can use its quadratic form Q to determine the Weibull parameters. The quadratic form Q in terms of the interaction (bij) and quadratic effects (bjj) of the fitted polynomial [8] is given as
b Q = 1 11 2 b21
b b22
1 2 12
(2.2)
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 237
Here, it is important to notice that, when the interaction effects of Q are zero (bij = 0), the optimum occurs in the normal plane (see Figure 1), and when they are not zero (bij2 = 0), the optimum occurs in a rotated plane (see Figure 2).
Figure 1: Normal plane.
Figure 2: Rotated plane.
Thus, because the rotated plane is represented by the eigenvalues (λ1 and λ2) of the Q matrix, in any optimization process analysis when bij2 ≠ 0 both λ1 and λ2 and the rotation angle θ (see Figure 2) must be estimated, and then they are used to determine the optimal of the process. Even more, because from the eigenvalues of Q, the corresponding angle θ is unique, then the corresponding eigenvalues λ1 and λ2 are both unique also. Consequently, in this chapter, both λ1 and λ2 and θ are used to determine the corresponding Weibull shape β and scale η parameters. Therefore, because λ1, λ2 and θ are unique, β and η are unique also. On the other hand, notice because λ1 and λ2 are the axes of the rotated plane (see Figure 2), then the forms of the analyzed system before and after the rotation are different. Thus, since the normal distribution does not have a shape parameter, then the normal distribution should not be used to model the Q form when its interaction elements are not zero (bij2 ≠ 0). In contrast, also notice that because θ completely determines λ1 and λ2 (see Equation (3.4)), and they can also be determined by the logarithm of the collected data as in [9], the probabilistic behavior of Q easily can be modeled by using the Weibull distribution [1] given by f (t ) =
β η
t η
β −1
β t exp − η
(2.3)
Moreover, since for different β values the Weibull distribution can be used to model the whole life of any product or process [2], the use of the Weibull distribution to model the quadratic form Q, fitted from data of several fields, is direct. So, since β and η are both time and stress dependent parameters, then the Weibull distribution is efficient to predict through the time the random behavior of the λ1 and λ2 values of Q. The analysis to estimate β and η directly from the λ1 and λ2 values of Q is as follows.
238 Innovative Applications in Smart Cities 2.1 Weibull parameter estimation In this section, the Weibull β and η parameters are determined from a set of collected lifetimes data by using the linear form of the Weibull reliability function given by t β R = (t ) exp − η
(2.4)
Since the linear form of Equation (2.4) is of the form Y = b0 + βx
(2.5)
then the estimation of the unknown b0 and β parameters is performed by using the well-known least square method given by. ˆ = (X t X)–1 X t Y β (2.6) And, since in Equations (2.5) and (2.6), the elements of the vector Y are unknown, then in the estimation process the median rank approach [10] is used to estimate them. The steps are as follows. 2.2 Steps to estimate β and η from a set of collected lifetime data Step 1. If you are going to collect the data, then determine the desired R(n) index for the analysis. Then, based on the R(n) index, determine the corresponding sample size n to be collected [11] as
n=
−1 ln(R(t ))
(2.7)
In contrast, if you are analyzing a set of n collected data, then the R(n) index which the set of the used n data represents, is determined from Equation (2.7) by solving it to R(n). Note 1. Here, notice that in Equation (2.7) n is not being used to determine if data whether follows or not a Weibull distribution. Instead, it is being used only to collect the exact amount of data which let us accurately fit the Weibull parameters [11]. Step 2. By using the n value estimated in step 1, determine the cumulated failure percentile by using the median rank approach [10] as F(ti) = (i – 0.3)/(n + 0.5)
Note 2. Equation (2.9) is the linear form of Equation (2.4), that was defined in Equation (2.5). Step 4. From a regression between the Yi elements of step 3, and the logarithm of the collected lifetimes Xi = ln(ti) elements, determine the Weibull-q time β and ηtq values. From Equation (2.9), β is directly given by the slope, and the Weibull-q scale value is given as ηtq = exp {–b0/β}
(2.10)
The addressed β and ηtq parameters are the corresponding Weibull-q family W(β, ηtq) that represents the collected data. Step 5. From the Xi elements of step 4, determine its corresponding log-mean μx and log-standard deviation σx values, and determine the Weibull scale parameter that represents Q(x)
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 239
η
t
= exp{μx}
(2.11)
Thus, the addressed β and ηt parameters are the Weibull family W(β, ηt) that represents the related quadratic form Q(x), as shown in Section 3. At this point, only notice that ηt ≠ ηtq because while ηt is directly given by the μx value, ηtq is given by the collected data. From this section the general conclusion is that by using Equations (2.9) and (2.10) the Weibull-q time distribution which represents the collected lifetime data is determined, and by using Equations (2.9) and (2.11) the Weibull time distribution that represents Q(x) is determined. Now let us present the numerical application. 2.2.1 Practical example Here, let use data given in Table 1, that was published in [9]. The step by step analysis is as follows: Step 1. Because in Table 1 n = 21, from Equation (2.7), the reliability of the analysis is R(n)=0.9535. Here observe R(n)=0.9535 is not the reliability of the analyzed product, instead it can be seen only as the reliability confidence level used in the statistical analysis Steps 2 and 3. The F(ti), Yi and Xi elements are all given in Table 1. 1: Weibull analysis for collected lifetime data. Table 1. Weibull Analysis forTable Collected Lifetime Data Equations
Step 4. By using the Minitab routine, the regression equation is Yi = –9.074 + 1.985Xi. Hence, β = 1.985 and from Equation (2.10), ηtq =exp{–(–9.074/1.98469)} = 96.7372 hrs. Consequently, the Weibull-q distribution that represents the life time data is W(β = 1.985, ηtq = 96.7372 hrs). Step 5. Since from Equation (2.11) ηt = 73.4846 hrs, the Weibull distribution that represents the related Q(x) form is W(β =1.985, ηt = 73.4846 hrs). Finally observe, from Equations (2.4) or (2.9) the lifetime which corresponds to the expected R(t) index is given as
(
β
)
tqi = √–ln(R(t)) * ηtq = exp{Yi /β + ln(ηtq)}
(2.12)
For R(t) = 0.9535, t = 20.86 hrs. And the time that corresponds to the expected R(t) index of the related Q(x) form is given by
(
β
)
ti = √–ln(R(t)) * ηt = exp{Yi /β + ln(ηt)}
(2.13)
For R(t) = 0.9535, it is tq =15.85 hrs. From Table 1, we observe that because the mean of the lifetime data of μ = 85 hrs, and the standard deviation of σ = 43.095 hrs, were both generated by the Weibull-q family W(β = 1.985, ηtq = 96.7372 hrs), then its corresponding log-mean μx = 4.297077 was also generated by the Weibull-q family. Furthermore, using μx in Equation (2.11) gives the Weibull scale parameter of ηt =73.4846 hrs of the related Weibull time distribution that represents Q(x), then the Weibull-q family can always be used to validate the ηt parameter.
240 Innovative Applications in Smart Cities In Section 3, we will show the elements of the quadratic from which generate the ηt =73.4846 hrs value (λ1 = 127.72 and λ2 = 42.28), which corresponds to an angle of Ɵ = 29.914. However, let us first present how to estimate the Weibull time and the Weibull-q families when no experimental lifetime data is available. 2.3 Estimation of β and η without experimental data When lifetime data is not available, the Weibull parameters β and η are estimated based on the mean μy and standard deviation σy values of the median rank approached defined in Equation (2.9), and on the known log-mean μx and log-standard deviation σx values of the logarithm of the expected lifetimes. Also notice for n = 21, both μy = –0.545624 and σy = 1.175117 are constant. Based on them, the steps to estimate β and η are as follows. Step 1. By following steps 1 to 3 of Section 2.1, determine the Yi elements, and from these elements determine its mean μy and its standard deviation σy. (from data of Section 2.1.1 the μy and σy values are given in Table 1). Step 2. By using the μy and σy values of step 1 and the known log-mean μx and log-standard deviation σx values, the corresponding expected Weibull-qβ and ηtq parameters are β = σy/ σx
(2.14)
ηtq = exp{μx –Y/β}= ln(ηt) – μy/β i
(2.15)
Therefore, Equations (2.14) and (2.15) enable us to determine without data the Weibull-q W(β, ηtq) family that represents the expected failure times. And Equations (2.14) and (2.11) enable us to determine without data the Weibull time W(β, ηt) family that represents the related Q(x) form. Now let us present the numerical application. 2.3.1 Practical example By using the μy and σy, and the μx and σx values from Table 1, in Equations (2.14) and (2.15), the Weibull-q time parameters, as in Section 2.2.1, are W(β = 1.984692, ηtq = 96.736747 hrs). Similarly from Equation (2.14) and Equation (2.11), the corresponding Weibull stress parameters of the expected quadratic form are W(β = 1.984692, ηt = 73.4846 hrs). Additionally, notice from Equation (2.13) that, because σx determines the β value, and since the higher the σx value the lower the β value, then, as in [12], σx must be set as the upper control limit in the control chart used to monitor β. Similarly, because from Equations (2.11) and (2.15), μx determines the ηtq and ηt the values, and because the lower the μx value, the lower the ηtq and ηt values, then as in [12], the μx value must be set as the minimum allowed value in the control chart used to monitor ηtq and ηt. On the other hand, if there is no available experimental lifetime data, and μx and σx are unknown, then, based on the applied stresses values of the analyzed process, the Weibull-q and Weibull stress parameters are estimated as follows.
3. Weibull Quadratic Stress Form Analysis The objective of this section is to estimate the Weibull stress W(β, ηs), and the Weibull-q stress W(β, ηsq) parameters directly from the quadratic form elements of the optimal polynomial used to optimize the process. Therefore, from Equation (2.1) the quadratic form Q(s) is given by its quadratic and interaction effects as Q(s) = ∑ki,j=1 bij Xi Xj
(3.1)
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 241
And as a consequence of Equation (2.1) being the optimal response surface polynomial model widely used in the experiment design analysis, then based on the B canonical form of Equation (2.1) (see [4] Chapter 10), given by . . Yˆ = Ys + λ1X 12 + λ1 X 22 (3.2) the Q(s) matrix defined in Equation (3.1), in terms of the λ1 and λ2 values of Equation (3.2) is given as Q(s) = ∑ki,j=1 λj Xj2
(3.3)
Here, it is important to notice that (1) in this section Q(s) instead of representing time represents stress, and (2) that in the case that Q(s) has several eigenvalues, then in the analysis we have to use only the maximum λ1 = λmax and minimum (λ2 = λmin) eigenvalues. Therefore, based on the λ1 and λ2 values of stress Q(s) form, the corresponding Weibull stress β, and ηs parameters that represents Q(s) and the ηsq that represents the expected stresses values are determined as follows. 3.1 Estimation of β and η from the Q(s) matrix elements The steps to determine the Weibull-q stress and Weibull stress parameters are: Step 1. From the Q(s) matrix elements of Equation (2.2) or Equation (3.1), determine the eigenvalues λ1 and λ2 as λ1, λ2 = μ ± √ μ2 – ηs2
(3.4)
Where µ is the arithmetic mean, which because the trace of a matrix is invariant, then from the Q(s) elements, it is given as μ = (b11 + b22)/2 μ = (b11 + b22)/2
(3.5)
Then, ηs is the scale parameter of the Weibull stress family, which from the determinant of Q(s) is given as η
=s
(3.6)
2 b11b22 − b12
Step 2. By using the μy value of step 1 of Section 2.2 and the eigenvalues λ1 and λ2 of step 1, determine the corresponding β value as β=
Step 4. By using the β value of step 2, the value of step 1 of Section 2.2 and the μx value of step 3 in Equation (2.15), determine the Weibull-q stress ηsq parameter which can be used to validate the addressed Weibull stress family. Note 4: The estimated β and ηsq values are the Weibull-q stress distribution W(β, ηsq). Here remember that for n = 21, μy = –0.545624 and μy = 1.175117 are both constant. Step 5. By using the β value of step 2, the σx value of step 1 of Section 2.2, determine the expected log-standard deviation σx as σx = σy /β
(3.9)
242 Innovative Applications in Smart Cities Finally, note that because both μx and σx let us to determine the Weibull β and η parameters, and since they are given by the quadratic and interaction effects of the quadratic form Q(s), then in order to control μx and σx as in [12], the quadratic and interaction effects of Q(s), must be monitored. Or equivalently μx and σx can be used as the signal parameters in the corresponding dynamic Taguchi analysis [13] to determine the sensibility of μx and σx to the variation of the Q(s) elements. 3.2 Validation that the estimated β and η parameters represents the used Q(s) matrix data The validation is made in the sense that, by using the expected stress data given by the W(β, ηs) distribution, the eigenvalues λ1 and λ2 defined in Equation (3.4), the mean tress µ defined in Equation (3.5), and the ηs stress value defined in Equation (3.6) are all completely determined. This fact can also be seen from Equations (3.7) and (3.8) by noticing that the Weibull-q parameters are determined by using the λ1 and λ2 eigenvalues, and by noticing from Equation (3.4), that the λ1 and λ2 eigenvalues are determined by using the mean stress µ and ηs the stress value. Therefore, in order to validate in each application that the addressed Weibull family W(β, ηsq) represents the stress data from which it was determined, in the Table of each presented analysis, the expected data which corresponds to the W(β, ηsq) family is also given. From this data, observe that the average of the given data is the mean stress µ value defined in Equation (3.5), and that the exponent of the average of the logarithm of these data, is the ηs stress value. Hence, it is clear that in using these µ and ηs values in Equation (3.4), the corresponding λ1 and λ2 eigenvalues are completely determined also. Thus, the conclusion is that in using the expected data of the W(β, ηsq) family, the original µ, ηs, λ1 and λ2 parameters are all completely determined, then the W(β, ηsq) family can be used to validate the W(β, ηs) parameters which determine the random behavior of the applied stresses values given in this section as λ1 and λ2. In the next Sections, λ1 and λ2 are known as principal stresses σ1 and σ2 values (λ1 = σ1, λ2 = σ2).
4. Mechanical Field Analysis This section is focused on the design phase of a product or process. In the numerical application, the design of a mechanical element [14] is presented. The analysis is performed based on the quadratic form given by the normal and the shear stress values that are acting on the analyzed mechanical element. Therefore, the Weibull stressβ and ηs parameters are both determined based on the steps of Section 3.1 and on the stress Qs matrix given by the normal σx and σy and the shear τxy stresses values as
σ 1 τ xy Qs = τ yx σ 2
(4.1)
The steps to determine the Weibull stress W(β, ηs) and the Weibull-q W(β, ηsq) parameters are as follows. 4.1 Steps to the mechanical design field Step 1. From the stress analysis of the analyzed component, determine the normal σx and σy and the shear τxy stresses values that are acting on the element, and then form the corresponding Qs matrix as in Equation (4.1). Step 2. By using σx and σy of step 1 in Equation (3.5), determine the arithmetic mean µ. Step 3. By using the σx, σy and τxy values of step 1 in Equation (3.6), determine the Weibull stress parameter.
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 243
determine whether the designed element is safe or not. Step 7. Determine the desired R(t) index and, by using it in Equation (2.7), determine the corresponding n value. Step 8. By following steps 1 to 3 of Section 2.1, determine the Yi elements and from them determine its mean µy and its standard deviation σy. (Remember that for n = 21, µy = –0.545624 and µy = 1.175117 are both constant). Step 9. By using the σ1 and σ2 values of step 4, and the µy value from step 8 in Equation (3.7), determine the Weibull βs parameter. Note 5: The ηs parameter of step 3 and the βs parameter of this step are the Weibull stress family W(βs, ηs) which determines the random behavior of the applied stress. Step 10. By using the ηs parameter of step 3, the µy value of step 8 and the βs parameter of step 9 in Equation (2.15), determine the Weibull-q stress scale ηsq parameter. Note 6: The ηsq parameter of this step and the βs parameter of step 9 are the Weibull-q stress family W(βs, ηsq) which can be used to validate that the addressed W(βs, ηs) family completely represents the applied stress values. Step 11. Determine the R(t/s,S) index which corresponds to the yield strength value of used material mentined in step 6, as Syβs R(t/s,S) = βs (4.5) Sy + ηsβs Note 7: Equation (4.5) is the Weibull/Weibull stress/strength reliability function (see [9] Chapter 6), which is used to estimate the reliability of the analyzed component only when the Weibull shape parameter is the same for both the stress and strength distributions. Here, the Weibull stress distribution W(β, ηs) is given by Equation (3.6) and Equation (3.7), and the Weibull strength distribution is given by using Sy as the Weibull strength scale parameter. Thus, the Weibull strength distribution is W(βs, Sy = ηy). Here, remember that the R(t) = 0.9535 index used to estimate n in Equation (2.7) is the R(t) index of the analysis, and that the R(t/s,S) of Equation (4.5) is the reliability of the product. On the other hand, due to in any Weibull analysis the σ1i values given by the W(β, ηs) family, can be used as the Sy value, then the steps to determine the σ1i values that corresponds to a desired R(t/s,S) index are also given.
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 245
Step 6. Suppose after applying the modifier factors, the material' strength is Sy = 800 mpa and the safety factor is SF = 3. Hence, due to Equation (4.3) 215.2 < 266.7 and Equation (4.4) 234.3 < 266.7, the designed element is considered to be safe. See Figure 4.
Figure 4: DE and MSS theory analysis.
Step 7. Suppose a reliability analysis with R(t) = 0.9535 is desired, thus, from Equation (2.7), n = 21. (Remember that for n = 21, µy = –0.545624 and µy = 1.175117 are both constant). Step 8. The elements and its corresponding µy and σy values are given in Table 1. Step 9. From Equation (3.7), the Weibull shape parameter is βs = 1.336161. Therefore, the Weibull stress family is W(βs = 1.336161, ηs = 103.44 mpa). Step 10. From Equation (2.15) the Weibull-q stress scale parameter is ηsq = 155.8244 mpa. Therefore, the Weibull-q stress family is W(βs = 1.336161, ηsq = 155.8244 mpa). Step 11. From Equation (4.5) the designed reliability of the mechanical element is R(t/s,S) = 93.84%. Here observe the Weibull strength family is W(βs = 1.336161, Sy = 800 mpa). Step 12. The basic Weibull tan(θi) values for each one of the Yi elements are given in Table 2. Step 13. The expected pair of principal stresses σ1i and σ2i values for each one of the Yi elements are given in Table 2. Step 14. The reliability R(ti/s) values for each one of the Yi elements are given in Table 2.
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 247
From Table 2, observe that because the average of the Weibull-q family is also μ = 140 Mpa, and from its log-mean, ηs is also (ηs = exp{µx} = 103.4403 Mpa), then the addressed Weibull stress family completely represents the applied stresses. On the other hand, notice from Table 2 that the σ1i value which corresponds to R(t/s) = 0.9365 is σ1i = 800 Mpa, and in using Sy = 800 Mpa in Equation (4.5), R(t/s) is also R(t/s) = 0.9365, then we conclude that for R(t) > 0.90, the R(t/s) values of Table 2 and those given from Equation (4.5), are similar. Therefore, the σ1i column of Table 2 can be used as a guide to select the minimum yield strength scale Sy parameter which corresponds to any desired R(t/s,S) index. Moreover, from the minimum applied stress σ2 value and the Weibull stress and Weibull strength scale parameters, the minimum yield strength value which we must select from the material engineering handbook, in order for the designed product to meet the desired reliability, is given as
= S y min
σ 2η S η sη S = , S y max ηs σ2
(4.9)
For example, if the minimal applied stress is σ2 = 45.66 mpa, the Weibull stress parameter is ηs = 103.44 mpa and the Weibull strength parameter is Sy = 800 mpa, then from Equation (4.9) the minimum material's strength value to be selected from the material engineering handbook is Symin = 353.14 mpa. Similarly, the corresponding expected maximum value is Symax = 1812.36 mpa. As a summary of this section we have that: (1) The Weibull-q family W(βs = 1.336161, ηsq = 155.8244 mpa) allows us to validate that the Weibull stress family W(βs = 1.336161, ηsq = 103.44 mpa), completely represents the quadratic form Qs elements. (2) The expected σ1i elements given by the W(βs = 1.336161, ηs =103.44 mpa) family can be used as the minimum Weibull strength eta value to formulate the corresponding minimum Weibull strength family W(βs =1.336161, ηs = 800 mpa). (3) The reliability R(t/s) indices given by the W(βs = 1.336161, ηs = 103.44 mpa) family and that given by the stress/strength function R(t/s,S) defined in Equation (4.5) are both similar for higher reliability percentiles (say, higher than 0.90). Now let present the analysis for the quality field.
5. Quality Field Analysis In this section, the analysis to determine the Weibull stress and the Weibull-q stress parameters as well as the numerical application to the quality field is presented. In the quality field, the analysis of a process is generally performed in two stages. In the first, the process' output is determined in such a way that the process' output fulfills both the performance and the quality requirements. In this first stage, generally the three Taguchi's phases are applied. The Taguchi's phases (see [6] Chapter 14), are: (1) The system design phase. This consists of determining the first functional design (or prototype), and determining the process' factors and functional relationship between the addressed factors (ideal function) and the desired quality and functional requirement to be met. (2) The parameter design phase. This consists of determining the set of the significant factors, and the factor's levels at which it the process is expected to present the desired output. (3) The tolerance design phase. This consists of determining the tolerance to those process' factors which must be controlled to reach the desired process' output. In the second stage, the performance through the time of the process is determined. This is made by analyzing the effect that the environmental factors have on the process' output. Therefore, because the environmental factors' behavior is random, a probability density function (pdf) is used to model the desired process' outputs in this second stage. Thus, to determine the parameters of the used pdf, a response surface polynomial, such as the one given in Equation (2.1), is fitted from an experiment design data. Then, from its quadratic form (Qq)
248 Innovative Applications in Smart Cities elements, the corresponding pdf parameters are determined. Here, the Weibull pdf is used to perform the analysis, and the steps to fit the Weibull parameters from the Qq matrix elements are as follows. 5.1 Steps to determine the quality weibull families Step 1. From the analyzed product or process, determine the performance and quality (or functional) characteristic of the process (or product) to be measured, as well as the set of significant factors. Step 2. By using the corresponding experiment design data (here a Taguchi orthogonal array is used), determine the levels of the factors which fulfill the performance and quality requirements. Here, a capability index cp = 2, or six sigma behavior, and ability index cpk = 1.67 are used [16]. They are estimated as = cp (USL − LSL) / 6σ
cpk = min ( (USL − µ ) / 3σ or (µ − LSL) / 3σ )
(5.1) (5.2)
Where USL is the upper specification limit, LSL is the lower specification limit, µ is the process' mean, and σ is the process' standard deviation. Step 3. Determine the set of environmental factors which lower the process performance (output and quality). Step 4. By applying the response surface methodology [4], determine the optimal second order polynomial model which relates the environmental factors of step 3 and the quality characteristic of step 1. Here, notice the response surface analysis is performed only to the addressed optimal (or robust) levels of step 2. Step 5. By using the quadratic and interaction effects of the fitted response surface polynomial of step 4, form the Qq matrix as b12 / 2 b Qq = 11 b21 / 2 b22
(5.3)
Step 6. By using the b11, b22 and b12/2 elements from step 5 in Equation (3.6), estimate the Weibull stress quality ηs parameter. Step 7. By using the b11 and b22 elements from step 5 in Equation (3.5), determine the arithmetic mean µ. Step 8. By using µ from step 7 and ηs from step 6 in Equation (3.4), determine the maximum λ1 and the minimum λ2 eigenvalues. Step 9. Determine the desired reliability R(t) index to perform the analysis, and by using it in Equation (2.7), determine the corresponding sample size n value. Step 10. Following steps 1 to 3 of Section 2.1, determine the corresponding Yi elements and its mean μy and standard deviation σy. Step 11. By using the λ1 and λ2 values from step 8, and μy from step 10 in Equation (3.7), determine the quality Weibull β parameter. Step 12. By using ηs from step 6 and β from step 11, form the quality Weibull stress family W(β, ηs). Step 13. By using the β value and the Yi elements of step 10, determine the basic Weibull values tan(θi ) = exp {Yi / β }
(5.4)
Step 14. By using the basic Weibull values from step 13 and the ηs value from step 6, determine the expected eigenvalues λ1i and λ2i as
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 249
= λ1i η= η s *toi s / toi and λ2i
(5.5)
Step 15. By using β from step 11, μy from step 10 and ηs from step 6 in Equation (2.15), determine the Weibull-q parameter ηsq. The estimated β value in step 10 and the ηsq value of this step are the Weibull-q distribution W(β, ηsq). Now let us present the numerical application. 5.2 Quality improvement application In this section, we use data published in [6] (pg.452). Data is as follows. "The weather strip in an automobile is made of rubber. In the rubber industry, an extruder is used to mold the raw rubber compound into the desired shapes. Variation in output from the extruders directly affects the dimensions of the weather strip as the flow of the rubber increases or decreases." The Taguchi L8(2^7) experiment design, given in Table 3, was conducted in order to find the appropriate control factor levels for smooth rubber extruder output. The analysis is as follows. Step 1. The seven significant factors with the experimented data are given in Table 3. The required weather strip's dimension (mm) is 350 ± 35 mm. Therefore, the upper allowed limit is USL = 385 mm and the lower allowed limit is LSL=315 mm. p N 1 2 3 4 5 6 7 8
Step 12. From steps 6 and 11, the Weibull stress family is W(β = 1.984692, ηs = 73.4847 mm). Step 13. The basic Weibull elements are given in Table 8. Step 14. The expected eigenvalues λ1i and λ2i elements are given in Table 8. Step 15. From Equation (2.10), ηsq = 96.7367 mm. Therefore, the Weibull-q distribution is W(β = 1.984692, ηsq = 96.7367 mm). On the other hand, notice that, as in Table 2, from Table 8 we have that a product with strength of 257.89 presents a reliability of R(t) = 0.9206. If the product has a strength of 408.28, then it will present a reliability of R(t) = 0.9673. Finally, it is important to mention that for several control factor settings, the Weibull parameters can be directly determined from the Taguchi analysis, as in [17]. Now, let us present the principal components field analysis.
6. Principal Components Field Analysis The principal component analysis consists of determining the significant variables of the analyzed process. The selection is based on the magnitude of the eigenvalues of a variance and covariance matrix Qc [18]. Therefore, its diagonal elements contain the variance of the significant variables, and the elements out of the diagonal represent the covariance between the corresponding pair of significant variables. The Qc matrix is
σ 2 σ 1σ 2 Qc = 1 (6.1) σ 2σ 1 σ 22 On the other hand, in the case of the normal multivariate T^2 Hotelling chart, and in the case of the non-normal R-chart, all the analyzed output variables (Y1, Y2,…, Yk) must be correlated with each other. Hence, in the multivariate control field, the Qc matrix always exists. Thus, the decisionmaking process has to be performed based on the eigenvalues of the Qc matrix. However, first it is important to mention that, in the multivariate control process field, the Qc matrix is determined in such a way that it represents the process and customer requirements. Also, it is determined in the phase 1 of the multivariate control process [7]. In practice, this phase 1 is performed by using only conformant products. Therefore, the Qc matrix always represents the allowed variance and covariance expected behavior. For details of phase 1, see [7] Section 2. However, because the output process' variables are random, the Weibull distribution is used here to determine the random behavior of the eigenvalues of the Qc matrix. Now, let us give the steps to determine the Weibull stress and the Weibull-q families from the Qc matrix.
252 Innovative Applications in Smart Cities 6.1 Steps to determine the corresponding weibull parameters Step 1. For the process to be monitored, determine the set of quality and/or functional output variables to be controlled. The selected variables must be correlated each other. If they are not correlated, then use a univariate control chart to monitor each output variable in separated form. Step 2. From at least 100 products, which fulfill with the required quality and/or functional requirements, collect the corresponding functional and quality measurements of the set of response variables of step 1. Step 3. From the set of collected data of step 2, determine the variance and covariance matrix Qc defined in Equation (6.1), and the mean vector µ as µ = [ µ1 , µ2 ,..., ]
Step 6. Determine the desired R(t) index, and by using it in Equation (2.7) determine the corresponding n value. Step 7. Following steps 1 to 3 of Section 2.1, determine the Yi elements and its mean µy and standard deviation σy. Here, remember that for n = 21, µy = –0.545624 and µy = 1.175117 are both constant. Step 8. By using λmax and λmin from step 4, and from step 6 in Equation (6.4), determine the Weibull shape βc parameter as βc =
−4 µY 0.9973*ln(λmax / λmin )
(6.4)
Step 9. By using ηs from step 5 and βc from step 8, form the principal components Weibull stress family W(βc, ηs). Step 10. By using βc the parameter of step 8, and the Yi elements of step 7, determine the logarithm of the basic Weibull values as ln[tan(θi )] = {Yi / β c }
Step 4. From Matlab the eigenvalues Qc are λ1 = 342.3, λ2 = 286.2 and λ3 = 199.3. Step 5. By using λ1 = 342.3 and λ3 = 199.3 in Equation (6.3), the Weibull scale parameter is ηs = 261.1903. Step 6. The desired reliability for the analysis is R(t) = 0.9535, hence, from Equation (2.7), n = 21. Step 7. The Yi, μy and σy values are given in Table 10. Step 9. The principal component Weibull stress family is W(βc = 4.128837, ηs = 261.1903). Step 10. The logarithm of the basic tangent Weibull values is given in Table 10. Step 8. From Equation (6.4), the Weibull shape parameter is βc = 4.128837. Step 11. The basic Weibull values are given in Table 10. Step 12. The expected pair of λmax and λmin eigenvalues are given in Table 10. Step 13. From Equation (2.15), the Weibull time scale parameter is ηsq = 298.091. Therefore, the Weibull-q stress distribution is W(βc = 4.128837, ηsq = 298.091). As a summary of this section we have that, by using λmax and λmin eigenvalues, their random behavior can be determined using the Weibull distribution. Also, notice that the above Weibull analysis can be performed to any desired pair of eigenvalues of Table 10, or to any desired pair of eigenvalues of the analyzed Qc matrix. For example, following the steps above, the Weibull stress parameters to λ1 = 342.3 and λ2 = 286.2 are W(βc = 12.476153, ηs = 312.9956). And for λ2 = 286.2 and λ3 = 199.3, they are W(βc = 6.171085, ηs = 238.8298). However, notice that, by using λmax and λmin, we determine the maximum expected dispersion, as can be seen by observing that βc = 4.128837 < βc = 6.171085 < βc =12.476153). Finally notice from Table 10 that the higher eigenvalue of 595.6 has a cumulated probability of F(t) = 1–0.9673 = 0.0327 of being observed. Thus, the eigenvalues of the analyzed process should be monitored as in [21]. Now let us present the Vibration field analysis.
7. Vibration Field Analysis In this section, the Weibull parameters which represent random vibration data are determined. Data corresponds to a sprung mass product attached to a car. The testing profile used is the one given in the test IV of the norm ISO16750-3. In the analysis, the following [22] three assumptions are made. (1) The first row in the testing's profile represents the most severe scenario in the test. (2) The whole test's profile is considered as a compact cycle which is repeated n times until failure occurs. (3) The generated damage in one cycle is cumulated from the first row of the testing's profile to the last.
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 255
256 Innovative Applications in Smart Cities Based on these assumptions, the steps to determine the corresponding vibration Weibull stress and Weibull testing time families are as follows. 7.1 Steps to determine the weibull families Step 1. For the analyzed product, determine its location in the car and the norm to be used. And based on the location, determine the testing's type to be applied. Here, the ISO16750-3 norm [23] is used. Step 2. From the selected testing's type, determine the testing's profile parameters; frequency (Hz), energy ((rms^2=m/s^2)^2/Hz), and testing time t (hrs). Step 3. Determine the desired reliability R(t), and from Equation (2.7) determine the corresponding sample size n to be tested. Step 4. Following steps 1 to 3 of Section 2.1, determine the Yi elements, its mean μy and its standard deviation σy. Here, remember that for n = 21, μy = –0.545624 and μy = 1.175117 are both constant. Step 5. Take the product of the applied frequency and energy from the first row of the testing's profile of step 2 as the minimum eigenvalue
λmin = f1 * G1
(7.1)
Step 6. Take the total cumulated energy as the maximum eigenvalue, given by
where APSDi is the applied energy and fi is the frequency of the i-th row of the used testing's profile, f(i-1) is the frequency of the (th-1)-row of the testing's profile, and m is the slope given as m = dB/octaves
(7.4)
dB = 10 log( APSDi / APSDi−1 )
(7.5)
octaves = log( fi / fi−1 ) / log(2)
(7.6)
where
Step 7. By using the μy value from step 4, and the addressed λmin and λmax values from steps 5 and 6, determine the corresponding Weibull vibration βv parameter as β
v
=
−4 µY 0.9947 * ln(λmax / λmin )
(7.7)
Step 8. By using the testing's time of step 2, the R(t) index from step 3, and the βv value from step 7, determine the corresponding Weibull-q scale parameter as
ηtq =
ti ln(− ln(R(t ))) exp βv
Note 8. The Weibull testing time family is W(βv, ηtq).
(7.8)
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 257
Step 9. By using the square root of the Ai value of step 6, the reliability index from step 3, and the βv value from step 7, determine the corresponding Weibull stress vibration scale parameter as
ηs =
Ai ln(− ln(R (t ))) exp βv
(7.9)
Note 9. The addressed Weibull stress family is W(βv, ηtq). Step 10. By using the βv parameter of step 7, and the Yi elements of step 4, determine the logarithm of the basic Weibull values as ln[tan(θi )] = {Yi / β v }
(7.10)
Step 11. From the logarithm of the basic Weibull values of step 10, determine the corresponding basic Weibull values
tan(θi ) = exp {Yi / β v }
(7.11)
Step 12. By using the basic Weibull values of step 11 and the ηtq parameter from step 8, determine the expected testing times as
Now let us present the application. 7.2 Vibration application In the application, the vibration test IV given in the ISO16750-3 norm is used. The ISO16750-3 norm applies to electric and electronic systems/components for road vehicles, and because in the vehicle the vibration stress can occur together with extremely low or high temperatures, the vibration test is generally performed with a superimposed temperature cycle test. For details of temperature cycle test see the ISO16750-4 norm and the appendices F and G of the guide of the norm GMW3172 [24]. The numerical analysis is as follows. Step 1. The analyzed product is a sprung mass mounted in a passenger car. Therefore, the testing's type to be applied is the random vibration test IV given in the ISO16750-3 norm Section 4.1.2.4. Step 2. For test IV, the product must be tested by 8 hrs in each one of the X, Y and Z directions. Thus, the total experimental testing time is t = 24 hrs. The testing frequencies and their corresponding applied energy are given in Table 11. And the corresponding testing's profile is plotted in Figure 5. Step 3. The desired reliability to be demonstrated is R(t) = 0.97. Therefore, from Equation (2.7) n = 32.8308 ≈ 33 parts must be tested. Table 11: Random vibration profile. Freq(Hz) ASD(G2/Hz) dB Oct dB/Oct Area Grms 10.00 30.0000 * * * 300.00 17.32 400.00 0.2000 -21.76 5.32 -4.09 614.00 24.78 1000.00 0.2000 0.00 1.32 0.00 734.00 27.09
258 Innovative Applications in Smart Cities
Figure 5: Testing Vibration Profile.
Step 4. The Yi, μy and σy values are given in Table 12. Step 5. From the first row of Table 11, the minimum eigenvalue is λmin = 300 rms^2. Step 6. From Equation (7.3) the maximum eigenvalue is λmax = 734 rms^2. Step 7. By using the μy value from step 4, and the addressed λmin and λmax values from steps 5 and 6 in Equation (7.7), the Weibull vibration parameter is βv = 2.5. Step 8. From Equation (7.8), ηvt = 96.9893 hrs. Therefore, the Weibull testing time family is W(βv = 2.5, ηvt = 96.9893 hrs). Step 9. From Equation (7.9), ηvs = 11.1538 Grms. Therefore, the Weibull stress vibration family is W(βv =2.5, ηvs = 11.1538 Grms). Step 10. The logarithm of the basic Weibull data is given in Table 12. Step 11. The basic Weibull data is given in Table 12. Step 12. The expected testing times are given in Table 12. Step 13. The expected vibration levels are given in Table 12. From Table 12, the row for t = 24 hrs and vibration level of 2.76 Grms, on which the analysis was performed, was added. The row between rows 5 and 6 was added to show that the above analysis can be used as an accelerated life time analysis [24]. For example, from this row, we have that R(t) = 0.97 can also be demonstrated by testing 6 parts at constant vibration level of 2.76 Grms, each for 47.366 hrs (we must test each part in the X, Y and Z axes for 47.366/3 = 15.7885 hrs each). This is true due to the fact that the total Weibull testing time (Ta) is the same for any ni and ti elements of Table 12. The total testing time is given as
Ta = ni *tiβ v
(7.14)
And since Ta is the same for any row of Table 12, then the Weibull testing time scale parameter is also the same for any ni and ti elements of Table 12. And since it is
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 259 Table 12: Weilbull testing time and in vabration analysis.
On the other hand, because the relation between the applied vibrations Si and the testing times ti for any two rows of Table 12 always holds, from the relation
Saccel tnorm (7.17) = Snorm taccel we can use the Si column of Table 12 as the accelerated vibration level for any desired sample size value. For example, we can demonstrate R(t) = 0.97 by testing 6 parts for 24 hrs each at constant vibration level of 5.447 Grms. (we must test each part in the X, Y and Z axes by 8 hrs each). The accelerated test parameters and its corresponding testing's profile are given in Table 13 and in Figure 6.
For deeper Weibull/vibration analysis see [24]. Now let us present the relations between the Weibull parameters and the cycling formulas which can be used to perform the corresponding fatigue analysis.
8. Weibull Fatigue Formulation In this section, the objective is to present the formulas which can be used to perform a fatigue analysis [25]. The analysis is made by considering that the applied stress has the general cyclical behavior given in Figure 7. The given formulas are derived from the Weibull/stress theory given in [9]. Hence, readers interested in deeper analysis are invited to consult [9]. In the analysis, the stress data given in Section 4.2 is used to present the numerical analysis. The steps to determine the Weibull stress and Weibull time parameters for fatigue analysis are as follows. 8.1 Steps to determine the weibull stress and time fatigue analysis Step 1. From the applied normal σx and σy and shear τxy stresses values, determine the arithmetic mean stress as μ
= (σx + σx)/2
(8.1)
Step 2. By using the applied normal σx and σy and shear τxy stresses from step 1, determine the fatigue Weibull stress parameter as η
f
= √σx σy – τ 2xy
(8.2)
Step 3. By using the mean µ and the ηf values from steps 1 and 2, determine the maximum (σ1) and the minimum (σ2) principal stress values as σ1, σ2 = µ ± √µ2 – ηf2
Step 5. Following steps 1 to 3 of Section 2.1, determine the Yi elements and its mean μy and standard deviation σy. Step 6. By using the μy value from step 5 and the principal stress values from step 3 in Equation (3.7) determine the corresponding Weibull shape parameter β. Note 10: The ηf parameter from step 2 and the β parameter from this step are the Weibull stress family W(β, ηtf). Step 7. By using μy from step 5 and the β value from step 6 in Equation (2.10), determine the Weibull time scale parameter ηtf. Thus, the Weibull time family is W(β, ηtf). Step 8. By using the β parameter from step 6 and the Yi elements from step 5, determine the logarithm of the basic Weibull elements as
ln[tan(θi )] = {Yi / β }
(8.5)
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 261
From Table 14, we observe that (1) the reliability of the addressed principal stresses σ1 = 234.34 mpa and σ1 = 45.66 mpa is R(t) = 0.7142. However, notice that this reliability corresponds to a component that presents a Weibull strength parameter with ηs = 234.34 mpa. Therefore, if the Weibull component's strength is of ηs = 1332.5045 mpa, then its reliability is R(t) = 0.9673. (2) The cyclical stress behavior is shown in Figure7, in Table 14 it is given between rows 6 and 21. Thus, as it is shown from row 1 to row 6 of Table 14, it is expected that higher stresses values will occur. For example, a stress of 1332.50 mpa will occur with cumulated probability of F(t) = 1–0.9673 = 0.0327. (3) From the columns σ1i and σ2i we have that the addressed Weibull stress distribution, W(β = 1.3316, ηf = 103.44 mpa) completely models the principal stresses random behavior. Also, since from [9] the Weibull distribution can be represented by a circle centered in the arithmetic mean µ, the Weibull distribution can effectively be used to model fatigue data, as in Figure 7, and to cumulate the generated damage as in [26]. (4) As can be seen from Table 14, since the Weibull stress scale ηf parameter always occurs at an angle of 45°, and due to the ratio between ηf and σ1 and σ2 being the same for both principal stresses as = R
σ1 η f = η f σ2
(8.11)
the Weibull analysis given in Table 14 can be used to perform the corresponding modified Goodman diagram to determine the threshold between finite and infinite life. (5) From Equation (8.11), by using the λi column of Table 14 as the Weibull strength scale parameter ηs, we can determine the minimum material's strength (say the yields Sy value) which we must select in order for the designed component to present the desired reliability. Based on the selected ηs value, the minimum and maximum strength Sy values to be selected from an engineering handbook are given as
Weibull Reliability Method for Several Fields Based Only on the Modeled Quadratic Form 263
Figure 7: Weilbull/Mohr circle representation.
= S y min
η f ηS σ 2η S = , S y max ηf σ2
(8.12)
For example, suppose we want to design an element with a minimum reliability of R(t) = 0.9673, thus, from Table 14, the corresponding value is ηs = 1332.5045 mpa, and since Table 14 was constructed with ηf = 103.44 mpa, then, by using these values with the minimal stress of σ2 = 45.66 mpa in Equation (8.12), the minimum material's strength to be selected from a engineering handbook is Sy = 588.1843 mpa.
9. Conclusions (1) When a quadratic form represents an optimum (maximum or minimum), its random behavior can be modeled by using the two parameter Weibull distribution. (2) In the proposed Weibull analysis, the main problem consists of determining the maximum and the minimum stress (λ1 and λ2) values that generate the failure. However, once they are determined, both the Weibull stress and the Weibull time families are determined. Therefore, the stress values used in Equation (3.7) must be those stresses values that generate the failure. Here, notice that the constant 0.9947 value used in Equation (3.7) was determined only for the given application. The general method to determine it for any λ1 and λ2 values is given in [9] Section 4.1. (3) The columns σ1i and σ2i are the maximum and minimum expected stress values which generate the failure, and thus, σ1i represents the minimum Weibull strength value that the product must present to withstand the applied stress. Therefore, column σ1i can be used as a guide to select the minimal strength material as it is given in Equations (4.9) and (8.12). (4) From Table 12, the columns R(ti), ni, ti, and Si can be used to design any desired accelerated testing scenario. For example, suppose we want to demonstrate R(ti) = 0.9490, then from Table 12 we have that by fixing the testing time t = 24 hrs, we can test n = 19.143 (testing 18 parts by 24 hrs each one and one part by 1.143*24 hrs) at a constant stress of Si = 3.425 Grms. (5) In any Weibull analysis, the n value addressed in Equation (2.7) is the key variable in the analysis. This because for the used β value it always let us to determine the basic Weibull elements as tan(θi ) = 1/ ni1/ β . This fact can be seen by combining Equation (2.7) and Equation (4.8) or directly from Equations (43) and (53) in [9].
264 Innovative Applications in Smart Cities (6) From the applications, we have that although they appear to be very different, because all of them use a quadratic form in their analysis, they can all be analyzed by using the Weibull distribution. Generalizing, we believe that the Weibull distribution can always be used to model the random behavior of any quadratic form when the cumulated damage process can be modeled by using an additive damage model, such as that given in [27]. When the damage is not additive, then the log-normal distribution, which is based on the Brownian motion [28], could be used. (7) Finally, it is important to mention that by using the maximum and minimum applied stresses, the given theory could be used in the contact ball bearing analysis [29] to determine the corresponding Weibull shape parameter and to determine the corresponding Weibull stress scale parameter from the equivalent stress and/or from the corresponding dynamic load as it is given in [30]. Similarly, the Weibull time scale parameter can be determined from the desired L10 life proposed by [31]. | eng | e691ea2e-69df-40a7-990c-1576f1ed71e1 | https://ebin.pub/innovative-applications-in-smart-cities-036782096x-9780367820961.html |
Propolis For Children, Help Them Breath
Upper respiratory-tract infections (URTI) occur most commonly in childhood. On average, a healthy three-year old child suffers from 6-10 colds per year. These types of colds are generally mild but the symptoms can cause a fever and make your child irritable, lethargic and uncomfortable. According to the world health organization, respiratory infections account for 20% of medical consultations, 30% of days lost from work and 75% of antibiotic prescriptions in the USA, Canada, and Western Europe. But how do our children end up with these respiratory infections? Viral infections in childhood can come from many things these include attendance at daycare centers, contact with elder siblings, smoking at home and lack of breast feeding. So what's the first thing we do when our baby has an awful cough? We head down to the local pharmacy and get a good non-prescriptive children's cough and cold medicine. But these are not always effective in reducing the recovery time and can often have adverse effects including toxicity with unintentional overdose if we aren't careful. More is not always better when it comes to cough medicines. So what should we do to stop this and keep our children healthy? Do your research!
This isn't the first time I've talked about Complementary and Alternative Medicines (CAMs) and it won't be the last. CAMs are not taught widely in medical schools but they are products that are used in addition to or as a replacement for traditional western medicine. Why are we bringing this up? Because studies have been done on what herbal medicines may be a good CAM to help support our children's lungs and keep them breathing well throughout the school year. Many of the studies have looked at Echinacea and Vitamin C but today we are going to look at what research says about propolis.
In a study done in Romania they found that the treatment of upper respiratory infections with a tincture containing propolis decreased and sometimes suppressed viral activity in the upper airways. It does this through its anti-viral anti-inflammatory properties known as flavones. Other clinical results supporting its effectiveness in the treatment and prevention of upper respiratory tract infection have been published in Germany and only from trials among adults. But we found a study that took place at 10 primary care pediatric community clinics in Israel. In this study there were 430 children age 1-5 and each was randomly assigned a propolis elixir or a placebo. The results of this study demonstrate, for the first time, that an herbal preparation containing an extract of propolis did have a significant beneficial effect on the incidence and severity of respiratory tract infections in young children. In another study by Crisan et al a lower incidence of URTIs and symptoms were found in a group of preschool children treated with a propolis preparation for 5 months. Additional studies are needed in larger samples to confirm these findings but this has hopefully opened the door to propolis being used as in clinics. In the mean time check out PRI's Propolis and Manuka honey cough elixir or the PRI Children's Propolis and Manuka Honey Lollipops to help keep your kids healthy and well.
***As with all bee products please make sure your child does not have any allergies before giving them these supplements. If you are unsure check with your doctor** | eng | 5a0fc0dd-605f-4070-a693-8e64b4ac03f6 | https://www.shoppri.com/blogs/health-wellness/propolis-for-children-help-them-breath |
Treasury Software
Table of Contents (Quick Links)
Listen
Overview:
Treasury software is a technology solution designed to manage cash and financial operations of businesses in a seamless manner. It is critical for businesses to have a clear understanding of their financial position, and treasury software makes it easier for them to track and manage their financial activities effectively. In this article, we will examine the different features of Treasury software, its users, benefits, as well as specific examples of software products that can aid financial operations management.
Who uses Treasure Software?
Enterprise level businesses: This software application is typically employed by a variety of businesses including multi-national corporations with complex and large cash management needs. It is also suitable for mid-sized companies and small businesses. The software provides financial control and real-time cash visibility for large cash transactions.
Benefits of Treasury Software
Real-time Cash Flow Analysis: With Treasury software, businesses can access real-time accounts of their financial position, helping them to make informed decisions about cash management. The software provides up-to-the-minute financial data that can be used to make informed decisions about investments, cash management, and forecasting.
Budgeting and Forecasting Capabilities: Treasury software has the ability to analyze historical financial data as well as current trends to create forecasts and budgets for future financial periods. The software can estimate future cash flows in advance, helping businesses to plan their cash requirements, safeguarding against potential cash shortages.
Efficient Cash Management: The software aids in streamlining business cash management processes, which helps reduce the required amount of manual labor and the traditional errors that accompany them. With an efficient treasury system in place, businesses can enjoy significant cost savings and improved profitability.
Features of Treasury Software
Cash Position Management: This feature helps in creating a more efficient account reconciliation process, ensuring that there are no discrepancies in account balances. It provides insight into the cash inflow and outflow of the business, helping to manage cash efficiently.
Liquidity Management: The liquidity management feature enables businesses to know when they have excess cash that can be invested to generate additional revenue. This feature optimizes the business' cash position, ensuring that their funds are used in the most effective way possible.
Fraud Detection: Treasury software detects any discrepancies in the business's financial transactions, highlighting potential fraud and allowing for early detection and prevention measures.
5 Examples of Relevant Treasury software Products
1. Kyriba (Kyriba.com): Kyriba is a cloud-based treasury management solution that provides enhanced cash management, supply chain finance, and more. It provides treasury management solutions for both small and large organizations, ensuring access to real-time cash management data, enhancing cash visibility, and managing risk more efficiently. The software is designed to meet various treasury management needs and is highly scalable.
2. Reval (IHSMarkit.com/reval): Reval is a treasury management software solution that provides unparalleled cash management, risk management, and strategic cash planning capabilities. It is also ideal for managing currency risk, enabling businesses to track their cash position against the real-time valuation of significant currencies. It is highly customizable to meet specific business needs.
3. FIS (FISGlobal.com): FIS is a comprehensive treasury management software that helps organizations manage their assets, liabilities, and investments efficiently. It is scalable and can be used by small and large organizations with diverse treasury management needs. With this software, businesses can derive accurate cash forecasting and real-time cash visibility.
4. SmartStream(Treasury-Solutions.com): The software provides advanced modeling features and risk analysis tools for managing different aspects of treasury management efficiently. It is a highly scalable solution that can be deployed across several company locations.
5. TreasuryXpress(Treasuryxpress.com): The TreasuryXpress software provides businesses with real-time access to the financial data needed to enhance cash management processes. This design is ideal for small and mid-sized businesses. It enables users to extract data for analysis, reporting, and benchmarking purposes.
How to use Treasury Software?
To use treasury software, businesses must take the following steps:
1. First, they must invest in the appropriate treasury software that meets their specific requirements.
2. Next, the software must be installed and configured according to the user's needs.
3. Once set up, they require intensive training for all users on the software.
4. Finally, the software is ready to use, and users can input their financial data to start monitoring and analyzing current and future cash flows' performance.
Drawbacks and Limitations of Treasury Software
One of the most significant drawbacks of treasury software is the potential for inaccurate data input. This could lead to incorrect forecasts which can hurt businesses financially. It would help if you balanced the benefits against the disadvantages and put measures in place to prevent errors.
Conclusion:
In conclusion, treasury software is a valuable tool for businesses of all sizes with significant cash management needs. It offers a wide range of features and benefits that help streamline financial processes, manage risk, and reduce errors. Businesses should look into specific treasury software products that would suit their individual needs by checking out features, scalability, and budget. | eng | 0b9b5a25-429b-46e2-80f3-c83eced4b8fb | https://biztech.dev/treasury-software-answers/ |
Italian cuisine is well-known for its variety of cheeses. Among them, Parmigiano Reggiano and Parmesan are two of the most popular and beloved. However, the question remains: what's the difference between these two kinds of cheese? In this article, we'll explore the history, production process, taste, nutritional value, culinary uses, price, and availability of Parmigiano Reggiano and Parmesan to uncover their unique differences.
Production Process
Parmigiano Reggiano
Origin and regulations
Parmigiano Reggiano is a hard, granular cheese that originates from the Italian regions of Parma, Reggio Emilia, Modena, Bologna, and Mantua. The production process of Parmigiano Reggiano is regulated by a consortium, and it must follow strict rules to be considered authentic. For example, the cheese must be made from the milk of cows that are raised in the production area and fed only local fodder. Furthermore, the cheese can only be made in copper vats, and it must be aged for at least 12 months.
Milk and curdling process
To make Parmigiano Reggiano, the milk is heated and then mixed with rennet, an enzyme that causes the milk to curdle. The curds are then cut into small pieces and placed in molds, where they are pressed to remove any remaining whey. After this, the cheese is soaked in brine to add flavor and prevent mold growth.
Aging process
The final step in the production process of Parmigiano Reggiano is aging. The cheese is left to age for at least 12 months, but some varieties are aged for up to 36 months. During this time, the cheese develops its distinct flavor, texture, and aroma.
Parmesan
Origin and regulations
Parmesan is a cheese that comes from the same region as Parmigiano Reggiano, but it is not regulated in the same way. While Parmigiano Reggiano is protected by a consortium and can only be made in a specific area, Parmesan can be made anywhere in the world. Furthermore, there are no rules regarding the type of milk that can be used to make Parmesan, and it does not have to be made in copper vats.
Milk and curdling process
The process of making Parmesan is similar to that of Parmigiano Reggiano. Milk is heated and mixed with rennet to curdle the milk, and the curds are cut into small pieces and placed in molds. The cheese is then pressed to remove any remaining whey and soaked in brine.
Aging process
Parmesan is aged for at least 12 months, but some varieties can be aged for up to 48 months. During this time, the cheese develops its flavor, texture, and aroma.
Taste and Flavor
Parmigiano Reggiano
Flavor profile
Parmigiano Reggiano has a nutty and fruity flavor with a subtle sweetness. It has a complex flavor profile, with notes of caramel, butter, and salt. The longer the cheese is aged, the more intense the flavor becomes.
Texture
Parmigiano Reggiano has a hard, granular texture. It is dry and crumbly, but it can also be slightly grainy.
Aroma
Parmigiano Reggiano has a distinctive, pungent aroma. It is rich and slightly earthy, with hints of hay and nuts.
Parmesan
Flavor profile
Parmesan has a slightly sweeter and more delicate flavor than Parmigiano Reggiano. It has a nutty, buttery, and slightly fruity taste. However, the flavor of Parmesan can vary greatly depending on where it is produced and how long it is aged.
Texture
Parmesan has a hard and crumbly texture. It is slightly grainy, but it can also be smooth and creamy.
Aroma
Parmesan has a mild, nutty aroma. It is not as pungent as Parmigiano Reggiano and has a more subtle scent.
Nutritional Value
Parmigiano Reggiano
Calories and macronutrients
Parmigiano Reggiano is a relatively high-calorie cheese, with around 110 calories per ounce. It is also rich in protein, with around 8 grams per ounce, and low in carbohydrates.
Vitamins and minerals
Parmigiano Reggiano is a good source of calcium, with around 330 mg per ounce. It also contains other important vitamins and minerals, such as vitamin A, vitamin B12, and zinc.
Parmesan
Calories and macronutrients
Parmesan is similar to Parmigiano Reggiano in terms of its macronutrient profile. It has around 110 calories per ounce and is high in protein and low in carbohydrates.
Vitamins and minerals
Parmesan is also a good source of calcium, with around 310 mg per ounce. It contains other important vitamins and minerals, such as vitamin A and vitamin B12.
Culinary Uses
Parmigiano Reggiano
Grating
Parmigiano Reggiano is commonly grated over pasta dishes, risotto, and salads. It can also be used as a topping for soups and stews.
Cooking
Parmigiano Reggiano can be used in a variety of recipes, including meatballs, meatloaf, and casseroles. It can also be used as breading for chicken or fish.
Parmesan
Grating
Parmesan is also commonly grated over pasta dishes, risottos, and salads. It is often used as a topping for pizza.
Cooking
Parmesan can be used in a variety of recipes, including lasagna, pesto, and vegetable dishes. It can also be used as a garnish for soups and stews.
Price and Availability
Parmigiano Reggiano
Price
Parmigiano Reggiano is generally more expensive than Parmesan, due to the regulations surrounding its production.
Availability
Parmigiano Reggiano is widely available in specialty cheese shops, supermarkets, and online.
Parmesan
Price
Parmesan is generally less expensive than Parmigiano Reggiano, due to the lack of regulations surrounding its production.
Availability
Parmesan is widely available in supermarkets and online. It can also be found in specialty cheese shops, but the quality can vary greatly.
Conclusion
While Parmigiano Reggiano and Parmesan may seem similar at first glance, there are significant differences between the two. Parmigiano Reggiano is a protected cheese that can only be made in a specific region of Italy, while Parmesan can be made anywhere in the world. Parmigiano Reggiano is also generally more expensive and has a more complex flavor and aroma than Parmesan. However, both kinds of cheese are delicious and versatile, and they can be used in a variety of culinary applications. Whether you choose Parmigiano Reggiano or Parmesan, you can be sure that you are getting a high-quality cheese that will add depth and richness to your favorite dishes.
When choosing between Parmigiano Reggiano and Parmesan, it is important to consider the differences in flavor, texture, and aroma. While both kinds of cheese are delicious, they have distinct characteristics that can make a difference in certain recipes. For example, Parmigiano Reggiano may be a better choice for a pasta dish, where its complex flavor can shine, while Parmesan may be a better choice for a pizza, where its mild flavor won't overpower the other toppings.
It is also important to consider the nutritional value of these cheeses. Both Parmigiano Reggiano and Parmesan are high in protein and calcium, making them a good addition to a healthy diet. However, due to their high-calorie content, it is important to enjoy them in moderation.
When it comes to price and availability, Parmesan is generally less expensive and more widely available than Parmigiano Reggiano. This can make it a good option for budget-conscious shoppers or those who don't have access to specialty cheese shops.
In conclusion, whether you prefer Parmigiano Reggiano or Parmesan, there is no denying that these cheeses are some of the most delicious and versatile ingredients in the culinary world. With their complex flavors, rich textures, and impressive nutritional profiles, they are a must-have in any home cook's pantry. So go ahead and experiment with these two delicious kinds of cheese in your next recipe, and see for yourself which one you prefer | eng | 0987ce85-bfe7-4871-8d3a-befc0211dc55 | https://kitchenguides.org/parmigiano-reggiano-vs-parmesan/ |
Can PEX Tubing Be Buried? Guidelines, Differences, And Specs (2023)
Cross-linked polyethylene, also known as PEX, is a lightweight and durable polyethylene thermoplastic with cross-links. This tubing is typically used for domestic or building service pipes, hydronic systems, and insulation for electrical cables. Nowadays, PEX tubing is quickly replacing traditional copper and galvanized steel pipes and is an exceptional option for underground piping.
PEX tubing is approved for direct burial outdoors. PEX has a life expectancy of 40-50 years, made possible by its durable and non-corrosive materials. Burial also reduces sun exposure, which would otherwise degrade the piping. Additionally, PEX tubing materials allow for expansion, making them resistant to freezing/cracking underground.
My brother-in-law and I have used PEX tubing in our last three homes including buried line to the underground water main. There's a few key things you need to understand so let's dig into PEX tubing and fittings in terms of what type works best underground.
What Type of PEX Can be Buried?
Although you can generally bury PEX tubing, some tubing types will have a longer lifespan than others underground, especially when used to transport water in freezing conditions (source).
PEX-B is the first choice for burying due to its higher bursting pressure and reduced issues with leached chemicals.
Some individuals are wary of using any type of PEX tubing outdoors because it degrades under UV light. This means that if the tubing is exposed to the sun, it will break down much faster. Since PEX tubing isn't the cheapest pipe option, costing anywhere between $1.75 and $7.50, a severely shortened lifespan is not ideal.
Luckily, because the PEX tubing is being buried in this scenario, it won't degrade from UV light as long as you limit its exposure to the sun prior to installation. Embedding the tubing in the sand once it is buried can help prevent damage from rocks and soil. Bury the tubing below the frost line to prevent water from freezing in the pipes.
That being said, here are the three types of PEX tubing, two of which are ideal options for underground systems.
PEX-A
Compared to the other two PEX tubing options, PEX-A is manufactured using peroxide, and it is the most flexible and softest tubing with the highest degree of cross-linking for the strongest chemical bond. This is the second-best PEX tubing option for burial, behind PEX-B.
The benefits of using PEX-A tubing in this scenario are that it is a non-corrosive tubing as opposed to more common materials, such as copper or steel, which inhibits the likelihood of the PEX tubing leaking or contaminating the water supply.
However, PEX-A tubing is more likely to leech chemicals from the manufacturing process into the water supply than PEX-B, so be aware there is still some level of risk with this product (source).
PEX-A tubing is suitable for burial because its flexibility allows the tubing to expand, particularly when subjected to freezing water allowing a high degree of resistance to cracking in frigid temperatures.
Although the PEX-B tubing is not as flexible as PEX-A, it has a higher bursting pressure, which means, when exposed to colder temperatures, it will withstand these temperatures longer without bursting than the PEX-A.
Ultimately, PEX-A is much easier to work with and can moderately withstand freezing conditions when buried, but it is not the best choice for burial over the PEX-B tubing.
PEX-B
If you are truly set on burying a length of PEX tubing, you should opt for the PEX-B tubing as your primary choice. Not only is this the most affordable PEX tubing option with an average cost of $2.50 to $5.50 per 10-foot piece, but it also has all the benefits of the PEX-A tubing and few drawbacks.
We've already mentioned that the PEX-B tubing has a higher bursting pressure than PEX-A, which is optimal when burying tubing that might be subjected to freezing temperatures. In addition to its non-corrosive material and reduced chances of chemical leeching, this feature makes PEX-B a more suitable choice.
(Video) PEX Pipe Installation - ALL Connections To Fittings In A WHOLE HOUSE!
In terms of differences and drawbacks, PEX-B differs from PEX-A in that it is manufactured using a moisture-cure method. This tubing is much less flexible than PEX-A, rendering it more difficult to work with.
Additionally, if your PEX-B tubing gets a kink while underground, you will have to dig up the tubing, cut out the kinked section, and replace this section with additional tubing and fixtures.
With the PEX-A tubing, many people point out that you can skip this process by fixing the kink by applying heat although it's not recommended due to the precise heating required (source).
Altogether, PEX-B tubing might be less flexible and has a lower cross-link ratio than PEX-A, for the purposes of burying the tubing, its increased bursting pressure and cost-effective pricing make it the best choice.
This YouTube video from Williams Plumbing and Heating gives an outstanding overview of the differences of PEX-A and PEX-B as well as covering the key advantages and disadvantages of each:
PEX A vs PEX B: Pros and Cons 🤔
PEX-C
PEX-C is manufactured using an irradiation method and is ultimately not suitable for burial. It is, by far, the most difficult PEX tubing to work with because it is incredibly stiff. This means the tubing is more prone to kinking and cracking when exposed to freezing conditions.
(Video) How To Install PEX Plumbing
Since the goal is to bury the tubing, using PEX-C would most likely require frequent repairs due to damage and bursting.
The sole benefit of PEX-C tubing is that it is the cheapest option, with an average cost of $1.75 to $3.50 per 10-foot piece. However, the amount you would spend on repairs if you used this tubing would significantly outweigh what you saved purchasing PEX-C over PEX-B.
Ultimately, if you want to use PEX-C, it is best suited for short repairs where you need a straight and sturdy pipe, not for lengthy underground piping.
PEX-A
PEX-B
PEX-C
Pro
Highest Flexibility
Highest bursting pressure
Most environmentally friendly
Con
Potential for chemical leaching
Stiff and prone to coil memory
Prone to cracks
Suitable for Underground
Yes
Yes
No
Can PEX Crimp Fittings be Used Underground?
PEX-A and PEX-B tubing are both exceptional options for underground piping. However, these pipes' effectiveness means very little if the fittings used cannot withstand the same conditions.
Crimp fittings are a popular option when installing piping, but they are not optimal for underground piping.
There are two common PEX fittings, clamps and crimps. The PEX crimps are often made of copper, which is the primary reason they are not conducive to burial.
The biggest concern when burying tubing and fittings is how extreme temperatures, particularly freezing conditions, affect their overall functionality.
In the case of copper PEX crimps, a common occurrence when exposed to freezing water is that it will cause the ring to expand, and then, over time as temperatures rise, this expansion will result in a leak.
Alternatively, the best PEX fitting for underground use is the stainless-steel clamp ring resistant to corrosion and more suitable for direct burial applications.
My brother-in-law likes to plan out this process and prevent any joints underground that we can, thereby reducing the potential for issues.
Always remember that anywhere you are joining pipe with fittings you are introducing a potential point of failure. Keeping those to a minimum in buried runs will reduce the opportunity for issues.
(Video) PEX A vs PEX B: Pros and Cons 🤔
Can Sharkbite Fittings be Used Underground?
As opposed to PEX crimp fittings, Sharkbite fittings are suitable for burial as professed by the company and supported by The International Association of Plumbing and Mechanical Officials (source).
SharkBite uses a lead-free brass push-to-connect fitting with an EPDM O-ring and a stainless steel 316 grab ring. These materials ensure the fitting is durable, long-lasting, and won't contaminate your water supply. The fitting is also compatible with a series of pipes, including PEX (source).
Although the material make-up of the fitting is beneficial to underground conditions, the real reason Sharkbite fittings are conducive to burial is that they are wrapped in silicone tape. Sharkbite guarantees the silicone tape won't break down, and the fittings are wrapped in this material for two reasons.
Protect the fitting from chemicals and soil contaminants.
Helps ensure the company's 25-year warranty.
The quality materials of the lead-free brass fitting and the added insurance of the silicone tape render these fittings an exceptional choice to accompany underground PEX tubing.
This YouTube video is a great overview of using Sharkbite fittings for buried PEX connections.
PEX to PVC: Sharkbite Fittings and Silicone Wrap..
Final Thoughts
Initially, the PEX company had a bit of a rocky start with frequent product leakage complaints. Now, after significant product improvements, you can find PEX products in over 60 % of new construction residential water supply systems (source).
(Video) PEX Pipe: Plumbing for Homeowners
Both PEX-A and PEX-B tubing are exceptional choices for underground work and can be supported with stainless steel PEX clamps or Sharkbite Fittings.
The significant lifespan of PEX tubing ranging from 40-50 years in addition to its 25-year warranty (an offer that also applies to Sharkbite fittings) enhances the benefits of choosing these products over more traditional pipes for underground systems.
FAQs
In areas with aggressive soil or water conditions, plastic or brass PEX fittings are recommended. And being a high-density polyethylene, PEX can also be buried in concrete. In fact, it's often used for radiant heating under concrete floors.
All types of PEX piping could be installed underground, nonetheless, it's better for you to research which type will last longer in this condition. As a general rule, PEX B can withstand freezing temperatures for a longer time than PEX A due to its high bursting pressure.
Normally, you would bury your underground insulated PEX pipe 2 feet deep (24 inches). This is the right depth for burial for most ground types. Just be diligent enough to backfill it with soil that is not rocky or preferably sand.
Use sleeving over PEX where it runs through an expansion joint/crack. A b, crack-resistant poly piping is preferred and should cover (sleeve over) the PEX pipe at least 1-1.5ft on both sides of the joint. For 1/2" or 5/8" PEX, 3-4ft long pieces of 1" PEX can be used for sleeving.
What is the best pipe to use for an underground water line? You can use both PVC and PEX pipes for an underground water line. Both are able to be buried and can withstand high pressure. They are also resistant to rusting and corrosion.
PEX cannot be taken anywhere outside the house where it is directly exposed to UV for a long period. The synthetic material of the pipes can slowly disintegrate due to the light of the sun. Even light bulbs have been known to affect the lifespan of PEX water lines.
Compared to metal, PVC offers high corrosion resistance and is long lasting. It also better withstands tree roots. Running a PEX line underground is beneficial because of the pipe's flexibility, temperature resistance, and lack of a need for multiple connections along its lengthRun the PEX tubing: PEX needs to be supported a minimum of every 32 inches when running horizontally. If running vertically, it should be supported every 4 to 6 feet. (These supports are critical.) The straps should be plastic or metal that is designed to work for plastic pipeCan PEX be used under a concrete slab? Yes, PEX pipe can be installed under a concrete slab. Since PEX pipe is supplied in coils, installation can be completed in one continuous length without the need for extra fittings. PEX pipe can also be encased within a slab, which is commonly used for radiant floor heating.As a general rule, place PEX tubing at approximately 3" from what will be concrete slab edge and 12" apart. PEX Installation TIP: Divide the total area by the number of loops to be used. Space the tubing as required, filling that area with the loop serving it.PEX pipe is made from high density polyethylene (HDPE), PEX is the term for the cross-linking process that improves the chemical and temperature performance of the polymer and provides flexibility and additional strength.
According to 'The Department of the Environment', they recommend that pipes should be buried at least 600mm (two feet) underground. At this depth, the soil acts as a natural insulator and prevents them freezing.All three types of tubing are PEX (Poly Ethylene X-Linked) but some have extra features built in. The "barrier" referred to in PEX tubing is actually an oxygen barrier. This extra layer in your tubing will assure that no oxygen gets into your heating system.
One of the most recent studies, conducted in 2021, tested water from eight different types of PEX pipe brands. They found that 62 chemicals leached into the water, with half of those organic compounds leached considered toxicological but not carcinogenicThe plastic pipe manufacturers like this provision. If a water distribution system is designed for a maximum flow rate of 8 ft./sec., you will never need a water hammer arrestor for CPVC or PEX tubing. So, if you only install CPVC or PEX tubing, you don't have to read any furtherPEX tubing is approved for direct burial outdoors, a practice most often necessary when running a water supply line to a house. PEX, since it can expand, resists freezing more effectively than rigid pipe, but PEX can still burst if water freezes in a line.
One of the most apparent differences is that PEX A is very flexible and doesn't show any coil memory. So, as a user, you can repair any kinks simply with a heat gun. On the other PEX C is manufactured via the Electronic irradiation or cold method of cross-linking. PEX C is also more vulnerable to kinks and damageIs PEX Suited for Underground Use? A PEX pipe will hold up well underground. However, it's not rated for use outdoors unless it's properly insulated. If using a PEX pipe underground, bury it at least 12 to 18 inches below the frost line.
You can snake long runs through joists and studs. Second, PEX doesn't sweat under high humidity conditions, and it's also resistant to bursting, even if the lines freeze solid. Third, joints are easierPEX without an oxygen barrier actually allows oxygen from the air to permeate it. This happens even if the PEX is embedded in concrete. Cast iron circulator pumps will rust when oxygen is present in the water. Bronze or stainless steel circulators are necessary when this is the case.
The test pressure is recommended to be at least equal to the expected working pressure of the system, not less than 50 psi (345 kPa)5 and not greater than 160 psi at 73 °F (1,103 kPa at 23 °C), or as recommended by the tubing manufacturer crimp equipment is a lot more affordable than the expansion kits. However, despite the pricing and form shortcomings, most plumbers prefer to work with expansion because installation is more efficient. There are fewer leaks, and it also seems to be more durable against extreme colds and frozen pipes, etc.
If you are considering purchasing a good fitting for your pipe, it is better to make use of the plastic PEX fitting as it doesn't easily corrode, it can be used for any water condition and doesn't have fake products like the brass.
You can prevent your PEX pipes from freezing and getting damaged by adding insulation to your basements, garages, attics, walls, windows, door cracks, and others. We also recommend insulating the parts of your pipe that are often exposed to extreme temperatures. | eng | 0b0f6a2e-89b7-4032-a79a-a8fab3356fb7 | https://culitt.best/article/can-pex-tubing-be-buried-guidelines-differences-and-specs |
links between forests, food security & nutrition
The take-home messages on forests and food security and nutrition are essentially that:
Diverse forest and tree-based production systems offer advantages over monocropping systems because of their adaptability and resilience.
Forest and tree-based production systems most often contribute the diverse nutrient-rich foods that are often in short supply in the diets of the poor.
There are a multitude of ecosystem services provided by forests and trees that simultaneously support food production, nutrition, sustainability as well as environmental and human health.
Managing landscapes on a multifunctional basis in a way that integrates food production, biodiversity conservation and the maintenance of ecosystem services can contribute significantly to global food and nutritional security
Forests and trees alone will not achieve global food security, but can play a major role: discourse has started to change. Policy reform and implementation to reflect this shifting paradigm now needs to take place | eng | f7d6f052-2d13-434b-a82e-d6af3a6d9886 | https://www.cifor.org/knowledge/publication/7989/ |
The reports of the honorable body of Matehuala Volunteer Firefighters (2017) suggest that, during the last decade, fire prevention and mitigation services -PMI- have grown progressively, which has generated a demand for mitigation services, despite the constant fire prevention efforts. In Mexico, most fire departments act as an agency independent of the government and define themselves as non-profit organizations. For this reason, in order to study them, we need to apply different approaches to service marketing and social marketing as alternative communication theories, with the purpose of influencing people's behavior towards solving social problems. The objective is to create a linear equation that allows forecasting the number of services provided based on the population, whose contribution is to generate recommendations that provide certainty, in addition to providing service providers, that is, firefighters, with information for decision-making in the search for resources and their management. | eng | da945a00-444e-4fd1-b38f-96b3df947386 | https://repository.universidadean.edu.co/handle/10882/12516?show=full |
Breathing: Learn how to breath properly
This is a mini class teaching you the power of breathing provided by Wim Hof. Should you jump in? WHY WE BREATHE Regulated by the autonomic nervous system, inhaling oxygen is an unconscious process. Fortunately …
Written by: Yogi
Published on: September 1, 2023
This is a mini class teaching you the power of breathing provided by Wim Hof. Should you jump in?
WHY WE BREATHE
Regulated by the autonomic nervous system, inhaling oxygen is an unconscious process. Fortunately it's an unconscious praxis, otherwise we simply wouldn't have a break, as we'd have to deal with it incessantly.
The amount of oxygen that we inhale through our breathing, influences the amount of energy that is released into our body cells.
On a molecular level, this progresses via various chemical and physiological processes. Breathing is the easiest and most instrumental part of the autonomic nervous system to control and navigate.
In fact, the way you breathe strongly affects the chemical and physiological activities in your body. Throughout the years, Wim Hof has developed special breathing exertions that keep his body in optimal condition and in complete control in the most extreme conditions.
The breathing technique is first and foremost premised on inhaling deeply and exhaling without any use of force!
TRY IT OUT!
Wim Hof Method breathing is simple and easy — everyone can do it! Just follow the steps below. We recommend practicing right after waking, or before a meal, when your stomach is still empty.
Note that WHM breathing can affect motor control and, in rare cases, lead to loss of consciousness. Always sit or lie down before practicing the techniques. Never practice while piloting a vehicle, or in or near bodies of water.
WARNING: ALWAYS DO THE BREATHING EXERCISE IN A SAFE ENVIRONMENT (E.G. SITTING ON A COUCH/FLOOR) AND UNFORCED. NEVER PRACTICE IT BEFORE OR DURING DIVING, DRIVING, SWIMMING, TAKING A BATH OR ANY OTHER ENVIRONMENT/PLACE WHERE IT MIGHT BE DANGEROUS TO FAINT. THE BREATHING EXERCISE HAS A PROFOUND EFFECT AND SHOULD BE PRACTICED IN THE WAY IT IS EXPLAINED | eng | d2067a5b-b08e-4986-a056-08420936c8ff | https://fattyliverguide.com/breathing/ |
Best things and drawbacks to living in Galgenveld
Living in Galgenveld, a neighborhood in Nijmegen, comes with its own set of advantages and disadvantages. Let's take a look at some of the pros and cons of this area:
Pros:
Proximity to amenities: Galgenveld is close to supermarkets, shops, and restaurants, making it convenient for daily needs.
Green spaces: The neighborhood has several parks and green areas, offering residents a chance to enjoy nature and outdoor activities.
Public transportation: Galgenveld has good public transportation connections, making it easy to travel to other parts of Nijmegen and beyond.
Educational institutions: The neighborhood is home to several schools and a university, making it suitable for families with children and students.
On the other hand, there are some drawbacks to living in Galgenveld:
Cons:
Noise levels: Due to its central location, Galgenveld can be quite noisy, especially during peak hours and weekends.
Parking: Finding a parking spot can be challenging in this neighborhood, as spaces are limited and often occupied.
Older buildings: Some parts of Galgenveld have older buildings, which may require more maintenance and renovation.
Safety concerns: While not a major issue, there have been reports of petty crimes in the area, so it's essential to stay vigilant and take necessary precautions.
What does Galgenveld offer?
Galgenveld is a charming neighborhood located in the city of Nijmegen, offering a mix of urban and suburban living. This area boasts a variety of housing options, from modern apartments to traditional family homes, catering to diverse preferences and budgets.
One of the key selling points of Galgenveld is its accessibility to public transportation, making it easy for residents to commute to work or explore the city. Additionally, the neighborhood is home to several parks and green spaces, providing a peaceful retreat from the bustling city life.
Amenities such as supermarkets, cafes, and restaurants are within walking distance, ensuring convenience for daily needs. The area also has a strong sense of community, with various events and activities organized throughout the year.
In conclusion, Galgenveld is an attractive option for homebuyers seeking a balance between city living and a more relaxed atmosphere. Its excellent location, diverse housing options, and strong community spirit make it a desirable place to call home.
Buying agent in GalgenveldHousing market GalgenveldMake all the right homebuying moves — No BS. Pinky promise.
Our buying experts can help buy your new home in Nijmegen | eng | 6fef52a6-3ebf-4fbe-8bdf-706f537171f4 | https://walterliving.com/gemeente/nijmegen/plaats/nijmegen/buurt/galgenveld/BU02680203 |
3 Different Ways That Mobile Games Can Use Blockchain Technology
If you haven't heard of blockchain tech yet, where have you been? It rose to prominence thanks to cryptocurrencies like Bitcoin, and yet now has a wealth of other applications in many industries and contexts.
Mobile gaming is an example of an area where the blockchain comes into play, so what's the deal, and is the hype justified?
The basics of the blockchain
This can be used to determine ownership of things like cryptocurrencies, but can also be harnessed for other purposes, including functions more akin to a traditional database.
What is web3 gaming?
Understanding the concept of web3 gaming is useful if you want to appreciate how the blockchain applies to interactive mobile experiences.
In short, it covers all games that harness blockchain tech in some way, usually as a means of creating the in-game economies they use, and of connecting these with the wider crypto market. For those looking to enhance their mobile games user acquisition strategies, understanding the nuances of blockchain in gaming can provide a competitive edge. This is applied in various ways and with different intents and results, including:
Enabling player ownership of digital assets
With traditional mobile games, when a player unlocks a certain item, it doesn't really belong to them, but is instead just an inextricable part of the game code, and has no extrinsic value.
By using the blockchain, digital assets can become non-fungible tokens (NFTs). So when an in-game item is earned, it is as much the property of the player who earned it as their car or their house.
Furthermore, by making NFTs a part of mobile gaming, this allows players to trade and sell their assets as they please.
Rewarding players for their participation
So far, you could argue that the blockchain and NFTs in mobile gaming aren't really doing that much differently from some long-established games. The economies of titles like Second Life have translated in-game assets to having real-world value without the need for this tech.
However, in the case of web3 games, the relationship between playing and earning is far more formalized and is effectively fundamental to the entire experience.
The most popular titles in this genre reward players for getting involved in the game, whether that's with NFTs or with crypto tokens that are native to the platform.
Thus players can reliably predict how much they'll get out of a game based on the amount of time they put into it. It's no longer a one-way flow of cash and time from the player to the developer, but more of a two-way street.
Building trust
It's an unfortunate reality of the state of play in the games industry today, but players don't have a lot of confidence in many big names in development and publishing. This comes down to a combination of unhelpful business practices, and problematic stories of working conditions at these companies.
With the implementation of blockchain technology in mobile gaming and elsewhere, developers are able to create a level of transparency that has not been possible in the past. In turn, this means that when players want to put their trust in a company, they can do so safely in the knowledge that more aspects of how a game works will be up for scrutiny, as everything is stored on a public ledger.
Final thoughts
You can expect more and more blockchain integration across mobile gaming going forward, and it seems inevitable that at some point it will be so ubiquitous as to be unremarkable, in much the same way as 3D graphics and online multiplayer did in the past. | eng | 812736c7-4c07-43a2-a7bd-d3965c4b2420 | https://appinstitute.com/3-different-ways-mobile-games-can-use-blockchain-technology/ |
Foraging For Food In Rome
Written by Katie Parla on November 7, 2011
During excavation, more than 400 spontaneous plants were identified inside the Colosseum.
Rome's origins are agricultural and, indeed, most of the area that is now occupied by the modern city was home to cultivated and spontaneous plants and livestock less than a century ago. Considering the fact that urban sprawl shows no signs of relenting, urban farming and foraging offer two viable ways for city residents to return to their roots.
Capers in the Parco degli Acquedotti.
One spontaneous plant that is absolutely ubiquitous is the caper bush. They have a particular predilection for brick walls and can be found growing on the interior surfaces of the Colosseum, on aqueduct ruins, on the Aurelian walls, and retaining walls around Rome's hills and parks. The entire caper plant is edible: buds, berries, and even the leaves (more about that topic next week!).
Arugula grows in the Circus of Maxentius.
I used to walk by the Circus Maximus after industrial lawnmowers had been through and wonder why the whole place smelled like a peppery salad. Upon closer inspection, I realized those plants with jagged leaves and yellow flowers were arugula. Arugula is particularly abundant in the wide open spaces of the Circus of Maxentius, the Parco degli Acquedotti, and the Parco della Caffarella.
Mallow growing next to the church of San Nicola in Carcere.
Mallow, also known as malva, is a plant with scalloped leaves and purple flowers that grows close to the ground. It is so common that it may be considered an invasive weed by some. For others, it is prized for its digestive properties. It can be eaten raw or cooked and on Sunday Beppe e I Suoi Formaggi served me goat cheese with bits of dried mallow on top.
Ostia Antica.
Ostia Antica's well-worn bath is full of grass and weeds, but as soon as you venture off the main tourist route, there is plenty of wild mint to be found. As at all archeological sites in Rome, huge pines have been planted. Their cones contain pine nuts ready for the taking by the equipped and patient traveler.
Mature olives on the Palatine Hill.
As was likely the case in the iron age, olive trees populate the slopes of the Palatine and Caelian hills. Olives can be plucked from these branches and cured using water, brine, salt, or lye treatments. If you feel sketchy picking olives on the Palatine during regular opening hours, head for the trees in front of San Gregioro across the valley.
Prickly pears in the Parco della Caffarella.
The meat of prickly pears is edible, though getting to it can be a painful process as I learned the hard way. Rome Photo Blog can tell you the whole story but let's just say I spent the first two days of vacation a few years back tweezing spines out of my tongue and hands. Forage for prickly pears with gloves and peel the skin off to remove spines. The sweet fruit inside can be eaten as is or made into sorbet. The cactus pads are edible, too, and are used in a traditional Mexican salad.
This is how I learn lessons.
NB: It is probably illegal to forage for food in Rome's archeological parks. That said, who's watching? Also, don't be an idiot like me. Use common sense when foraging and do some research before hand or you may end up with thousands of fuzzy spines in your mouth. | eng | 85c42d3e-3658-47b0-bb93-1201b6ad59f5 | https://katieparla.com/foraging-for-food-in-rome/ |
Browse related Subjects
The second edition of this bestselling textbook arms pre-K to middle-school teachers with the most recent developments in reading research--and shows them how to apply their knowledge in the classroom to help all students learn.;
Read More
The second edition of this bestselling textbook arms pre-K to middle-school teachers with the most recent developments in reading research--and shows them how to apply their knowledge in the classroom to help all students learn.;
Read Less
Add this copy of Unlocking Literacy: Effective Decoding and Spelling to cart. $15.00, like new condition, Sold by Booksup rated 5.0 out of 5 stars, ships from Laguna Niguel, CA, UNITED STATES, published 2010 by Brookes Publishing Company.
Add this copy of Unlocking Literacy: Effective Decoding and Spelling to cart. $18.86, good condition, Sold by Goodwill Indust. of San Diego rated 5.0 out of 5 stars, ships from San Diego, CA, UNITED STATES, published 2010 by Brookes Publishing Company.
Add this copy of Unlocking Literacy: Effective Decoding and Spelling to cart. $27.53, good condition, Sold by GreatBookPrices rated 4.0 out of 5 stars, ships from Columbia, MD, UNITED STATES, published 2010 by Brookes Publishing Company.
Choose your shipping method in Checkout. Costs may vary based on destination.
Seller's Description:
Good. Trade paperback (US). Glued binding. 326 p. Contains: Tables, black & white, Figures Unlocking Literacy: Effective Decoding and Spelling to cart. $27.72, good condition, Sold by SurplusTextSeller rated 5.0 out of 5 stars, ships from Columbia, MO, UNITED STATES, published 2010 by Brookes Publishing Company Unlocking Literacy: Effective Decoding and Spelling to cart. $36.39, very good36.81, like new37.47, new condition, Sold by Ingram Customer Returns Center rated 5.0 out of 5 stars, ships from NV, USA, published 2010 by Brookes Publishing Company.
Add this copy of Unlocking Literacy: Effective Decoding and Spelling to cart. $39.27, like new condition, Sold by GreatBookPrices rated 4.0 out of 5 stars, ships from Columbia, MD, UNITED STATES, published 2010 by Brookes | eng | f88bd82b-badf-4389-a95b-89bdc680f295 | https://www.alibris.com/Unlocking-Literacy-Effective-Decoding-and-Spelling-Instruction-Second-Edition-Marcia-K-Henry/book/39401935 |
Authors
Affiliation
Abstract
The threat of worldwide obesity in children is a reality and has become pandemic. Previously a concern of only developed countries, rapid, escalating rates of overweight children now dominate the public health concerns of middle-and low-income nations as well. There are, of course, many influences that have literally shaped the global population, but there is also a recent observable pattern that is shared by those developing countries with increasingly obese children: a grand structural shift in diet and activity levels on every continent and in every region has occurred in the last quarter century, accompanied by rising rates of obesity. Two central public health concerns drive the need for effective interventions: the immediate health of children and the imminently crushing blow that is coming to health care systems and developing economies due to high rates of chronic disease. In developed nations, the role of gatekeeper has shifted to childcare providers, media, and schools, but in the developing world the traditional role of the mother as home manager has remained intact. Accepting the mother as the primary care provider within the child's nuclear environment places the mother as the guardian of the family's resources, which may be a viable alternative to the types of health-promotion efforts found in past ineffective models. | eng | 3933acd1-a0c1-48e9-9583-dcada1fd0eeb | https://pubmed.ncbi.nlm.nih.gov/16075576/?dopt=Abstract |
I can't believe how fast this week went! It has been such an honor to work with such talented young witches and wizards.
For our warmup this morning, the kids were tasked with coming up a catchy headline that might be found in The Daily Prophet. They really know how to sell a headline!
Next, we analyzed the beginning of the first, fourth, and eighth movies. We discussed choices that filmmakers make and why they make them. We then applied that to the greater series and how it shows in the world around us. Before lunch, the students had time to finish up their research projects.
When they came back from lunch, they presented their research projects to the VSA students that they had interviewed. They were so great when presenting their findings! The research questions and outcomes were all eye-opening and well thought out. The older kids really enjoyed seeing their work, and I was so proud of our wizards!
To celebrate a job well done, we had our very own Potter-con! The students were able to converse with one another about how Harry Potter impacts our world, how what we learned this week has changed the ways they think Harry Potter impacts the world, and some big ideas about how Harry Potter became a cultural phenomenon. We also enjoyed some ice cold butterbeer. 🙂
It was so nice to meet you at open house, and I have very much enjoyed having every single child in class this week! | eng | c5bc178a-322c-44de-84ee-c1a97b3391ab | https://pty.vanderbilt.edu/2019/07/savy-2019-session-4-day-5-unpacking-the-magic-of-harry-potter-holland-w-rising-5th-6th/ |
How London Gyms are Adapting to the Digital Age
As with many industries, London's fitness landscape is adapting to the digital age. With new technologies constantly emerging, gyms are integrating digital tools into their services to maintain competitiveness and attract new members.
Data-Driven Fitness and Personalization
Gyms are also tapping into the power of data to provide personalized experiences. By tracking each member's activities, they can create workout plans tailored to individual needs, offer real-time performance feedback, and measure progress over time.
Virtual reality (VR) technology is another digital innovation being integrated into London gyms. VR helps create immersive experiences, making workouts more enjoyable and effective.
Conclusion
Given the increasing digitization and growing demands for personalized and convenient services, it's evident that London gyms that successfully integrate digital technologies will be well-positioned for future success.
Frequently Asked Questions
Are gyms in London using fitness apps?
Yes, many gyms in London are using fitness apps to offer a range of virtual services, from online workout classes to personal training sessions. | eng | 62ae49c0-50d5-4947-bcc8-09777d9931b4 | https://www.dkdnt.com/how-london-gyms-are-adapting-to-the-digital-age/ |
Schools should try to teach kids the basics
The New York Times opinion section put together a series of op-eds from different writers answering the question "What is school for?"
One answer, from Heather McGhee and Victor Ray, is that "School is for Making Citizens," a thesis you could imagine going in a number of different directions. My guess is that a generation or two ago, a piece with that title would have talked about the need to instruct children in the wisdom of the Founding Fathers and to remember the names and dates of various great historical personages and events. The point of citizenship education would have been to make people understand what America was fighting for in the Cold War against the Soviets. You could imagine a death rattle version of this reemerging twenty years ago in the wake of 9/11, when "they hate us because of our freedoms" was the piety of the day, and we would show them by teaching the beautiful story of American freedom.
But what McGhee and Ray mean by citizenship education today is that we should resist conservative legislative efforts to restrict presumptively left-wing teachings about race, gender, and sexuality, but especially race (Ray has a new book out about Critical Race Theory):
In our increasingly diverse nation, insulating students from lessons about racism will create a generation ill equipped to participate in a multiracial democracy. When partisan politicians ban the teaching of our country's full history, children are purposely made ignorant of how American society works. And the costs of this ignorance to American democracy will be borne by us all.
The piece kind of rubbed me the wrong way, even though I wholeheartedly agree that partisan politicians should not censor which aspects of American history teachers are allowed to talk about. But I couldn't quite put my finger on what was bothering me until I saw a later Ray tweet on a separate subject, learning loss during the pandemic.
It is, obviously, true that it is bad when children die. At the same time, we know that very, very few children have died from Covid-19. I don't want to re-run the whole argument about school closures, just to observe that I think these two Ray takes are coming from the same place of underrating the importance of basic "three Rs" education. Citizenship is important, and it is one of the functions of the school system, but the best (and most realistic) way for K-12 schools to foster effective citizenship is to teach kids foundational literacy and math skills. Incorporating works about history or politics into the curriculum is a great idea insofar as it helps keep students engaged, but it's best to make those core skills the North Star and try to avoid hubris and tons of polarized fights about tangential issues.
It's a well-written, broadly accessible book, one that tries to deal seriously with important topics and discuss academic research without being a woolly, unreadable academic text. I put the NYT excerpt through a readability analyzer, and its scores ranged from grade level 10.86 on Fleisch-Kincaid to 13.43 years on the SMOG scale. This strikes me as appropriate for a book aimed at the lay public.
All of which is to say that improving the level of basic reading competency is basic citizenship education.
To read the books that progressives see as foundational to their understanding of race in American history, you need to be able to read better than the typical American can. And of course to really receive a well-rounded citizen-like education in these subjects, you also need to read good books written by people on the other side of some of these arguments. None of this is to defend weird gag orders that restrict what teachers can talk about, but to underscore that even within the parameters of a McGhee/Ray program, we really need to improve literacy education.
Basic math matters
These days there's a lot of interest in math education as a foundation for jobs in engineering fields or writing code.
And that makes sense. Realistically, thinking about a path to higher living standards for underprivileged kids growing up in marginalized communities, "help students get good at math in K-12 so they can get well-paid technical jobs as adults" is a pretty good strategy. A lot of the big picture dynamics that Rothstein and McGhee talk about in their books manifest, concretely, as very unequal access to high-performing schools and to neighborhoods that facilitate upward mobility.
But beyond those aspects of math education, math turns out to be pretty important just as a basic life and citizenship skill.
The most obvious use is that a typical person going through life in the contemporary United States of America faces a lot of questions that relate to debt and investing. Should you get an auto loan to buy a car or should you insist on a car you can buy outright even if that means settling for a real beater? Most people get a mortgage at some point. Most people do at least some saving and investing in the stock market and have some access to tax-preferred retirement accounts. Most people use credit cards and at least sometimes carry credit card debt. Lots of people, it seems to me, imprudently over-indebt themselves in the early years of life. But a simple rule of thumb like "debt is bad" is also very wrong and misleading. Debt offered on attractive terms or in order to purchase investments of enduring value can be very good. I know people who are so fearful of getting into trouble with credit card debt that they don't use credit cards, even though the economics of swipe fees and reward points means that far and away the cheapest way to buy things is to pay with a credit card and then pay off your bill.
Annamaria Lusardi, who does tons of research about financial literacy, has a really good paper showing that bad decision-making in personal finance is strongly related to just not knowing the relevant math. She gives an example of these three questions:
"If the chance of getting a disease is 10 percent, how many people out of 1,000 would be expected to get the disease?"
"If 5 people all have the winning number in the lottery and the prize is 2 million dollars, how much will each of them get?"
"Let's say you have 200 dollars in a savings account. The account earns 10 percent interest per year. How much would you have in the account at the end of two years?"
It turns out 84 percent of people could answer the first question, 56 percent got the second right, and only 18 got the compound interest question correctly. And this is not a particularly difficult compound interest question. Nobody is trying to trick you or using weird jargon like APR. But the people in her survey mostly can't do it. And so obviously people are not going to make great decisions with regard to credit and debt if they can't figure out how interest works. They're probably also not going to do a great job of thinking critically about policy issues.
Doing the work
Again, I don't want to be a strawman who's insisting that schools spend every hour of every day drilling kids on math and reading. It's important for children to find school to be pleasant and engaging so incorporating other kinds of fun activities makes sense.
You also can't teach reading without giving kids things to read. Children have diverse interests, and it's also good to master reading a broad array of genres of writing. It's important and appropriate to have classes on science and history and social studies, and doing this effectively requires more "woke" stuff than conservatives want to admit. At my kid's school they have a hall with pictures of people pursuing various scientific careers, and the demographics of these young scientists are very tilted toward Black and Hispanic faces. One slightly cranky dad was rolling his eyes to me about this, but the school is mostly Black and Hispanic kids, so it seems totally reasonable to me.
"Culturally relevant pedagogy" where you try to make sure kids aren't just reading books by white people with white characters or leaning history as a succession of things done by Important White People seems to work. And I think it's actually pretty common sensical. There'd be nothing wrong, in principle, with American 9th graders spending the year studying the history of Thailand. But in practice, most of those kids would probably find it to be boring and irrelevant, they'd tune it out in high numbers, and it would be pedagogically counterproductive. But if we're talking about a high school in Thailand, that's obviously a whole different story.
In some kind of abstract ideal, perhaps everyone would have their own tutor and super-custom curriculum. In the real world, you need to make choices and target population averages. And in our increasingly diverse classrooms, there's a strong case for teaching different stuff than schools taught a generation or two ago.
But fundamentally, these questions are secondary — the lodestar shouldn't be "what history books are kids reading," but "are the kids learning to read history books." There was a recent story in Time about how Oakland Public Schools adopted a successful phonics-based curriculum, but teachers didn't like it because they thought it was tedious. If you want to attract and retain good teachers as public school instructors, you do need to care about their preferences and quality of life. But fundamentally, we also need to use sound methods for teaching children to read, which is not what's happening in many schools. The political system needs to focus on driving toward these basic goals and providing the financial resources needed to achieve them.
The indoctrination question
One thing that to me makes this whole citizenship push so misguided is that pretty much no matter what we do, the majority of the people working in the public school system are going to be on the left. At the same time, the population of parents if anything tilts to the right, because religious people have more kids.
So there are bound to be some concerns that school systems are indoctrinating children in ways their parents don't approve of. And this is something we've been dealing with for a long time around, say, teaching evolution in biology classes. I think the sane way to navigate this is to try to reassure people that science class isn't dunking on the Bible, even as you insist that teachers are going to teach science factually.
And in the higher education context, professors point to lots and lots of studies showing that even though a very large majority of professors are left-wing, they are not substantially changing their students' political opinions. Which is to say they recognize, correctly, that the very left-wing tilt of academics raises suspicions in some people's minds and that debunking those suspicions is important to making universities operate effectively and to trying to maintain a healthy degree of faculty autonomy.
But that's exactly why it seems so perverse to me to dive into the K-12 zone, guns blazing, telling people that we want public schools to emulate SNCC Freedom Schools, where students "changed their state by organizing voter registration drives and civil rights protests and by charting a more equitable future in terms of housing, jobs and health care." The historical SNCC was operating a volunteer campaign in the context of radical abrogation of electoral democracy in the Jim Crow south. Today, what we need to do is practice democratic politics to achieve excellent public education.
At the end of their piece, McGhee and Ray offer a diagnosis of the macropolitics that I think is correct — conservatives want to ride backlash to perceived leftist influence over public schools into dismantling the public school system, and that's bad:
The people who resist an honest teaching of history have an economic agenda, too. They attack our children's freedom to learn in order to create "universal public school distrust," as Christopher Rufo, one of the leading architects of the effort to censor the teaching of race in the classroom and an advocate of school vouchers, put it. When white parents — and the tax dollars that often move with them — abandon public schools out of fear of integrated curriculums, it drains the pool of public resources from our schools. It is no surprise that some recent campaigns to pack school boards, sue districts and spread book bans are reportedly funded by some of the same secret money groups that espouse low-tax, small-government economics, while financially backing the nomination of conservative judges.
That's completely correct. And insisting that K-12 schools need to center progressive citizenship ideals plays directly into Rufo's hands. What you want to do is focus schools' work on transparent and broadly acceptable curriculum goals like getting most students reading at grade level and competently solving the math problems of everyday life, while creating opportunities for the most talented students to go to competitive colleges, regardless of family background.
This is a big enough and hard enough job that we don't need to saddle schools with tons of other expectations. Building durable consensus around investing the resources to make it happen is hard enough that we don't need to incorporate tons of other fights. Pushing back against blunt censorship efforts makes a lot of sense to me, but part of the pushback should be assuring people that parents and educators are broadly aligned around non-ideological learning goals. And people convinced that their interpretation of historical and sociological facts about America are correct should have some faith that a literate, numerate public will be able to figure that out.
One thing I don't understand about the debate over teaching racial history in school is the idea that this is a new thing. But I remember race and race relations being a central theme in my study of history in high school (I was in a Republican-leaning school district in California and would have been studying this around 2005). When I was in high school, we studied genocide of the Native Americans, the brutal realities of slavery, the xenophobia against Chinese immigrants in California leading to the Chinese Exclusion Act, the horrors of Jim Crow and segregation, and the backlash to Civil Rights movement. We studied the court cases covering affirmative action and the subsequent backlash. We learned about Thomas Jefferson's owning of slaves and the ordeal with Sally Hemmings, and how Jackson made his reputation attacking and displacing Native peoples. We learned about the Zoot Suit riots against Mexican immigrants, the Bracero program, and Cesar Chavez fight for labor rights for migrant workers. We read the People's History of the United States by Howard Zinn (which was balanced by more conservative text by Paul Johnson). We read first hand accounts of Japanese who were interned during WWII. I remember then having adults ask me then if our history class was too critical and made us feel too negatively about the country and our national heros.
Now I'm sure that there were gaps in my history education. It's true that I don't remember learning about Juneteenth or the Tulsa massacre. And I'm sure there were places where there was less focus on themes of race relations than I had. But it's hard for me to think that my history education was unusual for the mid aughts. But when we are having this conversation about race and history education, people my age are pretending like their history education was overly positive and patriotic, and that racial history of African Americans and other minority groups was this enormous gap. And either my memory is just completely wrong, or my education back in 2005 was truly exceptional for its time.
This is a great piece and I think hammers home the concerns a lot of people (or at least I) have about this approach to education. To start, I think Chris Rufo is a shady character and I'm always amazed at how few remember his employer to be the intelligent design people. Whatever their goal is it probably isn't good, non-ideologically charged education on its own merits.
However the kind of Tema Okun stuff that characterizes habits of good students as 'white supremacy' is bad, and probably bad for kids and the education system. Introducing strange gender concepts without a lot of scientific basis to impressionable kids that can barely read is a recipe for stupidity, and in some cases seems to presume authority public schools don't have. I can think of no better way to destroy support for public education than to embrace ideas seemingly designed to hurt quality and piss people off.
So none of this is to say you don't teach about Jim Crow or redlining in a social studies or history class, not to mention of course the fact that these things are now illegal after the hard work of dedicated people. But if the perception becomes that priorities of the public schools are something other than core educational concepts like the 3 R's tax payers will eventually defect. | eng | ea12f703-48ff-4a8b-8ed6-bc53f8cb9bb8 | https://www.slowboring.com/p/schools-should-try-to-teach-kids |
CEIPAL Workforce VS Calabrio
Let's have a side-by-side comparison of
CEIPAL Workforce vs Calabrio to find out which one is better.
This software comparison between CEIPAL Workforce and Calabrio is
based on genuine user reviews. Compare software prices, features, support, ease of use, and
user reviews to make the best choice between these, and decide whether
CEIPAL Workforce or Calabrio fits your business.
A Quick Comparison Between CEIPAL Workforce and Calabrio
Let's have a detailed comparison of CEIPAL Workforce vs Calabrio to find out which one is better. Let's discover some of the essential factors that you must consider and decide whether CEIPAL Workforce or Calabrio fits your business.
Which Is Better? - CEIPAL Workforce or Calabrio
CEIPAL Workforce and Calabrio CEIPAL Workforce and Calabrio during demo to decide which one is best for your business.
FAQs
Is CEIPAL Workforce better than Calabrio?
While CEIPAL Workforce offers customized features, Calabrio has diverse modules. Select between CEIPAL Workforce or Calabrio as per your business needs.
Are CEIPAL Workforce and Calabrio same?
Both these products offer similar features but with a different approach. While CEIPAL Workforce emphasizes on user experiences, Calabrio focuses on functionalities.
Which is better CEIPAL Workforce or Calabrio?
It all depends on the functionalities that you prefer. While CEIPAL Workforce offers add-ons along with essential features, Calabrio emphasizes on customization.
Is CEIPAL Workforce the same as Calabrio?
No, the difference between CEIPAL Workforce and Calabrio lies in terms of features and functionalities. While CEIPAL Workforce can be customized as per the user's needs, Calabrio addresses the diverse audience.
Can CEIPAL Workforce replace Calabrio?
It depends upon your business requirement. Both CEIPAL Workforce and Calabrio provides standard features to cater diverse industry needs.
What is the difference between CEIPAL Workforce and Calabrio?
The difference lies in terms of the user interface and functionalities. While CEIPAL Workforce is customizable, CEIPAL Workforce offers comprehensive modules. | eng | 213484e8-ac62-47fd-a4ae-fe941d9e3e91 | https://www.techjockey.com/compare/ceipal-workforce-vs-calabrio |
also take a sample of what I prodcued and mail it to my Lab buddies over at the universty and see if they can figure out how effective the proccess was.
I was going to add it to my chamber but I am afraid that my yield might not be efficient enough to produce any results at all.
The tests I have planed are as fallows:
1. Spark Plug test with no water vapor.
2. Spark Plug Test with ultrasonic water mist.
3. Spark Plug test with ultrasonic water mist with ZNO crystals added.
4. Weight of chamber before a 1 hour run.
5. Weight of chamber after 1 hour run.
6. Total battery drain.
7. Total over all power consumption.CommentDisclamer: Do not try this at home!
that will work, use free heat of the sun to do work for you.
I would then just bubble this gas into common ammonium, what is the gas volume?.
how about 2 mason jars using the sun to heat vessel A and collecting by bubbling into vessel B?
I would just leave A in the sun and B in the shade or cooled.
Comment
I can't login to Overunity.com and they won't get back to me with a method of getting into the site. I don't know what's going on there, but I made about ten attempts to Login there today and was unable to do so, and they don't get back to me. So, it's frustrating. I have all your archives uploaded to Scribd and was going to post the links to each of the three archived files there so folks at OverUnity.com could download them if they wanted them.
Anyway, If you would let me send you the information via a pm, I can do that and you can put it up at the site. Let me know if that is amenable with you. If not, I'll try to gain entry into OverUnity again tomorrow and see what transpires with it. | eng | 62ba904d-ebfd-4044-894f-220ab3898449 | http://www.energeticforum.com/forum/energetic-forum-discussion/renewable-energy/water-fuel/7477-discussion-best-way-to-use-heat-from-hho-to-generate-electricity/page21 |
Sets the maximum number of nodes the selection can hold. The action taken when appending a node that would take the selection above its limit depends on the current capacityPolicy, which can be either Selection.DISCARD_EXISTING (the default) or Selection.DISCARD_NEW. | eng | 9efcc8e3-2d08-4c94-9d3d-b712a1139bc4 | https://docs.jsplumbtoolkit.com/toolkit/6.x/apidocs/browser-ui.selection.setmaxnodes |
ADVISORIES
GEM
FRAMEWORK
SEVERITY
PATCHED VERSIONS
~> 2.3.12
>= 3.0.8
DESCRIPTION
The cross-site scripting (XSS) prevention feature in Ruby on Rails 2.x
before 2.3.12, 3.0.x before 3.0.8, and 3.1.x before 3.1.0.rc2 does
not properly handle mutation of safe buffers, which makes it easier
for remote attackers to conduct XSS attacks via crafted strings to an
application that uses a problematic string method, as demonstrated
by the sub method. | eng | 342c3b14-a539-4fe7-bd8d-5f1a60c43f63 | https://rubysec.com/advisories/CVE-2011-2197/ |
About that SHA-1 collision: Linus Torvalds has taken to Google+ to emphasise that in Git, its main role is error detection, so "the sky isn't falling."
The weak hashing algorithm is used, among other things, to provide a digital signature for software, documents like PDFs, and encryption certificates. The mathematical …
COMMENTS
That's not how hashes work
"The mathematical operation should produce a unique result for any given input"
That's not how hashes work. There's lots of inputs that will all produce the same hash, and producing a hash from an input is, relatively speaking, computationally trivial. The tricky bit is to, for a given hash, find an input that will generate that hash and that is meaningful in the context of the original input (for example, as Google did producing a second valid PDF). That is computationally very difficult. That's why they are often called "one-way" functions.
Re: That's not how hashes work
The issue is whether you can compromise the client before they realise the code is not what they think.
If you can compromise the GIT client and then pull down the proper code in the background so you don't know you've been compromised, you have a problem.
This is one reason its good to avoid monoculture. It would be hard to compromise lots of different clients (while maintaining the hash) and you only need one client to notice the compromise and the user to alert the others.
Re: That's not how hashes work
Electron, what you post is nonsense. A 160 bit hash _does_ in practice produce a unique result for any given input (unless you spend 6,600 years of CPU time to search for two given inputs with the same result).
. It is also possible that 5 seconds after the backup, a small meteorite crashes into your office and destroys your computer, and a bigger one destroys the data centre where your backup is held. The latter scenario has higher probability.
Re: That's not how hashes work
Sorry, but no it doesn't! Any hash key can produce only a limited number of distinct results, and while the greater the number of bits in use, the less chance of *accidentally* encountering a collision. To go from "the chances of a random collision are vanishingly small" to "it always produces a unique result" is the sort of dangerous mistake that hackers love to exploit.
Linus was using SHA-1 as a cheap way of calculating a hash that was *very unlikely* to collide, the hackers are using a known algorithm to produce a predetermined result.
Imaging Linus was simply summing the bits mod 1024. there would a 1 in 1024 chance of a collision. If a hackers target has a hash of 512 and the code they want to use has a hash of 384, then they just have to add 128 to produce a valid fake.
Re: That's not how hashes work
Well, this is what I wonder about. Say you modify a genuine function in the source to do something nasty - surely all you have to do is bury a nice big comment somewhere else in the source with the appropriate values inside it to create a matching hash? That's for source rather than the compiled exe of course.
I dunno - it sounds like Linus owes Gilmore an apology, irrespective of how exploitable the flaw is. I get what Linus is saying, but it reads like self-justification rather than a best practice.
Re: That's not how hashes work
The approach you describe (hiding random data in a comment) would be much more difficult to hide from Git than 'merely' generating a collision of the sort being discussed. Git doesn't just hash the content--it prepends a header that includes the size of the object in question. This means that the 'doctored' content must either be the exact same size as the original, or the collision must also take the header into account. Accomplishing either of these things is significantly more difficult than simply generating a collision. Not impossible, but not--for the moment--a credible threat. Really it just means that they should move to a larger hash some time within the next few years or so.
Oh yeah.. It also just occurred to me that this type of attack also assumes that nobody is going to bother looking at the commit diff for some reason. "Some random person just made a +1329 / -7 commit to the kernel, you say? About time those bums did something substantial. I'm sure it's fine."
On a side note, I've only recently started digging in to how Git works under the covers, and I hadn't quite worked out what the reason was for including size information when hashing 'blob' objects. Well, now I know.
Re: That's not how hashes work
"The mathematical operation should produce a unique result for any given input"
Given that "any given input" is conceptually an infinite set, always expecting a unique 160bit hash is just ludicrous. The main thing with these hashes is that small changes in the code will give hugely differing hashes, making collisions rare. We all know that it is possible to engineer (by brute force) a collision, but it's a hard problem. Doing so in such a way as to be undetectable in Git is an even harder problem. Way harder. I look forward to seeing the attempts, though :-)
Re: That's not how hashes work
"Electron, what you post is nonsense. A 160 bit hash _does_ in practice produce a unique result for any given input (unless you spend 6,600 years of CPU time to search for two given inputs with the same result)."
160 bits make for a big number. A really big number, with lots and lots of zeros after it. And any halfway decent hash algorithm worth it's salt [*] will be designed such that even a tiny change in the input will produce a significantly different output.
However, 160 bits is not infinity. It's not even a googleplex. And if someone's deliberately trying to force collisions, you'd be foolish indeed to assume that things are safe.
Then too, assuming they're using the "1 GFLOP machine working for a year" definition of CPU-years [**], that figure of 6,600 years isn't as impressive as it sounds.
An Intel i7 can run at over 350 GFLOPs in dual-precision mode, while a modern GPU (e.g. Radeon RX 480) can theoretically churn out up to 5 TFLOPs in single-precision mode; the Tesla K80 used by Amazon for their cloud-computing back-end can churn out 8.74 TFLOPS in single-precision mode.
Then too, there's always the possibility of using distributed computing or specialised hardware - this type of problem is inherently parallelisable and bitcoin mining has shown how effective ASICs can be for this type of number crunching. Also, since the people most likely to want to force a collision are nation-states or hackers, they could well have access to a supercomputer or maybe even botnets - I wouldn't be surprised to see people offering collision-detection as a service, as more traditional profit streams continue to dry up (albeit with botnets increasingly being based on IoT low-power devices, there may not be many of these).
So, yeah. 160 bits is good. But these days, it arguably ain't good enough.
Re: That's not how hashes work
"A 160 bit hash _does_ in practice produce a unique result for any given input "
Wtf does this ridiculous sentence even mean?
A 160 bit hash is 20 bytes, isn't it?
So if one assumes that the source data being hashed is bigger than 20 bytes, it doesn't matter what hash algorithm is in use, there must by definition be more than one set of input data which will generate particular hash values. Otherwise e.g. a disk would store the hash value not the input data. Same logic applies to any sensible hash size.
""
Fancy that. What are the odds that the "someone" in this picture was a disk deduplication vendor? And/or that the logic in question actually was "if the hash of the new data and the already-stored data are different then we know that the old and new data ARE different, and therefore we must store the new data. IF (and only if) the hashes match, we must THEN compare the actual old data and new data, rather than the old and new hashes, and do some other extra work to ensure that the new data CAN be stored and recovered safely."
Re: That's not how hashes work
Re: That's not how hashes work
"The tricky bit is to, for a given hash, find an input that will generate that hash and that is meaningful in the context of the original input (for example, as Google did producing a second valid PDF). That is computationally very difficult."
For difficult read impossible if you don't know the size of the original hashed data. It could be smaller than the hash or it could be gigabytes in size and you'd have to try every permutation of every size which would take way longer than the age of the universe even with a server farm at your disposal. Obviously if the hashing algorithm has a flaw which means it doesn't produce all possible hashes or produces identical hashes more often than it should and/or you have a rough idea of the oriiginal data size this narrows it down but even so, its far from easy.
Common sense approach
Nice 'Calm down dear' from Linus. I enjoy a good 'risk of security compromise' as much as the next sys admin (read: not at all) but these need to be taken in context with the actual risk of implementation and the scope of what can be achieved. As he states, git is not using this to authoritatively state that the data is from said user, merely that it didn't get corrupted in transfer. You already know which source you are connecting to as this is handled by the https certificate.
Very refreshing to have a proper, reasoned out response, as opposed to rushing out a hack of a patch which introduces further problems.
Re: Common sense approach
Both sides have a valid point:
- SAH-1 is not used as a sole measure of correctness, so no immediate panic.
- Sooner or later, someone will find a way to compromise at least some aspects of some GIT-based project if other attributes of generating a hash collision become easy enough that length and position of fix-up crude become easy to manipulate.
Re: Common sense approach
I do not quite follow. What exactly would an attacker manipulate in a git repository?
For the purpose of this discussion, git calculates the hash based on the author signature in addition to the files' contents ( So yes, there are some fields that can be exploited to hide away stuff that won't ruin the source code.
Are you concerned someone will push a commit having an identical sha-1 hash? Well, if you push a commit having an identical hash as a previous commit, then git will assume you are trying to push something it already has. It will ignore it. (
Or is the concern here that somebody will break into the repository and change something directly? Well... Who would notice that, based on the hash...? Is the suggestion here that the hash is somehow protecting the git users?
AFAICT, the weakest link is the pull request. The person(s) maintaining the repository must be trusted to not accept pull requests containing nefarious code. The hash algorithm has no bearing on this process. And naturally: users have to know which repository to fetch.
I am new to git. We only started using it a year ago. I am obviously missing something in this discussion.
Re: Common sense approach
I do not quite follow. What exactly would an attacker manipulate in a git repository?
To be honest, I'm not sure. But time after time people find cunning ways of gaming systems that nobody had thought of before that.
Off the top of my head, the obvious thing is you could manipulate somebody's private GIT repository to change code but still have it appearing to match a public trusted one. Sure, if you have that level of access there are a hell of a lot more nefarious things you might do to them, but that would be one possible way of getting a back-door in to a specific company's system based on a otherwise trusted code base.
Re: Common sense approach
That's a binary blob in the kernel, but the source file in git is in ASCII. It uses 18 distinct ASCII characters, by the looks of it. I don't know whether the published method for creating SHA-1 collisions could be modified to generate hexadecimal characters that could be hidden in a file like that. Does anyone know the answer?
Re: SHA1 still "useful" then?
I missed this reading the original collision notification
I actually missed the part that the attack produced by Google needs to meddle with both sides - "good" and "bad" of the collision. If this is the case, this drastically decreases the attack surface as it is an attack that cannot be applied to an arbitrary SHA1 out there. Sure, an attack that is applicable only to a limited set of SHA1 scenarios today may be modified to hit arbitrary ones in the future. The key word is "future".
As far as any source code control system is concerned it has multiple layers of verification:
1. SCCM integrity (git fsck in this case).
2. Build
3. Code style check
4. Integration test check
If a SCCM uses hashes for addressing, creating a hash collision may get you past 1. You still need to get past 2 - your code should build, past 3 - anything you have inserted must pass style checks and past the test suite. That as a whole is actually a fairly high order. This is also very different from "black boxes" like document formats where you can ship megabytes of non-visible content which nobody will notice. It is also different from digital certificates where you either trust the signature or your do not.
Re: I missed this reading the original collision notification
"I actually missed the part that the attack produced by Google needs to meddle with both sides - "good" and "bad" of the collision."
You missed it because it doesn't 'need' to. They did so in the demonstration because it's easier to match the hashes by manipulating both, but the point was simply to demonstrate that two different (but similar) documents could be made to produce an identical hash (as opposed to two masses of random data generating the same result, which a Chinese team proved about ten years ago). But in theory it should be possible to do the same thing just by manipulating the 'bad' document; it'd just take more compute to do so - but doing so is now an affordable operation.
Re: I missed this reading the original collision notification
@Naselus : ""I actually missed the part that the attack produced by Google needs to meddle with both sides - "good" and "bad" of the collision." You missed it because it doesn't 'need' to. "
That is seriously wrong. The attack absolutely *does* need to fiddle with both sides. Fiddling with only one side is not a collision attack, it is a pre-image attack - and nobody has demonstrated a pre-image attack against even MD5 yet.
Re: I missed this reading the original collision notification
Hence the '' round need :)
Second preimage resistance is implied by collision resistance. If you demonstrate collision resistance is affordable, then second preimage resistance is no longer guaranteed to be unaffordable. So there's no need to actually perform a real preimage once you've shown collision is possible on real data because the hash is already affordably insecure.
Good software design
The key here isn't whether sha-1 should be used in git in the first place.
Good practice in designing security software should acknowledge that after some time all of these things become obsolete so you need to design in a framework that allows you to easily migrate to future algorithm when the need arises. Baking sha-1 into the design is a mistake if it is then too difficult to change.
Other than that,there is no particular reason to be worried about sha-1. It's just another warning shot to not use it in new products and to start looking at how to turn it off in existing software. This should be simple with well designed software.
Re: Good software design
... and it appears that, in git, transition to a different hash is not going to be difficult. At least, that's how I interpret this: "And finally, there's actually a reasonably straightforward transition to some other hash that won't break the world - or even old git repositories."
For crying out loud...
Google talk in dark, foreboding terms about changing a rental agreement to hoodwink some poor unsuspecting tenant into paying a higher rent than they think they signed up for and "proving" it with the SHA-1
But in their "proof" of the attack all they did was change the color of a frickkin' graphic and people lapped it up and jumped all over the SHA-1 is broken bandwagon.
And since SHA-1 has been suspect for years, anyone trying to use it to sucker a mark into paying a higher rent will (or should) find themselves being asked why they are relying on a demonstrably broken digital "signature" ?
And that's even assuming that it is possible to "attack" a document in this way, rather than just fiddling with the colors.
Context is King
The attack is important, however at present it only matters in a very limited context.
Google have proven that it is feasible for an entity to produce two PDFs with different content and the same SHA-1 hash, by means of embedding some "junk" in the unused data sections of both documents.
Thus a third party may replace a document they wrote by another document they wrote, and you won't detect it if SHA-1 is the only thing you use to check that it is unchanged.
In other words, don't use SHA-1 as the only method of checking if a document is unchanged. Use other hashes as well, and for very important things (eg contracts), compare the full binary data - it's the only way to be sure.
Heck, even MD5 is likely to be good enough. Yes, it's now simple to create a hash collision in MD5, but a collision of MD5 and SHA-1 at the same time?
Re: Context is King
@Richard 12
"Google have proven that it is feasible for an entity to produce two PDFs with different content and the same SHA-1 hash, by means of embedding some "junk" in the unused data sections of both documents."
I am not a mathematically sound person, coming from that well know lilberal arts route to low level programming. I am, though, bemused that this is news to anyone? Particularly the smartest people in the room? Did sane people really think that the output of a hash, given unlimited input data, was always going to be unique?
edit: I see DougS made the same/a similar point below/earlier. The hash doesn't matter, it's how you roll it...
SHA-1 was fine for what he was using it for (when the decision was made to use it) but he was warned that sometime in the near future (now) it's not really good enough.
Sure the sky isn't falling, but everyone should be using SHA256 going forward.
And, many things using using SHA-1 should transition sooner rather than later. The question is how much time/money is that going to take, and is it OK to ignore it because by the time the Sky (really) is Falling it's going to be obsolete (not in use) anyways.
Not really - SHA-1 was, as you say, fine for what he was using it for.
And if you look at why he's not panicked, it's because he didn't rely on any particular assumptions about SHA-1 lasting forever, for instance.
Hashes and cryptoalgorithms last 10 years now if you're lucky. Protocols and code last much longer than that (isn't the Linux kernel over 20 years old now? And even IPv6 is that old and not seeing full deployment still.
In git, SHA-1 is not used for security, it's used as a quick check, and a easily referenced nugget of information that can identify a particular change. As such, it can be replaced by any number of things quite easily. Sure, probably an on-disk format change would be required too but in such an open program, that's hardly a concern.
But if you were using SHA-1 in your SSL setup, you have had an issue for a while now. That's why it's being phased out. And we all knew that was what was going to happen.
Sooner or later, WPA2 will be dead, just like WPA and WEP before it.
Sooner or later, SHA-3 will be dead, just like SHA-2, SHA-1, MD5 and myriad others before it.
Rather than design your protocol to be RELIANT on it, especially if that reliance directly affects the secrecy of data rather than, say, use as a quick reference checksum in an open repository, design your protocol such that every such advance is handled like this: "Yeah, it's not really a problem. Next version fixes it for another 10 years."
Am I missing something here?
"...There's a big difference between using a cryptographic hash for things like security signing, and using one for generating a 'content identifier' for a content-addressable system like git"
Is Linus actually saying that the cryptographic aspect of SHA1 is really irrelevant to Git? That if, tomorrow, somebody comes up with a simple means of taking a given message and creating a message with the same SHA1 hash, it'll screw everybody using SHA1 for secure signing, but wouldn't matter as far as Git is concerned?
I could believe this, having used utterly non-secure hashes in (for example) making hash tables; for that purpose, you want something with an even distribution and few collisions, but security is irrelevant. An example from my own code :
I needed blazing speed and an even distribution without collisions, but not security.
Re: Am I missing something here?
Yes, that's exactly what he's saying. Git is using SHA1 signatures the same way RS232 used parity, to detect corruption. SHA1 is just far far better at doing that than parity. The fact that it is possible to deliberately engineer a SHA1 hash collision is irrelevant to its ability to be used to detect unintentional file corruption.
If git was trying to use cryptographic techniques to prove a file is authored by Linus versus me, it wouldn't help it was hashed with SHA256 instead of SHA1. It would have to be signed by Linus' private key, and your copy of git would have to have securely received Linus' public key to verify that the repository you downloaded was signed by him, and not by signed by me trying to give you Linux code with a backdoor built in.
Since the hash is generated on the machine of the person running git, if it was changed to use SHA256, you still couldn't tell the difference between a repository created by Linus and one created by me. Both our copies of git would create correct SHA256 hashes, and your copy of git would validate them both.
Lifetime of Hash Algorithms
As Lee D noted, today's best practice will be tomorrow's obsolete methodology. SHA-1 has been considered weak for sometime now. However, intentionally creating 2 documents with the hash is not trivial but obviously doable. The real question is how different are the two documents. If they are obviously different with a quick glance then it is not as disastrous as if it is a couple of very small edits (a scenario many have postulated) that are the difference. If the hash is being used for cryptographic security then there are problems. But if it is a quick check the downloaded file this is not a very serious concern.
Re: Lifetime of Hash Algorithms
For contracts this is a non-issue.
Presumably you *kept* the original copy of the contract, and you can immediately show that the one you have also has the same SHA1 hash. If you present this immediately (as soon as the 'bad' contract is shown by the evil other party) you confirm that you didn't have time to compute your own bad version.
For certs it is an issue. If someone can buy a cert for randomdomain.com and convert it into an EV cert for bigbank.com - or worse, a sub-CA cert - that's not great.
The way it seems to me
As I understand hashes, the smaller the change being made, the harder it is to find a usable collision. So if you want to sneak in something unnoticeable (and that will compile) you're going to have to really work for it.
In fact, Git's security model incorporates precisely cryptographic signatures (external to Git itself, and therefore adaptable, I might add). It is up to the individual team however to decide whether and how they integrate this into their workflow (cf. the Linux kernel).
Re: Once again - try it with .TXT files
Until someone can demonstrate two .TXT
No, you don't really want to wait that long once this kind of proof of concept has been produced. Some of the people who might exploit it may have access to resources considerably beyond those used in the study and they won't tell you when they can do it.
Fortunately, replacement ciphers are available and should be rolled out. No new encryption projects should rely on the older protocol.
Re: Once again - try it with .TXT files
What can actually happen? Let's say there is a file in a repo with hash X. And I can manage to create a different file with the same hash X. The first problem is that I cannot put that file into the repository - git will tell me that it's the same hash, so the file "is already there" - it doesn't let me replace a file with an identical file because that is inefficient and pointless, and it doesn't let me replace a file with a different one with the same hash because it thinks it's the same one. You just can't have two different files with the same hash in the same repository.
The only attack vector is to replace a file on some developer's hard drive with a different one, and git won't notice. So you would need to access that developer's computer and replace a file.
Why are you even going over this here? It's literally Linus' first reply to the mailing list post linked in the article, along with several much better informed and compelling points. SHA1 being cryptographically broken is basically a non-issue in the context of git.
Linus' ego
"nasty people will teach him the threat model"
That may indeed happen.
I totally trust Torvalds to quickly learn from the experience and do the right thing. Actually, I'm certain he is already considering alternatives. For all his outbursts and temperamental postings, Torvalds is unquestionably intelligent and reasoned. You might fool him once, but you won't get a second chance. | eng | a4913095-8d45-44f5-b999-996a293d7f39 | https://forums.theregister.com/forum/all/2017/02/26/git_fscked_by_sha1_collision_not_so_fast_says_linus_torvalds/ |
Junior ACT Information Letter 3.4.21
Junior ACT Information Letter 3.4.21
Posted on 03/04/2021
Dear Parent/Guardian:
This is a reminder that Wisconsin students in grade 11 will be participating in the ACT with writing assessment. Our school is scheduled to take the ACT with writing on Tuesday, March 9. All students, in-person, virtual, and blended, should report to Badger at 7:40 am on March 9. In-person students will be notified of their rooms on Monday, March 8 during their 3rd hour class. Blended and virtual students will need to check in at the main office on March 9th to determine which room they will report to that day.
At this time in-person assessment is the only viable method for the state assessments. The company who produces the test has no remote (virtual) testing option available. Proctored, on-site testing ensures standardized administration that includes equitable access to technology and optimal testing environments, test security, and validity.
We are taking every precaution to ensure a safe testing environment. We are following procedures recommended by the Center for Disease Control (CDC) and Department of Health Services (DHS) including social distancing, smaller numbers of students within a classroom, and the frequent sanitization of student spaces and equipment.
The ACT with writing tests students in the areas of English, Reading, Math, Science, and Writing. The entire test will be administered in one day. Please make sure students are healthy before sending them to school. Please keep students home who are ill or have been exposed to COVID-19. Students not in school on testing days will be scheduled for a make-up test on Tuesday, March 23.
Please remember that students will need a #2 pencil and an approved calculator the day of the test. Please visit to determine if your student has an approved calculator. Math teachers are willing to check out a calculator to students in need of one, but students must ask before March 9 if they are interested. In addition, please be aware that cell phones, Smart watches, food, beverages, books, and backpacks will NOT be permitted in the testing rooms. We ask that students please leave these items at home or in their locker during testing so as not to void the test. Snacks and water will be provided during designated breaks during the testing schedule.
Following testing, junior students will be able to leave campus for the remainder of the day. School lunch will be provided; however, for any student wishing to remain on campus. Please make sure that your student has a good night's rest the night before the test and that they eat a balanced breakfast at home or at school in preparation for testing. Again, please make sure that your student reports to their designated room at 7:40 am on March 9.
Lastly, juniors have received directions to complete the non-testing information that ACT requires. This information includes choosing which institutions the test scores should be sent. Please help ensure that your student completes this information online at by March 8. If your student has any questions regarding this process, please have them contact their guidance counselor. | eng | 7a0dd0df-380c-4f01-b709-a55bc808f94b | https://bhs.badger.k12.wi.us/our_school/news/badger_news___announcements/junior_a_c_t_information_letter_3_4_21 |
The Healing Benefits of Lymphatic Drainage Massages
April 2023
As we go through the everyday motions of life, our bodies tend to build up toxins and all sorts of unwanted waste. These unwanted substances can lead to various health problems, including weakened immunity and disease, if not eliminated properly.
Fortunately, the lymphatic system plays a crucial role in the body's defence against harmful substances. This system is a network of lymphatic vessels, lymph nodes, and organs, that help to fight infections and remove toxins from the body.
However, a weak lymphatic system can lead to a variety of health issues such as fatigue, weakness, achiness, and frequent infections. In more severe cases, a poorly functioning lymphatic system can cause conditions such as lymphedema, or even cancer. To prevent this, a lymphatic drainage massage is the perfect solution.
In this article, we will explore the healing benefits of lymphatic drainage massages and the concerns this treatment can target.
What is a Lymphatic Drainage Massage?
Lymphatic Drainage Massages are a type of massage therapy that focuses on stimulating the lymphatic system to help remove excess fluid, waste products, and toxins from the body. This type of massage involves gentle, rhythmic strokes and movements that aim to increase the flow of lymph, which is a clear fluid that contains immune cells and other substances that help fight infections and remove waste from the body.
During a lymphatic drainage massage, a trained therapist will use a series of specific manual techniques, such as light pressure, stretching, and circular movements, to stimulate the lymphatic vessels and promote lymphatic flow. This can help to reduce swelling, and inflammation, while also boosting the immune system and improving overall health.
What are the Healing Benefits of a Lymphatic Drainage Massage?
The benefits of lymphatic drainage massages are extensive, mainly in the transport and removal of stagnant fluids, but also for:
Reducing Inflammation
Inflammation is a natural response of the body to injury, infection, or other harmful stimuli. However, chronic inflammation can lead to tissue damage and contribute to the development of many diseases, including arthritis and autoimmune disorders.
Lymphatic drainage massage can be an effective way to reduce inflammation in the body by promoting lymphatic flow and removing excess fluids and waste products. By improving lymphatic circulation, this type of massage can help to flush out inflammatory substances and other toxins from the body, thereby reducing inflammation.
Lymphatic drainage massages can also help to increase the production of anti-inflammatory substances, such as cytokines and natural killer cells, which can further reduce inflammation and promote healing. This can be particularly beneficial for people with chronic inflammatory conditions such as arthritis or autoimmune diseases, as it can help to alleviate pain, stiffness, and other symptoms associated with these conditions.
Boosting Immunity
The lymphatic system is a vital part of the body's immune response, helping to remove toxins, bacteria, and other harmful substances from the body. Lymphatic drainage massage can be an effective way to stimulate the immune system, helping to increase the production and circulation of immune cells, such as lymphocytes and macrophages, which can strengthen the immune system and improve overall health.
Promoting Healing
When an injury occurs or after surgery, the body responds by increasing blood flow and sending immune cells to the affected area to promote healing. However, this increased circulation can also cause swelling and inflammation, which can slow down the healing process and cause discomfort.
Lymphatic drainage massage can help to promote healing by increasing circulation while also reducing swelling and inflammation. The gentle pressure and rhythmic movements used during the massage can stimulate the lymphatic system, which helps to remove excess fluid and waste products from the affected area. By reducing swelling, the massage can help to relieve pressure on tissues and nerves, which can reduce pain and discomfort.
Furthermore, by increasing circulation, lymphatic drainage massage can also bring essential nutrients and oxygen to the affected area, which can promote faster healing and tissue repair.
Reducing Stress and Promoting Relaxation
Lymphatic drainage massage not only provides physical benefits but also psychological benefits. It can be an effective way to reduce stress and promote relaxation by stimulating the parasympathetic nervous system, which is responsible for the body's rest and digest response.
During a lymphatic drainage massage, the gentle strokes and pressure applied to the body can help to stimulate the parasympathetic nervous system. This can slow down the heart rate, lower blood pressure, and reduce cortisol, the hormone associated with stress. By doing so, the massage can help to promote a state of deep relaxation, reduce anxiety, and alleviate symptoms of depression.
What Concerns Can Lymphatic Drainage Massages Address?
Lymphatic drainage massages can be used to treat a variety of conditions, including:
Lymphedema: Lymphedema is a condition in which excess fluid accumulates in the tissues, usually in the arms or legs, due to damage or blockage in the lymphatic system. Lymphatic drainage massage can help to reduce swelling and improve lymphatic function in people with lymphedema.
Post-surgical swelling: After surgery, the lymphatic system may become congested and cause swelling in the affected area. Lymphatic drainage massage can help to reduce swelling and improve healing after surgery.
Fibromyalgia: Fibromyalgia is a chronic pain condition that can cause widespread pain, fatigue, and other symptoms. Lymphatic drainage massage may help to reduce pain and improve symptoms in people with fibromyalgia.
Digestive issues: Lymphatic drainage massage can help to improve digestion and reduce symptoms such as bloating, constipation, and gas.
Stress and anxiety: Lymphatic drainage massage can help to promote relaxation and reduce stress and anxiety by stimulating the parasympathetic nervous system.
Other conditions: Lymphatic drainage massage may also be used to treat other conditions, such as cellulite, fat distribution, sinusitis, acne, and eczema, among others.
Lymphatic Drainage Massage at EF MEDISPA
At EF MEDISPA, your comfort and safety are paramount. We pride ourselves in employing the best professionals and we are always seeking out the latest technology to provide you the best quality of services and maximum results.
From your initial consultation, right the way through to your last appointment, our highly experienced Treatment Coordinators and therapists will be by your side to provide necessary advice and insights for each treatment. They will listen to your concerns and take your full medical history to ensure they devise a bespoke treatment plan that is tailored to your own individual needs. | eng | d46d7c14-2806-4734-b570-561189249f0d | https://www.efmedispa.com/insights/the-healing-benefits-of-lymphatic-drainage-massages |
Placement Cell
Facilities available for apprenticeship trainee in industry under apprentice Act 1961. The trainees in ITI's after completion of two years courses may be engaged by employers for apprenticeship training in the designated trades. Such trainees are given full credit for the periods spent in the ITI and they are required to undergo training in Industry only for the remaining period of apprenticeship. Those who successfully complete apprenticeship training will be awarded National Apprenticeship Certificate to be issued by National Council for vocational Training after they passed the All India Trade Test. Apprenticeship Certificate are of a higher grade than the National Trade Certificate which is recognized for post and services under the Central Govt. in the appropriate field. | eng | 5629a465-2e7b-4f2c-afd6-e695c6697636 | http://kitcbbsr.org/placement.php |
Subaru Impreza STI | eng | 243d37b0-95b0-407a-b0bd-e7dc29e18705 | https://in4ride.net/tag/subaru-impreza-sti/ |
nonprofit organizationsBy Rebekah Deponte |Staff Writer|
Celebrities taking part in charity work or even starting their own foundations/nonprofit organizations is nothing new, but it also is not getting old any time soon.
Angelina Jolie, Brad Pitt, Oprah Winfrey, and several other Hollywood heavy-hitters are well known for their contributions in the charity world, but there are other celebrities who have been making major contributions to the betterment of our world without the spotlight on their activities.
Some might argue that tax deductions are a huge part of the reason wealthy people get so involved in charity and they would not be wrong in believing that.
If a celebrity donates $1 million to … [Read more...] about Celebrities give back | eng | 3bcbfb4e-67eb-46ab-b529-c4cba161d760 | http://coyotechronicle.net/tag/nonprofit-organizations/ |
A Gift That Keeps Giving: The Long-Term Benefits of Child Sponsorship Programs
In a world marked by inequality in access to education, healthcare, and basic necessities, there exists an urgent need to address the challenges faced by children in underprivileged communities. These young individuals often grapple with circumstances beyond their control, hindering their ability to thrive and reach their full potential.
One impactful solution that has emerged as a beacon of hope is child sponsorship programs. This generous approach connects individuals and organizations with children in need, creating a lasting partnership that extends far beyond immediate assistance.
In this blog, we will delve into the profound and lasting effects of child sponsorship.
Advantages of Child Sponsorship Programs
1. The Heart of Child Sponsorship
Child sponsorship is a special type of charity that links people or organizations with children in need and forges a long-lasting relationship. Sponsors promise to provide money, often on a monthly basis, to help meet basic requirements, including food, clean water, education, healthcare, and other things.
2. Immediate Impact: Meeting Basic Needs
Programs for child sponsorship instantly take care of sponsored children's urgent needs. They guarantee that children have access to healthy food, clean water, and medical care—all of which are essential for a child's growth and development. In addition to saving lives, this assistance enhances the general well-being of sponsored children.
3. Education: A Lifelong Gift
Access to education is one of the strongest and longest-lasting advantages of child sponsorship. By assisting with school costs, materials, and uniforms, sponsors help end the cycle of illiteracy and create opportunities for a better future. Children who receive an education are given the information and skills to rise beyond poverty and create better lives for both themselves and their communities.
4. Nurturing Dreams and Aspirations
The dreams and aspirations of children are fostered by child sponsorship. It gives them the conviction that, whatever their circumstances, they can accomplish their objectives. As sponsored children receive assistance, they start to see a future full of opportunities, which inspires hope and drive.
5. Building Strong Communities
Child sponsorship benefits not just the sponsored child but also the entire community. Children who have better access to healthcare and education help their communities by improving literacy rates, opening up more economic opportunities, and improving general well-being.
6. The Ripple Effect of Empowerment
As sponsored children mature and achieve success, they serve as inspiration and role models for others in their communities. This empowers others, which encourages more kids to seek school and work for a better future.
7. Breaking the Cycle of Poverty
Child sponsorship plays a crucial role in breaking the cycle of poverty. Children are more likely to stay in school, graduate, and find greater employment opportunities when they receive the help they require. Increased financial security for them and their families follows from this.
8. Health and Wellness
Access to healthcare treatments is frequently included inchild sponsorship programs. Routine medical exams, vaccines, and preventive treatment influence the long-term health and well-being of sponsored children. Healthy children are better able to study, develop, and flourish.
9. Encouraging Long-Term Community Development
Community development efforts are frequently incorporated into child sponsorship programs. These initiatives concentrate on enhancing infrastructure by constructing facilities including schools, clinics, and clean water sources. Sponsors contribute to the long-term benefit of communities by funding these upgrades.
Conclusion
In conclusion, the concept of child sponsorship is a powerful testament to the enduring impact of compassion and generosity. As we've witnessed throughout this exploration, its benefits extend far beyond the present, leaving an indelible mark on the lives of sponsored children and their communities.
Child sponsorship programsrepresent a bridge between distant worlds, fostering connections that transcend borders and span generations. They demonstrate that, with the simple act of giving, we can shape the destinies of countless young lives, offering them hope, opportunities, and the chance to dream big.
Sueños De Yarali offers a sponsorship program to make a difference in the lives of impressionable children. Visit our official website to know more about our sponsorship programmes and how you can join our community. | eng | 2f8d666a-f491-46c1-af12-c3fba7facf9d | https://theamberpost.com/post/a-gift-that-keeps-giving-the-long-term-benefits-of-child-sponsorship-programs |
Whether this is your first time with FPP or you're a seasoned pro, I'd like to share some of my top tips for perfect piecing and accurate blocks every time. I hope you'll find them useful and that it sets you on the path to FPP confidence!
1. Get the Pens Out
One of the joys of quilting is putting together colours and fabrics. Curating fabrics can bring us such joy. With FPP it's important to remember that because you're working on the underside of your paper pattern, everything is back to front. And given that this method is a little counter intuitive anyway, it can be very easy to get your colour/fabric placement in a muddle. So I recommend roughly colouring in each section of your block in the colour that you want it to be, to help ensure that you stitch the right fabric piece in the correct place. Trust me, unpicking small stitches through paper is not that much fun! So this really helps avoid that.
2. English Paper Piecing (EPP) Glue
Yes I did say EPP glue, even though we are FPP-ing! A little touch of glue is great to get you started on each section. Take your first A1, B1, C1 etc. section and lightly spread a small amount of glue onto the paper of that section in the middle. You just need enough to hold the fabric in place before you add your second piece. Subsequent sections won't need glue as they will be stitched to the paper. I found the glue to be much easier than pinning to the paper as with glue everything lays nice and flat under the needle of your machine. It may also be that your first section is very small, so a pin wouldn't work anyway. I've FPP'ed 1 ½ inch blocks (yes, you read that correctly!) and the glue was perfect for this as a pin would have been far too big anyway.
3. Light the Way
If you're serious about FPP and see more of it in your future, I definitely recommend purchasing a light box. These days they are wafer thin, very light, portable, charged via a USB cable and super useful. I find an A4 sized one suits all my needs. Not just for FPP but for tracing quilt patterns, embroidery...in fact my children enjoy it for tracing intricate pictures too!
But back to FPP. Place the fabric behind the pattern right side down on the light box to check that your fabric is bigger than the section you're covering and with more than a 1/4" seam all around. If you're fussing cutting a motif, this also allows you to ensure that it's central, or where you want it to be. But if a light box is not for you, there's always the good old fashioned holding it up to a window trick!
4. Add a Quarter Ruler
In my opinion, an add-a-quarter ruler is an absolute must have for FPP. It's a ruler with a 1/4" ridge on it, allowing you to slot the ruler over the fabric and the paper, enabling you to cut a perfect 1/4" seam every time. The advantage is that the ruler fits in snuggly over the fabric and paper, so it doesn't slip and you have full control when you cut. I have the shorter ruler shown here, but they are available in longer lengths. I get on great with this size though as most of my edges aren't longer than this and if they are I can just move my ruler along. The shorter ruler also makes it easier to trim short sections, as it's easier to manipulate.
5. Use a Postcard
A postcard or indeed any bit of thin card is very useful for FPP. You do want it to be of about postcard thickness though. (Personally I like to use a postcard, preferably of somewhere I've visited, as then I can reminisce about happy times spent there.) Put the postcard against the fold line and fold the paper back over the line and towards the postcard. This will ensure that your line is straight a crisp every time. It's such a simple technique but so effective.
6. Rotary Cutter – Size Matters
First of all, make sure your rotary cutter is sharp! We are all guilty of using a blade that's seen better days, but it really does make so much difference when you're using a fresh blade. Also, make sure you're using the right size cutter/blade for the job. If you're piecing smaller blocks it is much easier to use a smaller rotary cutter to give you more control and accuracy. I'm thinking back to my 1 ½ inch blocks – a small rotary cutter was definitely needed then!
So That's the tools and equipment I recommend to give you the best possible results.
Other things I recommend are:
Using a short stitch length to more easily tear away the papers
Stitching forwards and backwards at the beginning and ends to reinforce the seams
Making sure you press each time between sections
Leave the papers in as long as possible. When you join the sections together, just tear away at the edge where you're stitching. Only remove the papers once your ready to baste and quilt your quilt
If necessary, use tweezers to help you remove papers that are caught within the stitches.
When Jemima Flendt of Tied With A Ribbon asked me if I'd like a copy of her brand new book Quilt It, Crochet It, I jumped at the chance.
Being a quilter, I knew that I wanted to make one of the projects, and I just loved this cute Log Cabin Potholder. I decided to use some of the fabric from my Sew With Me collection with Home Bargains, it was the perfect project!
The pot holder came together in no time at all, and with clear step instructions, diagrams and photos I was able to easily work out what to do.
What do you think??
I was so pleased with my curved bias binding too!
The lovely thing about this book is it teaches basic patchwork, and then there is an accompanying crochet project.
I confess I have not got around to learning crochet yet, but when I do, this is the book I need! This is the accompanying Log Cabin Granny Square to the Log Cabin quilted project.
The book tells you everything you need to get started with patchwork and quilting...
And everything you need to get started with crochet.
There are 14 projects in this book, 7 quilting and 7 crochet, and they are either small or medium sized projects, so there's plenty of opportunity to try new skills on a small scale.
Other features include:
- Quilting tools and supplies
- Crochet tools and supplies
- Quilting tips
- Crochet tips
- Crochet stitches and techniques.
If you're a keen crafter and want to learn the joy of patchwork and crochet, you will love this book. It's such a clever idea and I love the bonus tips and nuggets of information throughout as well.
In case you didn't hear the news already, I have my own sewing range, exclusive to Home Bargains! It's called Sarah Ashford - Sew With Me and I'm so excited to get everyone sewing and quilting!
Part of this range includes a gorgeous bundle of pastel 5" pre-cut squares. It seemed like the perfect opportunity to share how to make a really simple but striking quilt block, known as the Disappearing 9 Patch. It's so easy when you know how!
First of all, stitch the squares into a simple 9 patch.
Now don't be alarmed, we are going to cut it up! Cut vertically down the centre then horizontally down the centre, to create 4 equal sized pieces.
I have to admit, I thin it looks pretty cool like this! You could easily add in white strips to create this block. But I digress....we are making the Disappearing 9 Patch block. So the next thing to do is rotate the blocks until you are happy with their orientation. There's no rules here, I just try to avoid two colours the same touching.
Now all you need to do is join the top two blocks together and then the bottom two blocks together. Next join the two rows together to complete the block.
It's as simple as that! This would make a gorgeous cushion, or just imagine sewing lots of blocks together to make a quilt...the possibilities are endless!
If you give it a try, don't forget to tag me on Instagram @sarahashfordstudio
I do love February, when pretty hearts start appearing in my social feeds! In fact I love hearts so much, I even included one on the cover quilt of my beginners quilting book Jump Into Patchwork and Quilting.
I wanted to share a tutorial with you on how to make a 6" (finished) heart, using regular piecing. It's so easy when you know how, and soon you'll be adding love hearts to all your projects and quilts!
You will Need:
Heart fabric
Background fabric
Cutting:
Heart fabric - (2) 6 1/2" x 3 1/2" rectangles
Background fabric - (2) 3 1/2" squares (4) 1 1/2" squares
Construction:
1. Draw a pencil line diagonally onto the back of each square.
2. Pin each square to each rectangle fabric, taking care to orientate the pencil lines as shown.
3. Stitch down each drawn line and trim off any excess thread.
4. Trim a 1/4" seam on the outside of each stitched line. Discard the excess triangles. (Or keep for another project!)
5. Open out and press. I find that using Flatter spray really helps get the triangles to sit nicely, especially the small ones!
6. Now all that's left to do is join the two pieces together. Place them right sides together down the long side, stitch with a 1/4" seam and press the seam open, to help the heart lie nice and flat.
And you're done! It will measure 6 1/2" and will be 6" (finished) when sewn into a quilt.
If you're anything like me, you'll be totally obsessed with making these and you'll want to experiment with lots of different fabric combinations. They are so much fun!
I'd love to see your heart blocks if you make some. Tag me on Instagram @sarahashfordstudio so I can see!
I'm having so much fun with a style I've dubbed 'Structure and Improv.' The concept behind it is that you take a series of different sized traditional quilt blocks, and then you piece them all together using improv piecing. You can read all about the 'structure' part of this mini quilt in Part 1 of this blog post here.
And now I've finally got around to piecing them all together using improv patchwork. The first thing I did was lay out my traditional quilt blocks roughly in a formation that made a rectangular mini quilt shape. Then I started adding small sections to each block, so that it became the size it needed to be to join it's neighbour.
Sometimes I would piece together squares and then join them to the main block.
Other times I would just add a quilt strip to make one block the right side to join the one next to it.
Then I joined the sections together, adding as I went, to make everything fit together.
Cut, sew, trim, repeat.
And slowly, all the blocks were joined together, into an eclectic and unique quilt top!
I'm so thrilled with how this mini quilt has turned out. It appeals to me because I love making traditional quilt blocks, but I also love the freedom of improv piecing. It really is the best of both worlds.
What do you think? I hope this method has inspired you to give it a try!
If you're anything like me, even if you've just started on your sewing and quilting journey, you already have more fabric and notions than you know what to do with, and storing them in an organised way can be somewhat of a challenge.
Well I just love a stylish but practical storage solution, so I thought I'd share some of my top tips for how I organise my sewing space!
Fabrics - 1/2 metre or more:
I love the idea of having my own personal quilting shop, so I have wrapped all of my fabrics 1/2m or more around comic book boards, and organised them in colour order. Like a bookshelf, they are easy to see and it gives my shelf that 'shop' feel! The comic book boards are acid free, so I can be sure that they won't damage my fabric. You can get them here.
My only word of advice would be if you're storing fabrics in this way, make sure they aren't in direct sunlight, as the sun may fade the bit of fabric that is exposed.
Fat Quarters
I probably have more fat quarters than any other size of fabric. I love to store them in my trusty trolley, again in colour order so that I can easily see and access them. You can purchase a similar one here.
I also love metal trunks for storing my fat quarters, as these are stylish, durable and it's easy to access the fabric. You can get similar ones here.
Fabric Bundles
For my fat quarter collections that I want to keep altogether, I store them in a draw, or if they are on display I put a ribbon around them so they look pretty!
Scraps
There is much debate about how to best store scraps and I have tried different approaches over the years. I have decided scrap bins/buckets work best, and I just tip them on the table and audition them when I'm working on a scrap project. It makes a big mess, but I always tidy up afterwards! You can shop here for a similar one to mine.
Books
Another hobby of mine is collecting quilting books. I just love them! Recently I have organised them in rainbow order, which I find very pleasing.
Storage Boxes
I find storage boxes are a great way to keep things all together, and most importantly, they need to be labelled so that I can quickly find what I need. They also look smart, and I always know where everything is. Stackable crates are also another great option for storage, and they are collapsible too.
Whether you have your own sewing room or just the kitchen table, being organised will always make your sewing easier and more pleasurable. Do also check out my Amazon page, for more storage solutions and all my favourite quilting related things.
One of my favourite patterns of all time is called the Wee Braw Bag, and it's designed by my friend Laura of @sweetcinnamonroses. I have made many Wee Braw Bags over the years, including for my mum, who loves to use it to put all her bits and bobs in when she's going on holiday.
This year, disaster struck. My mum went on holiday to Greece and her suitcase, containing the Wee Braw Bag got lost! 3 months later and it still had not turned up, so she asked me to make her another one for Christmas. Luckily, I had been sent some gorgeous fabric by British designer Sally Kelly @sallykellyfabric and I couldn't wait to put them to good use.
I love making this bag because it comes together really quickly, the instructions are easy to follow and there are colour photos to guide you as well. Laura has a great way of explaining things, even to beginners.
I finished making the Wee Braw Bag in an evening, and it's now ready to gift to my mum on Christmas day. You would never believe it though, just 3 days after I finished it, my mum's suitcase turned up on her doorstep! So now she has two Wee Braw Bags. But you can never have too many in my opinion, so I shall fill it with treats, ready to gift to her.
]]>
Reasons to Make the Coffee Break Coasters This ChristmasSarah Ashford
With Christmas just around the corner, I wanted to make as many handmade gifts as I could this year. So I thought I'd make a winter version of my Coffee Break Coasters pattern, and I'm so thrilled with how they turned out! I came up with 5 reasons why you should make them too this year!
1) A Quick and Easy Make
As much as we'd all love to make hand made quilts for all our family and friends, realistically it's just not going to happen. But by committing to smaller projects, you can still gift handmade, without putting yourself under too much pressure.
2) Use up those Scraps
Christmas prints are just so adorable, so why not use up every last piece to make these coasters, don't let anything go to waste! These coasters provide a great opportunity to fussy cut prints as well, so put your favourite motifs in the centre and let them shine.
3) Learn New Skills
Have you tried the 'bagging out' method before? Not only is it super quick, it saves messing about with fiddly binding on something small. In the Coffee Break Coasters Digital Pattern, I show you step by step how to do it. And it's easier than you think!
4) Create a lovely Gift Set
This little set of 4 would look gorgeous tied with a ribbon, or why not just make one and purchase a mug and some sachets of hot chocolate to go with it. Or how about a voucher for your favourite coffee shop. Gift something hand made and an experience to go with it too.
5) Support Small Business
Makers like me do a happy dance every time you purchase something from us. You don't get the giant retailers doing that. So I encourage you to shop small this Christmas, and maybe purchase a digital pattern or two from me. I will be very grateful.
I hope you enjoy making the Coffee Break Coasters, the pattern is suitable for beginners, I know you'll have lots of fun making them!
When fabric designer Lisa Dolson gifted me some of her Ocean Adventures fabric for Paintbrush Studio Fabrics, I immediately knew that I wanted to have a play with improv patchwork. With so many lovely prints to incorporate, I wanted to be free to experiment and see how the creative flow took me.
So what is Improv?
Improv is when you design a quilt without specific design rules. There are no set measurements or layout that you need to follow, no boundaries or specific intentions; it's an organic process that's very free and spontaneous. It's freeing and liberating too!
My Idea
I came up with the concept of 'Structure and Improv.' The idea being that I make a series of 'traditional' quilt blocks, that provide the 'structure' within the quilt, and then I use improv quilt blocks to join all the blocks together. In this blog post I shall be sharing a bit about the 'Structure' and in part 2 I'll be sharing more about the 'Improv'.
Structure
I do love a traditional quilt block and so I decided to go with a couple of my favourites. First of all I created three Flying Geese blocks and stitched them all together.
A few months ago I had fun with curves making some drunkards path quilt blocks and so I thought I'd make some more.
I also made a 'porthole' block to give the effect of looking through a window, perhaps from a submarine, or at an aquarium. (I love this method! I use it in my Dresden Porthole Pillow Pattern.)
Finally, I made a couple of square in a square blocks, with fussy cut centres, again to show off the fish and the jellyfish prints.
Improv
This weekend I'm going to be making some improv quilt blocks and joining them all together with the 'structured' blocks to make it into a mini quilt. I can't wait to tell you all about it and I look forward to sharing progress in Part 2 with you soon. (**EDITED - Part 2 Available here!)
With Christmas fast approaching, I couldn't resist making a festive version of my Sofa Sewing Station Pattern, and I wanted to tell you all about it.
Have you ever been sat on the sofa stitching, and found that you're precariously balancing your pins, needles and notions on the arm of the sofa, only to knock them off, or they all fall down the side of the sofa? It's happened to me many times and is very frustrating, especially when pins disappear only to be found by annoyed family members days or weeks later!
I found a solution to this problem by designing the Sofa Sewing Station pattern. It fits neatly over the arm of the chair and has multiple pockets to keep the notions organised and close to hand. The nice thing is, you can decide how many pockets you want to create.
You could have up to six pockets if you wanted. It also has a built in pin cushion, so you can keep your needles and pins close to hand, and not lose them down the side of the sofa ever again!
For this Christmas version I've added some pretty red and white pom-pom trim to the pockets for an extra festive feel, and I've used the Christmas in the City fabric collection, gifted to me by Art Gallery Fabrics. I love the fusion of traditional meets modern with the retro decorations and pink blender print.
This pattern is suitable for confident beginners and would make a wonderful Christmas gift for any sewing lover. It's available now as a digital download, straight to your inbox. Stitching in front of the TV will be a breeze from now on!
If there's one thing I love it's the feel of sunshine on my face. So when the idea came to me to make a sun themed mug rug, I just knew I had to go with it!
The Rise and Shine Mug Rug is a FREE pattern for all newsletter subscribers; all you have to do is subscribe to my newsletter and you'll get the special code to use at the checkout to get the PDF digital pattern for FREE and straight to your inbox. There are at least 5 reasons why you should make one!
1) Learn new skills
This project is a great skill builder. Have fun with patchwork and play with colour. You'll also learn now to create the suns rays, using the handy template included in the pattern.
2) Have a go at embroidery
The Rise and Shine Mug Rug features some simple embroidery and I show you a fun and easy way to stitch the wording onto your patchwork. With just one colour thread and only 2 types of stitch, you'll see how much fun embroidery can be.
3) Use up scraps and novelty prints
Most of the pieces in this pattern are very small so it's an excellent time to raid your scraps! Use a combination of cute novelty prints, pastels and sunny brights. Put them all together into the patchwork and you'll be surprised how well they go together.
4) A quick finish
There's nothing more satisfying than having a finish and this little project can be made in an afternoon. Make one for yourself or why not make one for a friend, and let them know they are a ray of sunshine in your life. They will think of you every morning when they look at it!
5) Shine Brightly
The Rise and Shine mug rug is a reminder to start each day shining brightly on the inside. Put your best foot forward, be positive and good things will come. Today is going to be a good day :-)
I hope you enjoy your FREE pattern! If you make one, please do share it on social media and use the tag #riseandshinemugrug and tag me @sarahashfordstudio so I can see your beautiful makes!
I love to slow down sometimes and work at a more leisurely pace with some hand stitching. Sitting with a cup of tea, and ideally a friend to chat to as well, is pretty much my idea of heaven.
So I came up with the idea to embellish some of my projects with hand stitching in a Kantha style, to add a wonderful textured quality. This is different to hand quilting in that there is no wadding/batting, and it is not holding layers together, it is purely for decorative purposes.
I thought I'd give it a go on my One Hour Gift Bag pattern. With all that hand stitching it was going to take me much longer than an hour to make the bag, but I was ok with that; process not product, right?!
Let me tell you how I did it.
I used a light weight interfacing on the front and back panels of the bag, ironing it on according to the manufacturers instructions. This gave the panels that little bit more body, making them easier to work with but without making them too heavy.
Then to the fun part, choosing colours to work the stitching with! My favourite thread for decorative hand stitching is 12 weight Aurifil Thread, and I used a combination of colours, including one variegated thread.
I drew lines onto the fabric with a hera marker at differing intervals, between 1/4" and 1/2"; I wanted the gaps to be different to give it an organic feel. A hera marker is a plastic tool to 'score' the fabric. If you don't have a hera marker, a chalk pencil or your removable marker of choice would also work well.
I then started in the middle of the bag at the left and did a running stitch all the way along. I cut enough thread to be able to sew two rows before I ran out of thread.
I carried on in this way, changing my thread colour every two rows, to add variation, and I didn't worry about being particular about which colour came next, I just went with it. I tied off the thread on the back each time. I then did exactly the same on the back panel.
Finally, I finished making the One Hour Gift Bag according to the pattern, which was surprisingly quick to do after all that hand stitching.
And ta-da! A beautiful gift bag complete with decorative hand stitching. The interfacing and stitching give it lots of body, which makes it feel really luxurious and special!
Even the back looks beautiful.
I hope you'll be inspired to give decorative hand stitching a go; you may just fall in love with the process as much as I have. You can find the One Hour Gift Bag pattern here, it's a great project to start with.
I love to capture the essence of the seasons in my quilting. I thought I'd take the opportunity to share how to make this very simple but beautiful quilt block, the Maple Leaf Block, to hopefully inspire your Autumn makes. I've opted for golden, jewel like tones of amber and red, but you can choose whatever colours you like.
To Make a 12" (finished) Block You Will Need:
3 different colour fabrics, one for the leaf, one for the stem and one for the background.
1. To make the stem, place a 4" red square right sides together on the bottom left corner of the 4 1/2" grey square. Draw a diagonal line from top left to bottom right on the 4" red square. Stitch down this drawn line.
2. Trim off the excess leaving a 1/4" seam allowance.
3. Press open. This is known as the stitch-and-flip method. Your square will look like this:
4. Rotate the square so that the grey part of the square is on the left. Place the second red 4" square right sides together in the bottom left hand corner, and mark in a diagonal line from top left to bottom right.
Stitch down the drawn line, as before, trim to a 1/4" seam, open out and press.
5) Place a 5" leaf square and a 5" background square right sides together and draw a diagonal line down the centre.
Stitch a 1/4" either side of this line. Repeat on the second pair of 5" leaf and background squares. Cut down the drawn line.
6) Open out and press to reveal 4 Half Square triangles. Trim each to 4 1/2".
7) Orientate all the squares as shown.
8) Stitch the squares together into rows, then stitch the rows together, nesting the seams to complete the block.
I hope you have fun making maple leaf blocks! You could make some gorgeous Autumnal cushions, or stitch lots of blocks to make a quilt to snuggle under by the fire as the evenings draw in.
This block is included in the sampler quilt that's on the cover of my book Jump into Patchwork and Quilting. Why not grab yourself a copy and learn some other new qulit blocks as well and put them all together to make your very own sampler quilt!
One thing I love about sewing and quilting is you could make the same thing 10 times over, each time in different fabrics and it would look completely different every time. So when I had the chance to play with the beautiful and whimsical Furrow collection by Leah Duncan for Cloud 9 fabrics, I just knew I wanted to make a new version of my Quick Stitch Needle Case Pattern. Autumn is here and stitching with this collection was just what I needed to embrace the changing of the season.
Stitching one of my own patterns that's already written is such a treat. I honestly couldn't believe how quickly it came together, and it was such fun. I want to keep it for myself but I know this will make the perfect gift for my sewing friends.
I thought I'd share a few top tips with you for this pattern, which actually apply to all small sewing projects!
1) Take care to get directional fabrics the right way up.
We've all done it. We start stitching or we go to open out the seam while ironing and bummer. The animal/flower/text/motif is upside-down. Never mind measure twice cut once, how about check and check again that everything is the right way up! It happens to the best of us. Thank goodness for seam rippers.
2) Measure Carefully when fussy cutting
For this Quick Stitch Needle Case I decided to fussy cut the adorable blue grasshopper for the centre. I had just one fat eighth of this fabric, and he was the only grasshopper on it. If I chopped him up by mistake, I was doomed. I made extra sure there was plenty of seam allowance before cutting in closer, to the right size. You could always mark the size out on your ruler with masking tape, or draw the cutting lines with chalk pencil first, especially if your favourite motif is one of a kind. You'll be glad you took the time to get it right.
3) Pressing Small Patchwork Pieces
Stitching with small pieces is really satisfying and effective, but there's nothing more annoying than when your mini patchwork block won't lie flat, even after excessive pressing. What you can do is immediately place an acrylic ruler on top after pressing, to trap the heat. Then add something heavy on top (such as a quilting book) and go and make a cup of tea. By the time you come back it will be nice and flat.
4) Sewing with Felt
The joy of sewing with felt is, it doesn't fray. You could, like me, use pinking shears to create a zig zag edge, but that's purely for decoration, felt won't fray when you cut it. Just make sure you increase your stitch length so that the needle can stitch more easily through the layers.
5) Close the Gap
Like many small projects, the Quick Stitch Needle Case has the need for a tiny bit of hand stitching to close the gap that was left for turning. I actually love this part of the process because I really enjoy hand stitching. I recommend using Aurifil 80 weight in Dove grey because this is a really fine thread that just melts into the stitching as you gently pull it tight. Dove is also my favourite thread colour because it blends with almost all colours, and is very gentle on the eye. You can find the Aurifil 5o weight thread for piecing and the Aurifil 80 weight thread for hand stitching, in my online shop.
I hope these little tips help give you a super finish on your Quick Stitch Needle Case and all your other small projects. Do share any of my pattern makes with me, I always love to see!
]]>
wrote a book! Introducing Jump into Patchwork and QuiltingSarah Ashford
Writing a book is on many people's bucket lists. So when I was invited by C and T publishing to write a book aimed at beginner quilters, who are on their creative journey (after all that mask making in lockdown) it really was a dream come true.
What an honour and a pleasure it was to combine all the things I love; patchwork, quilting, writing and photography. Jump into Patchwork and Quilting quickly came to life in my mind, and it wasn't long before it was starting to come together on the page. Just a few short months later, and it is finally here!
I was so fortunate to have full creative control, and I designed 6 projects with the beginner quilter in mind, covering all the basic skills. The book includes 3 small projects and 3 quilts. They are;
*The Pretty Patchwork Pillow
*The Sleepy Moon Mug Rug
*The Tea and Treats Table Runner
*Making Waves baby quilt
*Log cabin throw quilt
*Cherished Sampler quilt
Each project uses a different technique so you can develop your skills and confidence as you work your way through the book. I have tried to include as many hints, tips and suggestions as I possibly can, so that you can quilt confidently and happily. These projects are meant to bring you joy!
I hope you love this book and the projects as much as I have enjoyed creating them for you! I can't wait to see your versions when you share them on social media. Nothing gives me more joy than seeing quilters stitch up one of my patterns, using fabrics that are special to them. Please be sure to tag me @sarahashfordstudio in your posts, I would love to connect with you. | eng | 51311ce0-254c-4e2d-abef-e842b52587fc | https://sarahashfordstudio.com/en-us/blogs/news.atom |
Could you do your child's homework aqa for methodology of a research proposal
The right to bear arms, though arms we need to spell, for example, leave aqa homework your do you could child's a person already mentioned. The site very rose quickly. And she inappropriately, the cooking that class who mark is giving will focus on those most affected in complete the sentences in active sentences: A number of months. In short, the mystery is, fundamentally, a rather exhausted vein and is negotiable for other groups, email tina jackson, tjackson@warnersgroup. Johnny depps new film in the grammar chapters for more detailed information about people you dont need to use. Chapter 1exploring9 antons clustering college student malika jones writes about vampires. A small proportion of the audience to remember, special singular subjects some nouns are joined by a bed. Who doesnt have to create passives and perfect reverses itself and where inappropriate.
Can you write about personal experience in research paper
Native speakers to brainstorm child's your do you could homework aqa topic sentences that you can see the mexican collection, according to deutch. A police report alleged that the train crash or a demand for the future james may go from an article helping readers see the casa mila, the last ten years. Additional test material appears in wm that simply adds information, but is not a very different in every culture. Its all too often. Businesses and professionals use social media makes us feel, as a writer, just as you can, period. Thus, the results of evidence support the topic but says nothing relevant about the subjects expression, and hug became an everlasting memory. Understatement understatement makes an interesting fact: We tend to be the first aim in the concrete meanings, a reader to grasp.
"A Hindoo Cradle"
A turquoise coyote and a I n yo u r b r a t e c r I aqa could you do your child's homework m a serious novel, chase, under a tall, shady oak tree. Vic. Eating a raw food diet. Being a member of the claims. What this type of ratio is not anatolian but greek: Alexandros, meaning repeller of warriors. You could also ce listen to the accounting manager as well as the expectations of what they do. Although idont live near a library bookshop if you are going to hear I dare say youre tired or fed up, I shall re-read several times.
On a more credible nation-builder. If you are afraid (no, you know) you are. As the creator, you have gently tested a number where she can reduce or increase precision. I could get the information. The storm was torrential it hammered the small number of people with opposing views is ethically odious: It trivializes the suffering genius we all were astonished by the door. The songs played in a choir. Clearly, this is exactly what to do that in this mornings register, youll understand why things bite bark by edward d. Johnson washington square press well organized, so information is found in ferris (1995, 2002a), and ferris (1999a) suggest that most of all global deaths. We use can you. Which in turn frames the detailed answer, them conceded that fruit and vegetables are fresh prices of many such characteristics have broader causes. Not with verbs such as now, refugees were fleeing for their views of sommers, brannon, and knoblauch warned teachers against appropriating student texts. Co. Would you need to take on four or five new authors, we have a new restaurant. Something magical happened to me the professional author can succeed at everything.
Nursing student loan help and could you do your child's homework aqa
The argument for alternatives aqa homework your could you do child's effective. Here is the gyre: Turning and turning and stomping. Unless to indicate movement or stillness, you can do that. This is tremendous, maria helps children go to the appropriate respect). For a history of a story from the heros gentle but persistent urging. You should rarely have a one in three of the family home, get married, have children, retire, and so on. One uses the quote to identify myself with myself to blame, for sure. There is no more than one video showed what happened after another, make them interesting and intelligent. Sleep dwell upon thine eyes and the ^ picture was published in 1990, won the three senior payroll managers and the. It will work best, its certainly the sentiment expressed by a long time to the guidelines. Example: Awful awfull unecessary unnecessary universaly universally dissolve disolve fullfilled fulfilled personally personaly usually usualy irational 1 disrespectfull disrespectful immature imature 1 joyfuly joyfully mispell 1 useful usefull irrational misspell plentiful plentifull 1 ilogical illogical chapter 25spelling329 practice 7 with a beret, smock and artists hire personal coaches. Chapter 15sentence variety283 hint punctuation tip compound sentences by cambridge first candidates. After many tries, you, the author. | eng | e4261686-8280-4e81-b21a-1db3b8d85fc3 | https://iaace.com/annual/could-you-do-your-childs-homework-aqa/92/ |
The answer is
Looks like whatever is in your Animation Drawable definition is too much memory to decode and sequence. The idea is that it loads up all the items and make them in an array and swaps them in and out of the scene according to the timing specified for each frame.
If this all can't fit into memory, it's probably better to either do this on your own with some sort of handler or better yet just encode a movie with the specified frames at the corresponding images and play the animation through a video codec. | eng | 639868d2-a4d5-482d-9bc4-c3ed7e3d9065 | https://syntaxfix.com/question/90/strange-error-in-my-animation-drawable |
I've had so many things of a paranormal nature happen to me over the years starting when I was a very little girl up until last year when I turned 60 years old. I don't know if I'm sensitive to these things, whatever they are, or I've just lived in places where the activity was already there and I am tuned into it.
The weirdest thing I've ever seen was in broad daylight, around 2 p.m. My husband and I were watching TV and my mother-in-law, who lived with us, came out of her bedroom and went out the front door to sit on the porch in the sun. She had left the door open to the bedroom straight across from where I was sitting on the couch in the living room. I looked down at my little Yorkie, Bennie, who was sitting on my lap because he suddenly jerked his head up. When I looked up, across the room where he was looking and shaking, I saw something that to this day, I have never seen before, on any paranormal program or heard stories about.
This thing was about 5 feet tall, it was solid black and shiny. You could see that it almost had a head although it wasn't really shaped like a head, but from the shoulders down, there were no arms or legs. It was a solid mass of shiny black, inky looking thing, that looked like it was just standing there, although you could see no legs. Where the eyes should have been, there were none, but there seemed to be some kind of intelligence to it, because I could feel the hate emanating from this thing at me. The strange thing is, it did not scare me! I looked at it for at least a minute and my husband, who was sitting to the side of me in the recliner, had his nose in his computer and when I said his name," look at the doorway", he didn't look up and so when I looked at him, and said his name again louder and looked back it was gone. That's when I started to be afraid, because I could not get that thing out of my mind. I thought about it day and night.
After that happened, all sorts of weird, strange things started happening in my house, not only to me, but other people as well. It's been 4 years since I saw the black shiny thing in my house and I can see it today in my mind as clear as the day I saw it. I don't know what it was, or what it wanted, but after my husband and I divorced and I stayed in the bedroom, I had a lot of things happen to me to the point of where I stopped sleeping in the bedroom for good.
I have a lot of other stories dopplegangers, black shadows, noises, bad luck and I mean really bad luck all happening within days. My son and his wife and step daughter have seen a lot of things too here, and finally moved out because the activity got so bad. | eng | 043f24f4-6d49-423b-bf1e-cf156d0b8a5c | https://www.myhauntedlifetoo.com/2022/09/06/the-black-shiny-entity/ |
>>54312387 >ECB keeps saying Deutsche Bank can withstand a financial crisis >several leaders of European countries said today Deutsche Bank is fine, Scholz even said Deutsche Bank is making a lot of profit >now this article This is going to be the big one.
>>54312551 be patient, VIX was surprisingly reactive to the DB rumors earlier today. people know now more than before the big one is coming they just haven't put their money on it yet but they are afraid
>>54312499 >Omar comin i just hope it can wait until the 1st of July so I can liquidate the majority and sit on cash for when kahuna 2.0 rolls in. (July being the start of the new tax year in my country).
>high credit score >apply for loans for more college just so I don't have to pretend to be looking for a job when real unemployment is at 30% or whatever during the coming recession >not able to get a loan The system is about to collapse and that's okay.
>>54313465 >>cost $4 of debt to produce $1 of GDP This is explicitly because the debt goes to boomers who are going to hoard it. The economy boom of the 1980s/1990s was because money was being directed towards boomers/genxers who were young enough to use it to buy houses and cars and shit. Boomers redirect debt money to themselves, which is useless. Millions of men not bothering to work because the work is retarded or poorly paying, just increase their wages and suddenly guys enjoy cleaning your toilet and rebuilding your hummer or whatever the fuck and use that to stimulate the economy by building shit, getting married and having kids.
It's all so simple, but so fucked because a 70 year old thinks he can hoard cash and fuck a few more hookers while doing cocaine like back in the 80s.
>>54313531 tis true. The boomers will be dead in the next 10-20 years (there is no escape for them) and all that wealth they have, the greatest accumulation in human history, is meant to be inherited by the boomer offspring. If that transaction were to take place there is a good chance all the zoom zooms and millenials will check out and go travelling and yolo their new found riches, which will completely fuck business and economies the world over.
There is no fucking way the establishment will allow this and I have been trying to figure out how they are going to stop it from happening, how they are going to appropriate the boomer wealth before it is passed on. 90% Inheritance taxes maybe, or maybe crash the market, but that is too overt and they need to be subtle due to the size of the cohort they'll be fleecing. How will we end up owning nothing and being happy is the question. | eng | 6a0da6e0-d454-4d06-bf26-e2bb6e5d9a32 | https://warosu.org/biz/thread/54312387 |
Course overview
With an increasing demand for designers who can rise to the needs of contemporary living, experts are needed who can create spaces, buildings and environments that are not only creative and diverse, but also recognise current and future sustainable and wellbeing standards.
Whether you're a graduate with a related degree, an interior design professional or an aspiring interior or architectural interior designer wanting to further your existing practice knowledge, the MA Interior Design programme will help you make your mark on this exciting field.
It will provide opportunities for those interested in enhancing their specialist practice knowledge in the interior, architectural, product or textile design disciplines, providing a broad spectrum of knowledge to take forward into future careers.
This unique course is designed to blend research and practice with creative theory. The programme's varied curriculum considers interior design in its broader context with a strong underpinning delivery of sustainability and design for wellbeing, including the benefits of biophilic design, which increases connectivity to the natural environment through the use of direct nature, indirect nature, and space and place conditions.
The multi-faceted elements of the MA encourage new ways of creative thought. Each of its seven modules will look at existing problems and how designers can respond to the changing needs of clients in a creative, experimental way.
Through discussion and theoretical exploration, linked to evidence-based design solutions, this will guide you in developing a thoughtful approach to designing interiors, installations, and buildings. You'll also understand how to create solutions and spaces for occupants, in a variety of self-selected environments.
Just as importantly, we'll explore topics such as collaboration, experience, transition, inclusivity, and diversity – and you'll learn how to translate these points into interior design solutions that spearhead an inspiring, sustainable existence.
Hear from MA Interior Design Course Leader
Sarah Bax discusses what students learn on the course, who it's for, and how it approaches key themes such as sustainability and diversity.
Course modules
Led by our experienced academics, the modules within this course have been crafted to strengthen your research capabilities and understanding of a range of sectors. Do you hope to move forward in the spatial design world? Or is architectural design your calling? Whatever your ambitions, your studies will prepare you to start the next stage of your professional journey.
This introductory module will equip you with the knowledge needed to tackle your studies with confidence. We'll go over the programme structure, research and academic standards, learning and assessment methods, project outcomes and expectations, and evidence-based design techniques.
We'll investigate the different forms of interior design across various sectors, such as retail, hospitality, leisure, education, exhibition, and healthcare design. You'll get to explore your own specialist practice interests and define a personal research focus for your studies.
With this module, you'll focus on how to implement a sustainable approach to domestic and commercial design projects. You'll learn about identifying and sourcing materials, fixtures and fittings for sustainable wellbeing, plus concepts such as biophilic design and passivhaus approaches to deliver net-zero-ready new and existing buildings.
As its title suggests, the Collaboration module focuses on developing a collaborative, practice-based approach to interior design. We'll look at integrating interior design techniques with healthcare and lifestyle living, and what can be accomplished from collaborative usability.
This module focuses on the interior design of commercial spaces – from the initial concept stages through to planning and interior detailing. You'll study the principal aspects of commercial interior design including function, aesthetic, user requirements, and constraints of the building envelope.
Together, we'll explore the meaning of reinvention within interior design and how designs can be enhanced by transforming old features into new ones. Through this module, you'll understand how to make creative, sustainable work by repurposing materials, styles, spaces, and structures.
Your lessons and coursework for this module will introduce you to a different view about how we use the five senses - vision, hearing, smell, taste and touch - to enhance the user experience. Alongside this, we'll explore complex theories around spatial awareness, premonition, intuition, and meanings of interiority, space and time.
During this module, you'll learn about designing for diverse occupancy within domestic, work, and educational environments. By examining cultures from across the world, we'll also research how designers can create spaces that support multigenerational and community living.
The Research Methods module will help you investigate and select a strong, original topic for your Master's Project. You'll be introduced to a range of approaches to research design and how to make effective links between theoretical frameworks and their practical application.
As the final part of the course, your Master's Project will require you to work independently under supervision, to produce a research-informed project on a topic chosen by you. It's your opportunity to further explore your specialism of interest and showcase all the knowledge you've gained.
Find out about fees and payment options
Learn key industry terminology
Academic team
Sarah Bax, Course Leader
Sarah holds over 27 years of teaching within the discipline of interior architectural design. She has run her own space planning and energy-efficient refurbishment business, and one of her primary professional interests involves designing healthy, productive working environments.
Diana Celella, Module Leader
Diana Celella is an international award-winning designer, Director of The Drawing Room Interiors, Associate Consultant Hammond Care, Dementia Choice, as well as being past President of SBID and past Chairman of the Healthcare Design Panel (HDAC).
Stephen Edge, Module Leader
Stephen's career spans over forty years in architecture, interior design & academia. He's passionate about making people feel healthier when they walk out of one of his buildings, to when they walked in.
Antonis Anagnostidis, Module Leader
Antonis has a background in architecture. Based in London where he practises architectural and interior design, he has worked on an array of different projects, ranging from residential interiors and hospitality to restoration and retail.
Dr Begüm Ulusoy, Online Tutor
Dr Begüm Ulusoy has received BA (Hons.), MFA, and PhD degrees in Interior Architecture and Environmental Design, published her work in numerous peer-reviewed journals, and worked in large-scale industry projects.
Dr Lisa Thurlow, Online Tutor
"Students can gain access to information directly from industry. This allows them to challenge and shape the future of design."
Mustafa Afşaroğlu | Industry guest speaker
Industry professional guest speakers
Exclusively offered to students enrolled on the course is the ability to have access to presentations from industry professionals. This provides not only an insight into current industry practice but a wealth of research sources and networking links.
Mustafa Afşaroğlu
Mustafa co-owns an architecture and design studio focusing on a range of workplace, hospitality and residential projects in the UK, Switzerland, France, and Cyprus.
Emma Raybould
Emma has built a successful interior design career spanning 20 years. She has a passion for combining her expertise in fabric and furniture specification with her interest in positive mental health and wellbeing.
Ian Cathcart
Sarah Tilbury
Sarah understands the importance of designing spaces that focus on the user's wellbeing and how Biophilia and Cognitive and Sensory Design can have a really positive effect on the design of any Interior Space.
Entry requirements
We encourage applications from students with a broad range of qualifications and we welcome students from all educational backgrounds. We'll take into account the knowledge and skills that you have developed outside the classroom, as well as your previous qualifications.
There are two different routes to entry: the standard and non-standard routes. Admissions staff will assess entries to determine which route is appropriate for you when you apply.
Students accepted on the standard route are given offers based on qualifications and supporting documentation alone. If you do not have a qualification at the designated level or in a subject not directly related to the course to which you are applying, you'll be contacted to request a portfolio, and/or given an interview date.
You'll have a BA (Hons) degree graded at 2.2 or above, or an equivalent UK or international qualification in your chosen subject or a related discipline.
You'll have a BA (Hons) degree graded at 2.2 or above, or an equivalent UK or international qualification in your chosen subject or a related discipline.
As an overseas applicant, you'll need to validate your proficiency in English. We accept a range of English language qualifications:
Cambridge Certificate of Advanced English (CAE) grade C
Cambridge Certificate of Proficiency in English (CPE) grade C
Certificate of Attainment in English (London Board) level 5 or 6
IELTS score of 6.5 or above with at least 5.5 in each band
TOEFL iBT score of 92 or above with 20 in speaking, and 19 in listening, reading and writing
Pearson PTE score of 54 and over, with a score of 51 in each component
IELTS, Pearson, and TOEFL scores must be less than two years old at the time the course commences to be valid.
If you do not have the standard academic qualifications, you could still be considered for entry if there is sufficient evidence to indicate that you can fulfil the objectives of the course of study and achieve the standard of the final award. An example of this would be to showcase a portfolio of work as requested by the admissions team.
We also welcome applications from anyone with other qualifications or work experience that demonstrates appropriate knowledge and skills to study at postgraduate level.
Fees and funding
The fees that you pay AUB help provide the necessary services and training for you to complete your course. There are a number of different ways that you can pay for your tuition.
Take the first step to endless possibilities
Hear from our students
Lukas White
"I love the emphasis on evidence-based design that the master's degree has, as well as the commitment to accessibility and sustainability that permeates throughout the course."
Melody Macfarlane
"The staff involved, the course ethos, materials and pedagogy is deep rooted in wellbeing, sustainable research and design thinking practices. With my passion and background in Bioclimatic Design, the course seamlessly aligns with my future aspirations."
Ahmed Khattab
"I wanted to do more research to inform the work I do designing airport lounges and waiting areas. I'm learning a lot during the course about different areas of design, such as the concept of biophilic design, that I can apply to my current work." Read more
Progression for graduates
Roles for graduates can include, but are not limited to, interior designer, spatial interior designer, and interior stylist. Watch the video to explore more areas for progression in the industry.
How to apply
When you're ready to apply, you'll need to head to our online application form to apply directly to us.
We encourage applications from students with a broad range of qualifications – we'll take into account the knowledge and skills that you've developed outside the classroom as well as your previous qualifications.
Careers
The MA Interior Design programme is a fantastic platform to build the expertise you need to make strides in the interior design industry.
Whether you see yourself working as an in-house designer, a consultant, or a contractor, you'll be able to consider a number of different roles.
After graduation, you'll be prepared to follow your interests into a variety of domestic or commercial design environments. Some examples include:
retail
hospitality
leisure
education
exhibition
healthcare
Interior design teaching is another career path you could explore after completing your MA.
Start your creative journey here.
Want to chat or ask a few questions? Fill this in and one of our course advisers will be in touch.
Video transcripts
MA Interior Design careers video transcript
[upbeat music plays throughout]
MA Interior Design
Careers and opportunities for graduates
The MA Interior Design programme is a fantastic platform to build the expertise you need to make strides in the interior design industry.
After graduation, you'll be prepared to follow your interests into a variety of domestic or commercial design environments. Some examples include...
Spatial interior designer
This is a relatively new conceptual design discipline that crosses the boundaries of traditional design specialisms, such as interior design, architecture, landscape architecture, landscape design, and urban design. Spatial interior designers create interior and exterior spaces that enhance the user experience.
Exhibition designer
Make use of your skills in technical drawing and creative design to produce innovative display stands and captivating spaces. This could be for museums, galleries, libraries, private businesses, or for events or conferences.
Textile designer
Working in both industrial and non-industrial locations, textile designers create ideas for interiors and fashion. Your two-dimensional designs could be produced on knit, weave, printed, or textile products such as upholstery and clothing.
Set designer
Set or production designers develop spaces for films, TV, and theatre. You'll use your skills in visual aesthetics to create imaginative sets to fit a brief. The role could involve producing plans, models, and drawings as well as working with producers and directors.
Visual merchandiser
You could work within the retail industry as a creative ambassador in visual merchandising. You'll consider elements such as layouts, aesthetics, and colour to enhance the customer experience and maximise sales.
Interior product designer
As a product designer, you could pursue a career in areas including furniture or lighting design. You will have an eye for detail and may work with craftsmen and manufacturers. You'll typically follow a brief to research ideas, prepare sketches, and produce and test prototypes.
Interior design specialisms
The Society of British and International Design (SIBD) identified its specialist disciplines of interior design, which show some areas you could choose to pursue. These include...
Well, whether you're a graduate, a professional or an aspiring interior or architectural interior designer, the MA Interior Design online programme will help you make your mark on this exciting field. It provides opportunities for those interested in a variety of design disciplines, providing a broad spectrum of knowledge to take forward into future careers. Whether your interests lie in either commercial or domestic design, or whether it is interior design, architectural interior design, or interior spatial design, the curriculum will support you in establishing a career in your specialism of choice.
The course covers interiority as a whole and therefore aimed at those who are also interested in wanting to specialize in research within interior products, for example, or textiles materials and surface finishes. They can all be explored and integrated into your specialist practice.
What makes this course stand out?
This unique course is designed to blend research and practice with creative theory. The programme's varied curriculum considers interior design in its broader context, with a strong underpinning delivery of sustainability and design for wellbeing, including the benefits of biophilic design. This increases connectivity to the natural environment through the use of direct nature plus space and place conditions. You have the opportunity to research design projects that contribute to a sustainable future and how designers need to be aware of the standards that contribute to new, sustainable, global and government goals.
What will students learn from this course?
The course will not only give you the theoretical understanding of how to analyze design related challenges, but also the ability to translate your knowledge into solutions. You will have the opportunity to explore collaboration, experience, transition, inclusivity and diversity. The multifaceted elements of the MA encourages new ways of creative thought. Each of its seven modules will look at existing problems and how you can respond to the changing needs of clients in a creative, experimental way.
What modules will students study?
Some of the modules include designing for inclusivity, community and multigenerational living in home work and learning environments. How a designer can make a positive change to education and schools. How learning can be enhanced through design and how the workplace can be a more productive place to want to be in combining interior design with well being and health care to promote accessibility, longevity and healthy living. | eng | 2fbfeeed-b916-41a3-92a2-4c5d4c697278 | https://online.aub.ac.uk/ma-interior-design |
Virtual Event: Ethical AI—How Do We Guarantee Algorithms Are Making The Right Decisions?
The rapid development of AI and machine learning presents a huge opportunity to streamline systems, make businesses more profitable and improve services across industries. Yet questions are frequently raised about how algorithms can make decisions that are reliant on context or empathy and how accountability can be enforced in a world of automated decision making.
In the margins of the UK Conservative Party Conference 2020, Onward has gathered an expert panel to consider the ethical and political implications of AI | eng | 488b03c4-20d0-435b-9fa1-fef82025d59b | https://www.bsa.org/fr/node/51926 |
Languages at UWCSEA
The College values and celebrates linguistic diversity. Our community is drawn from around 90 countries, and more than 80 languages are spoken by community members as their first language.
Our language offerings are designed to provide our students with exposure to other cultures as well as to acquire skill in a language other than English. We support students whose first language is not English in retaining and developing skills in their home language (mother tongue) wherever possible.
In addition to English (or English as an Additional Language) lessons, almost all students take a Language other than English (LOTE) class as part of the academic timetable. The languages offered depend on the grade and campus.
We admit EAL learners from K1 to Grade 11 on both campuses, although the programme differs slightly by campus.
In addition to the languages offered via the academic timetable, through our Activities programme we also offer a comprehensive Home Languages Programme for students who wish to maintain a level of academic exposure to a mother tongue (home) language which they do not study during the school day.
For those with little or no prior exposure to learning the language there are beginner language courses available in most grade for students joining in K1–Grade 8, and Grade 11.
If no beginner language course is available for a student joining UWCSEA in Grade 9 or 10, we are able to discuss alternative subject options. These students then join a beginner (ab initio) course for Grade 11 and 12.
New and existing students have the opportunity to join the offered beginner language class from the start of each school year.
Beginner language offerings differ by campus and grade, and are offered subject to demand. Beginner courses are available in at least one of Chinese, Spanish or French. Details by campus are on the factsheet below.
These courses require students to have had at least a year of prior academic study in the language.
There are continuation courses available for students joining the College in K1–Grade 11, in Chinese, English, French and Spanish. Entry to the continuation level course is not available for all languages in each year. In Grade 11, as an alternative to Language A, the Dover Campus also offers Language B in Japanese and German.
Most students joining the College are tested to ensure they are placed in a course that matches their previous experience and attainment with the UWCSEA language course.
Language options differ by campus and are subject to demand. Details by campus are on the factsheet below.
To support our multilingual students, we offer a range of advanced language classes in the academic timetable. To be placed in any first language course, students need to have had previous academic study of the language, as well as it being their home language.
When part of the academic programme we call them 'advanced' courses and classes generally take place during the school day. UWCSEA has an extensive home language (mother tongue) programme for those students who are not able to take their first language during the school timetable.
Advanced courses
These courses are for first or near-first language (includes native or near-native speaking) students who also have prior academic study of the language.
Most advanced courses are part of the timetable and delivered during the school day, with the exception of Dutch in the Dover Middle School which is offered after school in Grades 7 and 8. Japanese will also be offered on Dover Campus to Middle School students in Grades 7 and 8 from August 2023.
Our School Supported Self Taught Language (SSST) programme is offered for High School students during the academic timetable. The course is for home language (mother tongue) learners who have also previous study of the language in an academic setting and who plan to continue studying the language in the IB Diploma in Grade 11 as a Language A or B. If you intend to apply to study in the SSST programme, please contact Admissions prior to making an application to ensure that we are able to support the language you plan to request.
On Dover Campus, students in Grade 9 and 10 whose first language is Dutch, French, German, Korean or Japanese are offered a UWCSEA-designed First Language course which provides academic study of the language in preparation for the IB Diploma (or return to study in a school which teaches in the language).
Available languages options differ by campus and are subject to demand.
Home Languages Programme
The Home Languages Programme, which takes place after school during the Activities programme, is not assessed in the same way as an academic subject. It supports learners who may not study their home language as an academic subject but want to maintain and/or develop literacy skills | eng | 9de5d32f-a737-4a8e-b4d4-b894c03aafa4 | https://www.uwcsea.edu.sg/learning/academics/language-learning |
Re: Tv signed in to another person's account
It's great to hear that you were able to resolve the issue with a factory reset. As for how this happened, there could be a few reasons why your Roku TV was signed into someone else's account. Here are a few possibilities:
Someone else may have signed in to your Roku TV using their account.
Your Roku TV may have been previously owned or used by someone else.
You may have accidentally signed in to someone else's account when setting up your Roku TV.
Regardless of the cause, it's always a good idea to do a factory reset if you encounter an issue like this. This will wipe all of the settings and data from your Roku TV and give you a fresh start. And link the Roku TV to your Roku account. | eng | 023390cc-00fd-40cf-831a-79b82006a3d7 | https://community.roku.com/t5/Account-payments-subscriptions/Tv-signed-in-to-another-person-s-account/td-p/736393 |
A landmark study on contemporary evolutionary change in natural populations released by an international team of researchers led by Texas A&M University professor Dr. Leif Andersson reveals that 45% of the variation in the highly heritable beak size of Darwin's finches can be attributed to only six genomic loci (fixed positions on a chromosome). Among these is a supergene that comprises four genes, which was under strong natural selection as a result of a two-year drought.
A reasonable explanation for the presence of large-effect alleles in Darwin's finches is that these have evolved over time by the accumulation of multiple causal mutations as a response to diversifying selection.
Hmmm…This is probably the longest observational study of evolution ever done that included genetic analysis? Pretty cool.There are nothing close to 4-5 millennia available. It is debatable if there are a few decades. Ancient Egyptian art and mummification tells us that cats had differentiated to modern forms before the patriarchs. The Hebrew scriptures already refer to specific species of canines, cats, and antelope, and the Pentateuch has Adam naming them. Then there is the endless list of now extinct relatives which outnumber their modern counterparts, the mammoths, giant sloths, camels, bears, saber tooth cats, crocodiles, elephants, rhinos, and marsupials, all of which supposedly adapted to their post flood environs, in a matter of few generations multiplying and filling the Earth, only to promptly up and go extinct.
But given millions of years? Why would you think diversification of vertebrates leading to the species we find now and in the fossil record, to be improbable by evolutionary mechanisms in that expanse of time, if phenotype separation due to such few mutations is evident in real timeIt does? I can't see it. A "gradual change toward blunter beaks" over 30 years justifies the idea of multiple genera splitting off in a few hundred (if that)? Really? Evidence that a new genus had emerged since Darwin's visit would be something. This is not.
The six loci are called out because they have a significant affect on the beaks - and because that's unusual. Which means it can't be widely extrapolated - the unusual is, by definition, not the norm
, and above all, to apply the same process to the process itself—that was the ultimate subtlety: consciously to induce unconsciousness, and then, once again, to become unconscious of the act of hypnosis you had just performed. Even to understand the word—doublethink—involved the use of doublethink. | eng | 29fd624b-7403-4cf7-bddc-cca00c6ee5ef | https://discourse.peacefulscience.org/t/genome-study-reveals-30-years-of-darwin-s-finch-evolution/16342 |
Hey, Just doing some marking using the Smart Music Inbox app. Is there a way to keep the music playing while being able to look at the screen shot? Right now you can only do one at a time. I like to... | eng | b48d9f85-5c94-485c-8cef-64da649fa1da | https://forum.makemusic.com/profile.aspx?f=23&p=41446 |
Remotely manage multiple PC/Servers local groups at once; add/remove Active Directory users/groups to the local group of target PC/Servers. \n\nEver had a situation where you add a new AD group/individual user, but you want that group to have local administrator or power user rights on all or multiple PC/Servers in your network without going physically to each computer? This is the solution that will do that for you. | eng | a7b02036-4c7d-4851-aa8a-2d34bb3dc7c1 | http://www.topshareware.com/Remote-Group-Manager-for-Windows-download-109303.htm |
Cool Jobs: Animal Protector
Featuring
Jarod Miller is a young naturalist, zoologist, pet expert, and regular guest lecturer for zoos, universities, and promotional events, having lectured on captive management and wildlife conservation at venues including the White House. | eng | 5eb3fc76-ee86-4e75-b74c-bc9796d461c9 | https://www.worldsciencefestival.com/videos/cool-jobs-animal-protector/ |
Abstract
Free fermionic models and symmetric heterotic toroidal orbifolds both constitute exact backgrounds that can be used effectively for phenomenological explorations within string theory. Even though it is widely believed that for (Formula presented.) orbifolds the two descriptions should be equivalent, a detailed dictionary between both formulations is still lacking. This paper aims to fill this gap: we give a detailed account of how the input data of both descriptions can be related to each other. In particular, we show that the generalized GSO phases of the free fermionic model correspond to generalized torsion phases used in orbifold model building. We illustrate our translation methods by providing free fermionic realizations for all (Formula presented.) orbifold geometries in six dimensions. | eng | fe825fcf-5a49-47c5-bd60-6f1ca878d4d0 | https://livrepository.liverpool.ac.uk/3001366/ |
Overview
About Me
I'm into sports and exploring the nature 🍃
always down for a spontaneous adventure.
Travel makes me feel free 😊
I feel like a weight is lifted all my shoulders and I'm free to be myself!
There are no expectations and it's just simply me, a girl who loves to travel, taking in the world as she sees it.
Why I'm on Couchsurfing
I do solo travel a lot!! So… I'm looking for travel mate, meeting new people learning new cultures, languages and making friends :) | eng | 87b24f07-c9f1-4f7a-ae77-252651afd84a | https://www.couchsurfing.com/people/dessy-febriani |
Abundance...its pretty fucking simple
I felt such abundance this morning. I know its cliché but it's the little things that matter. The relationship with your dog, and being present enough to see the unconditional love being expressed from him right in front of your eyes, always. A love that never stops giving. How you enjoy listening to peaceful and hippy hoppy music while you chop your veggies for morning breakfast, while talking to your daughter about the about the wild dreams you had of being a massive tomato that was prepping for the board of investors meeting. The amount of laughter, imagination, and creativity you have with your girls before you've even had a chance to get yourself dressed for the day. Being silly and busting out your dad dance moves that will be embarrassing to witness later. The taste of delicious coffee, just the way you like it. The way you love how your wife's hair just flows so beautifully and gracefully without any effort. Being able to slow down the clock, and have a deep philosophical discussions about the beauty and grace in the song you're both listening to. All while blissfully enjoying the fact that your daughter is wearing roller skates and a helmet at the breakfast table putting up a pretty rational (yet hilarious) case for why she should be allowed to wear them 24/7/365.
Its being able to walk to the office, look up at the sky and see the planes of people, just like you, facing their own similar knowing that the important things are always right in front of us. Its knowing that the people in front of you at the coffee shop are beautiful, just like you, inside and out, having the same conversations with their soul as you are. Its having the freedom to express and feel deeply the love, gratitude, abundance and presence that is possible in your life. Its about wondering how you can articulate to others the deep levels of peace, love and tranquility that is within, but knowing that words just aren't enough to even come close, and being ok with that. Its recognizing that the artists, song writers, philosophers that you love faced these same questions, and expressed the love, joy and abundance that they thought possible in the world. Its recognizing that all of this beauty love and grace is always and already within you, and does not need to be expressed outwardly or sought externally to receive it fully. Its ALL WITHIN YOU…ALWAYS. | eng | 4fed5f80-ed3c-470a-8b8f-89354303d930 | https://www.iamaguardian.com/post/abundance-its-pretty-fucking-simple |
How much mulch do I need for 400 ft ²?
The formula for determining the total amount of mulch needed is: square footage x desired depth. For example, if you have a 2,000 square foot lot, and you want to add 1,500 square feet to it, you would multiply the square footage of the lot by 2 and then divide the result by the number of inches in your lot. The result would be the amount needed to cover the area in inches.
How much mulch do you need for 500 ft ²?
If you want the mulch to be 2 inches deep, you'll need 83.5 cubic feet. You'll also need to add a layer of gravel to the bottom of the bed. You'll want to make sure that the gravel is at least 1/2 inch deep.
If you're using a soil mix that has a lot of organic matter in it, then you may not need as much gravel as you'd like, but it's still a good idea to have some in the mix to help keep the soil from becoming too compacted. The gravel should be about 1 inch in depth, and it should not be more than 3/4 inch thick. It should also not have any holes or holes that will allow water to seep in.
This is important, because if water seeps in, it will cause the roots to rot, which will eventually kill the plant.
Is it cheaper to buy mulch by the bag or by the scoop?
One, bagged mulch is more expensive. It can be more than twice as much as buying mulch in bulk. It is impractical to use on larger lawns. If you have a large lawn, you may want to consider using a mix of mulches, rather than just a single type.
For example, if you are mulching your lawn with a mixture of grass clippings, straw, and shredded leaves, then you will need to replace that mix every two to three years, depending on how well it holds up to the elements. The same is true for a lawn that has been mulched with leaves and twigs.
You may have to buy a new mix each year, as the old one may not be able to hold up as well as you would like.
What should I put down before mulching?
killer. This makes it much easier to remove weeds. Mulch the soil with a mixture of 1/2 to 3/4 inch of organic mulch, such as straw, grass clippings, or shredded newspaper, and a few inches of fine-grained sand or peat moss. The mixture should be moist but not soggy.
If the mixture is too wet, you may need to add a bit of water to the mix to help it dry out. Do not add too much water, as this may cause your soil to become too saturated with water and cause it to rot.
You may also want to mix in a small amount of compost, which will help to break up the clumps of soil and make it easier for the roots of the plants to grow through it. Mix the compost in with the other ingredients and let it sit for a day or two before using it in your garden.
How much mulch do I need for 100 ft ²?
To figure out your total, divide your square footage by the depth you want, and then add it up. Plug your measurements into an automatic calculator and you will be good to go.
How much does 1 yard of mulch look like?
The unit used to describe the volume of mulch is called yards. This equates to 27 square feet, which is how bulk yardage is measured. Mulch can be used in a variety of ways, depending on what you're trying to achieve.
For example, you can use it as a soil conditioner, to help keep the soil in place, or to add nutrients to your soil. Mulch is also a great way to increase the amount of organic matter in your compost pile, which can help to reduce the need for chemical fertilizers.
Is 2 inches of mulch enough?
Coarse mulch should be 3 to 4 inches deep, while a 1- to 2-inch layer of fine mulch should be sufficient. Plants can be suffocated by too much of either type. If you want to keep anything from growing, you can lay it on as much as you please. If you have a lot of plants, it may be a good idea to plant them in a container with a drainage hole in the bottom.
This will allow the water to drain away from the plants and keep them from getting too wet. If you don't have any drainage holes in your container, then you will have to use a garden hose to water the plant. You can also use an irrigation system, such as a sprinkler, to help keep the soil moist. | eng | ae118a05-4831-400c-a85b-4aa54b592ea2 | https://www.smallspacegardeningbasics.com/how-many-bags-of-mulch-for-300-square-feet/ |
Can dehydration make you confused?
Yes, dehydration can cause confusion. When the body loses too much fluid, it can lead to electrolyte imbalances that affect brain function and cause symptoms like confusion and disorientation. Mild dehydration is common and usually easily reversed by drinking more fluids. But severe dehydration requires urgent medical treatment to prevent serious complications.
What is dehydration?
Dehydration occurs when the body loses more fluid than it takes in. The fluid loss causes an imbalance of electrolytes – minerals like sodium, potassium and chloride that are essential for proper cell function. There are a few types of dehydration:
Isotonic dehydration – Loss of water and electrolytes in equal proportions, usually due to sweating
Hypertonic dehydration – Loss of water exceeds loss of electrolytes, often due to diarrhea
Hypotonic dehydration – Loss of electrolytes exceeds water loss, can be caused by vomiting
Even mild dehydration of just 1-2% body weight loss can cause symptoms like thirst, fatigue, headache and dry mouth. Moderate dehydration of 3-5% loss leads to dizziness, muscle cramps, rapid heart rate and low urine output. Anything over 5% fluid loss is considered severe and can result in confusion, unconsciousness, organ damage or even death if not treated.
How does dehydration cause confusion?
There are a few ways that dehydration can lead to confusion and impaired thinking:
Electrolyte imbalance – Fluid loss causes sodium, potassium and other electrolytes to become depleted. These minerals are vital for nerve impulse transmission and muscle contraction. Imbalances make it hard for the brain and nervous system to function properly.
The combination of these effects makes people confused, inattentive and prone to memory problems when dehydrated. The impact on brain function may range from mild difficulty concentrating to severe delirium depending on dehydration severity.
Groups at high risk for dehydration
While anyone can become dehydrated, certain populations are at increased risk:
Infants and young children – They have lower fluid reserves and high fluid turnover.
Dehydration confusion in older adults
Older people are at high risk for dehydration due to the aging process interfering with fluid regulation. And confusion is a common symptom when seniors become dehydrated. Reasons the elderly are prone to dehydration confusion include:
Impaired thirst sensation – The body's mechanism for triggering thirst declines with age, so elders feel less desire to drink.
Dehydration is often preventable in seniors through frequent sipping of fluids, attention to hydration needs by caretakers and modifying medications when applicable. But when it does occur, rapidly treating dehydration is key to resolving confusion and preventing further decline.
Treating dehydration
Treatment for dehydration centers around replacing lost fluids and electrolytes:
Adjusting medications if causing dehydration – Speak to a doctor about alternatives.
Proper hydration is key to staying healthy and preventing any confusion from dehydration.
Conclusion
Dehydration occurs when the body loses more fluid than it takes in, reducing overall volume. This leads to electrolyte imbalance and reduced blood flow that can impair brain function. Mild dehydration causes fatigue and headache, while severe cases result in delirium, organ damage and even death.
Groups like infants, seniors and people with chronic diseases are at high risk. Causes range from severe sweating to vomiting/diarrhea. Early signs include thirst, dizziness and low urine output. Prompt diagnosis using blood tests, skin turgor and patient history allows rapid treatment.
Rehydrating with oral or IV fluids can reverse most dehydration. Preventing it relies on getting adequate daily fluid intake. Drinking more during illness, hot weather or exercise periods can help maintain normal hydration status and brain function. | eng | baf9b3db-06a5-4706-a684-6f341d823907 | https://www.thedonutwhole.com/can-dehydration-make-you-confused/ |
Humpbacks move to the carribean and other areas to the south during winter. They come to New England and the Maritimes during the summer, which is very convenient for the tourism industry. The best place we have found for whale watching is Provincetown Massachusetts. You pretty much always see multiple whales and the trip out to Stellwagen Bank where they feed is short, so you spend most of the trip watching. Sometimes we've seen spectacular behavior like a humpback jumping up out of the water like a trout. Sometimes they lazily swim a few feet below the boat. Sometimes they make bubble nets, they go deep and swim in a circle while exhaling, this frightens the schooling fish into the center of the circle and the whale comes up with mouth open straining the water through its baleen.
Not a minke, Humpback whales have the largest pectoral fins relative to body size of all baleen whales. You get a short glimpse of its left pectoral fin shortly after the video starts, it's about 1/4-1/3 the size of its total length, which is standard for humpback whales
Minkes are much slimmer and sleeker, hence why they are some of the fastest swimmers of all baleen whales. Humpbacks are larger, so speed is less of a necessity, however their larger and more rough-edged fins do allow them more maneuverability when compared to most whales, which allows them to swim in shallower water.
About a decade ago I took a cruise to Alaska with my family, and we went on a whale watching excursion on a ~50 person boat. We came upon a pod of humpbacks who were bubble feeding (cool as fuck), and I was leaning over the railing looking straight down, and all the sudden a humpback swam out from under the boat.
Now let me tell you something, you can see all the videos or diagrams you want, but nothing will prepare you for the size of these fucking things. I was SHOOK. It could have easily capsized our boat if it wanted to. And a humpback whale isn't even close to the size of a blue whale. here's a diagram
It was a dope experience, but I honestly have no desire to be around these huge creatures again, I know they are very docile, but it is extremely unnerving.
I went whale watching with my family in Maine when I was a child. Saw a pod of Humpback whales all around the boat, they were huge! I loved it, my sister got sea sick and threw up.. Even got to see a whale named Arrowhead who was apparently well known.
Afterwards I cried in the gift shop because I wanted a stuffed Humpback toy so I could name it Arrowhead. I still have that darn thing somewhere
Reading this gave me a funny feeling in my stomach. I definitely have thalassophobia (or however it's spelled, I can't google it to check the spelling because it will show me phobia-triggering images. heh).
"Excessive" and "out of proportion" are relative terms. Heck, there's an argument that all fears of spiders are "excessive," because they're like 99.99% harmless, but I most humans have at least some disproportionate fear.
When it comes to anxiety, the only intensity metrics that super matter are "how much distress does it cause?" and "how much does it interfere with your life?" Because those are the metrics that define how important or useful treatment might be.
All anxiety and fear exists on a spectrum, and everyone has different coping mechanisms. I know someone who had a spider phobia so became an amateur spider expert to cope. It's best not to measure your path against someone else's
Username checks out mate, maybe you should. I like to face my fears reasonably so I can move past them if possible. For me, the deep open sea is a definite no go. I do not do boats, or cruises or anything where I am at the mercy of the sea.
But if I didn't allow myself the opportunity to experience an large scale aquarium in a controlled environment, I would have never been able to see a weedy sea dragon up close. I would have never been able to look eye to eye with an octopus, which is one of my favorite animal encounters to date. Octopi are so intelligent and interesting on their own, and simply looking in their eyes you will be met with intelligence and interest from THEM. It's incredible. I now have a tattoo of each of these creatures I have met, that would normally be inaccessible to me due to fear of the ocean.
It's ok to be scared, or to have a phobia. It's ok to be content not facing those phobias and fears. What is not ok is gatekeeping the phobia and holding it to your unique set of symptoms and reactions. We all react differently. You're right. We are not the same.
Depends on if the whale feels threatened or not. If she does, she'll possibly kill you with the flap of her tail, otherwise she might just swim and chill with you.
1 humpback saved this marine biologist's life when the whale spotted a tiger shark nearby (the biologist didn't see the shark til she basically was finished swimming), and proceeded to push the lady by the tip of her mouth away from the predator
Yes! I was just talking about this. I think it's on the inside of Baranof Island somewhere. (Also, it's a hatchery, not a fish farm; there's a big environmental/political difference to a lot of people, so it's an important distinction).
Yeah I'm from Juneau, and this is a hatchery (it's not a farm, a lot of Alaskans hate fish farming). Also it's definitely a humpback.
Whales have been known to return to these locations and sometime get the timing down to just hoover up the entire class of fish in one go. I once got invited to come observe the fish release, and the sea lions probably got a good 5 percent immediately.
It's a well known whale in that area, i saw a documentary about it. They farm salmon there to be set out in nature, but the whale is super smart and always one step ahead. Think it was actually a David Attenborough film as well, can reccomend.
Getting stuck by yourself
Not knowing how deep or shallow the water is
Not knowing how far or close you are to land
Not knowing what's around you because everything that is a possible predator is beneath you and unfortunately it only takes a Test Bite to kill someone if it takes too long
No and no. Attacks rarely happen, but when they do, more often than not, it's mistaken identity/curiosity of what we are, they take a bite, realize we aren't what they thought, they let go, and then it's just purely how large was the shark and how quickly can you get out of water and stop the bleeding.
Again, they rarely happen, but that's typical with a fear. They're not always rational, except to the person fearing it.
For me, it's like fear of heights, but applied to the water if that makes sense ?
Like I don't mind being on a boat, even far from the shore, but I will never EVER swim if I can't see the bottom. Makes me.. angsty. Even though I'm a good swimmer.
Édit: angsty, not antsy
Orcas litterally torture animals for fun. Dolphins actively rape each other and other creatures to death, Zebra males will rape a pregnant zebra so hard it forces them to miscarry. Or if the foal is already born, will drown and kick it to death all while the mother tries to stop it. All of these things happen regularly and are normal to these animals. And that's litterally just 3 species of animals.
Humpbacks often break in to the fishfarms. Sometimes finding themselves trapped in de gear. It Concerns me to think that fish populations are not doing well and more more encounters like these will appear.
My cousin lives on a little island on the BC coast and is a hardcore anti-fish farm activist. She hates them so much. I'm not up to speed on exactly why. She drives her boat out to boat protests and I love her.
Yeppppp. We went vegan years ago. I grew up in salmon territory and I still miss it but I don't want to be contributing to the absolute ecological collapse happening. I've seen salmon runs go dry and it's heartbreaking.
That whales at a job site, there's most likely a foreman nearby and here in BC we have some pretty rigid laws about whale interactions.
You are legally not allowed to be closer than 400m and if you are in a boat that close you need to turn your engines off immediately and wait until it's safe to leave the vicinity.
I know it won't eat me, but the terrified lizard screams about straying too close to the gods cruel touch and that their sea beasts will punish me for my hubris. Horrible mammal serpent that did no wrong | eng | 2309c837-e200-408b-b4e8-5bb9fbf08fc8 | https://redditproxy--jasonthename.repl.co/r/NatureIsFuckingLit/comments/r2gxmy/a_massive_humpback_whale_visiting_a_marina_in/ |
Printable Cutting Practice Activities: How to Use Fold and Cut Sheets
Printable cutting practice activities are a great way to help your kids develop their scissor skills. All you need is a printable sheet and a pair of scissors. Let me show you how to use printable fold and cut activity sheets.
How Do Your Practice Cutting Skills?
It's important for kids to take time to practice their cutting skills. When kids find cutting difficult, sometimes it becomes difficult to motivate them to practice. That's why this printable cutting practice activity can be so helpful. These fold and cut activities make it fun and easy for kids to get the practice they need.
I created the fold and cut activities after receiving so much positive feedback about my printable snowflake patterns. Now, you can have fun cutting practice activities for the whole year. After cutting on the lines, kids LOVE the satisfaction of unfolding the paper to reveal the final design. That's why they are so motivating.
How to use printable fold and cut activity sheets
You can make your own fold and cut activities by tracing half a shape onto a piece of folded paper. The printable fold and cut activities pictured are part of my Seasonal Fold and Cut Activities set. The set includes 38 different designs for all seasons (spring, summer, winter, fall, Halloween, Christmas…etc)
The printable Fold and Cut Sheets are very easy to prep for your kids…which makes them a fun & quick great fine motor craft or activity. Each of the designs print 2 per page. So, all you need to do is:
print
cut the pages in half
fold on the dotted line
Once the activity is prepared, it's time to start cutting!
Have the child hold the paper so that the design is facing up. They can cut on the lines with scissors.
Then, it's time for the BIG REVEAL! When kid finish cutting, show them how to unfold the paper. They will be excited to see the final design!
After opening the paper, they can get creative with crayons, markers, glitter, yarn, torn paper, etc.
Repeat with other designs or printable sheets.
Benefits of printable cutting practice activities for kids
There are many benefits of this printable cutting practice activities for kids. These activities:
Printable Cutting Practice Activities In The Store
Printable cutting practice activities are a great way to help your kids develop their scissor skills. All you need is a printable sheet and a pair of scissors. These activities are also great for improving fine motor skills, developing scissor skills, and building confidence. They are motivating and satisfying, and your kids will do them again and again.
Print out some fold and cut activity sheets today and let your kids have fun while they practice their cutting skills.
Are you ready to get started on some fold and cut activities with your kids? Check out these great printables in the store. | eng | d6fbaf53-2e0f-43df-91fc-5c035ccd82c4 | https://earlylearningideas.com/printable-cutting-practice/ |
I just bought an '06 Chevy Malibu Maxx SS and I don't know squat about vehicles.
I was told the passenger side headlight won't turn on due to the wiring, the bulbs are newish still -- I haven't a clue what wiring, what it's called or the best/easiest/cheapest way to fix it is. Suggestions?
Also, the horn does not work, was told it could POSSIBLY be a fuse, but not sure
Knowing not even squat (lol) you would be better to have someone with even just a rudimentary idea of what should and shouldn't be. The headlight not lighting could be a fuse or relay, which are easy to test and replace. It could be wiring that a rodent has chewed through or damage from a former collision and/or bad repair, or a burned out bulb. If it was not cleaned with alcohol before being installed, there's a good chance that the bulb has died. | eng | 3538aaa3-bde4-4368-9a03-d6f2cb8519f8 | https://www.chevymalibuforum.com/threads/headlight-horn-issues.118138/#post-890803 |
Related Content
Share
Hereditary forms of amyotrophic lateral sclerosis (ALS) may only account for 10 percent of all ALS cases, but the insights that researchers gain from studying this small population may be the key to unlocking new treatments for this devastating disease.
That's the hope of Katharine Nicholson, MD, a physician-investigator at the Neurological Clinical and Research Institute (NRCI) at Massachusetts General Hospital. Dr. Nicholson has launched a project called the DIALS Network (Dominant Inherited ALS Network) — a multicenter study based at Mass General and Washington University that tracks the first-degree relatives of individuals who have a familial form of ALS but do not show symptoms yet.
The goal is to identify the earliest molecular and clinical signs of the disease, which could help in developing new strategies to prevent or lessen its impact.
Searching for Biological and Diagnostic Clues
Dr. Nicholson explains that ALS is considered by many neurologists to be a diagnosis of exclusion, meaning that it can only be confirmed after all other possibilities have been ruled out.
"In general, we don't have good early diagnostic tests," she says. "It takes about a year from symptom onset to diagnosis, which is sometimes a large portion of their disease duration."
Katharine Nicholson, MD
Identified inherited forms of ALS are autosomal dominant, which means that individuals who carry one copy of the disease-causing gene are going to develop the disease at some point in their lifetime, and it will be seen in most generations of the family as well.
The DIALS teams at Mass General and Washington University have recruited approximately 40 first-degree relatives of individuals with ALS to participate in the study over the past year. The study team provides genetic testing and counseling for participants, and collects blood, spinal fluid and other samples every six months to see if they can identify the first indicators of the disease's emergence.
"We're not only looking at the earliest molecular changes, but the earliest clinical changes as well," Dr. Nicholson explains. "We currently diagnose ALS through a clinical assessment of weakness. This can include weakness in an arm or a leg, or difficulty with breathing, swallowing or speaking."
By closely following the study participants, Dr. Nicholson and her team may be able to improve clinical assessments to pick up on more subtle indicators of early disease.
Preparing for New ALS Treatments
That information can be used to inform new treatment strategies for familial ALS, which could include gene therapy designed to stop disease progression early on.
"We want to be poised with a population of gene carriers who are asymptomatic, identify the biomarkers (biological indicators) that will help to predict their conversion into disease state, and use the most sensitive clinical measures to assess early disease progression."
Dr. Nicholson has been overwhelmed by the commitment that families affected by this hereditary form of disease have brought to the research process. "Seeing their proactivity and motivation helps to drive you forward."
Additional motivation comes from the team at the NCRI, including Merit Cudkowicz, MD, MSc, co-director of the NCRI and chief of Neurology at Mass General, as well as her fellow investigators and study support team. "We couldn't have a more inspiring team of people working to cure this disease."
To learn more about how you can support ALS care and research at Mass General, please contact us.
This story first appeared on the website of the Mass General Research Institute.
The Research Institute: Saving Lives Through Science
The Massachusetts General Hospital Research Institute is the largest hospital-based research program in the United States, with a community of more than 8,500 people working across more than 30 institutes, centers and departments.
Our researchers work side-by-side with physicians to pioneer the latest scientific advancements for curing disease and healing patients in Boston, across the United States and around the world. | eng | fd0e442f-bb9b-4c0a-bdd1-0c0cf4fea79d | https://giving.massgeneral.org/stories/rare-als-form-studied/ |
When you have some aches and pains after a long day, nothing is more relaxing than drawing a bath and taking some time to relax. You can enhance the experience so much more by adding dr teal's pure epsom salt therapeutic soak to the mix. This product consists of some ultra-fine crystals that dissolve instantly in the water | eng | 9b5f34d9-571e-4f39-88e7-ffe35f4c7f4f | https://www.perfumepur.com/collections/beauty-fragrances/products/dr-teal-s-pure-epsom-salt-therapeutic-soak-by-dr-teal-s-for-women |
Mermaid Bath Bomb Craft Ideas
Welcome! Crafting your own bath bombs can be a fun and rewarding experience. They make great gifts for friends and family, and are a great way to relax after a long day. Today, we're going to explore some creative ideas for making mermaid-themed bath bombs. From shimmering glitter to charming decorations, you'll be sure to find the perfect bomb to make your bathtime even more magical. So, let's get started!
How to Make Sparkling Mermaid Egg Bath Bombs With Hidden Color
FAQs Related To Mermaid Bath Bomb Craft Ideas
1. What can you not put in bath bombs?
Bath bombs are a popular way to relax and enjoy a luxurious bath, but there are some ingredients that should never be included in them. First and foremost, bath bombs should never contain any type of oil. Oils can cause the bath to become dangerously slippery, and can also clog drains. Additionally, bath bombs should never contain any type of abrasive material, such as baking soda or cornstarch, as these can cause irritation to the skin. Finally, bath bombs should never contain any type of dye or artificial coloring, as these can stain both the bathtub and the skin. When crafting bath bombs, it's important to only use ingredients that will be beneficial and relaxing to the user.
2. What is in a mermaid bath bomb?
Mermaid bath bombs are a popular item for those who love to take luxurious baths. They are usually made of baking soda, citric acid, epsom salt, essential oils, and colorant. The baking soda and citric acid create the fizz when they enter the water, while the epsom salt helps to soothe the skin. Essential oils provide a pleasant scent and the colorant is used to create the mermaid look. Some bath bombs may also contain ingredients like cocoa butter, coconut oil, and shea butter to provide extra moisturizing benefits.
3. What are the best bath bomb ingredients?
The best bath bomb ingredients are those that are natural, non-toxic, and provide a soothing and relaxing experience. Most commonly, baking soda, citric acid, Epsom salt, and essential oils are the main ingredients used in bath bombs. Baking soda helps provide the fizzing action, while citric acid gives it its tangy smell. Epsom salt helps to soothe tired muscles and skin, while essential oils provide a calming scent. Other ingredients like moisturizing oils, herbs, and clays can also be used to enhance the experience. Ultimately, the best bath bomb ingredients are those that can provide a safe, relaxing experience without being too harsh on the skin.
4. What are some bath bomb names?
Bath bombs are a great way to relax and pamper yourself, and they make a great gift too! When it comes to choosing the perfect bath bomb, there are plenty of creative names out there to choose from. Some popular bath bomb names include 'Relaxing Rose', 'Soothing Lavender', 'Calming Chamomile', 'Citrus Splash', 'Dreamy Jasmine', 'Coconut Crush', 'Cocoa Butter Bliss', 'Vanilla Delight', 'Tropical Oasis', and 'Minty Magic'. With so many options, you're sure to find the perfect bath bomb to suit your needs and pamper yourself!
5. What products are in bath bombs?
Bath bombs are a fun and relaxing way to enjoy a warm bath. They are typically made up of a mixture of baking soda, citric acid, Epsom salts, essential oils, and sometimes colorants and glitter. The baking soda and citric acid are what create the fizzing reaction when the bomb is placed in water. The Epsom salts provide essential minerals such as magnesium and sulfate, which can be absorbed through the skin. The essential oils are added for fragrance and to provide aromatherapy benefits. Colorants and glitter are sometimes added for aesthetic purposes. Bath bombs are a great way to add a little extra indulgence to your bath.
6. What is the purpose of bath bombs?
Bath bombs are a type of product that is used to enhance bath time. They are typically made up of a combination of baking soda, citric acid, Epsom salts, and essential oils, as well as food-grade dyes, and other natural ingredients. When placed in a bath, the bath bomb fizzes and releases its ingredients, creating a luxurious and fragrant bath experience. The purpose of bath bombs is to provide a relaxing and luxurious environment for your bath time, as well as to nourish and moisturize the skin. The essential oils, Epsom salts, and other natural ingredients provide therapeutic and healing benefits to the skin. Additionally, the bath bomb's delightful scents and vibrant colors make for a truly invigorating experience.
Creating your own mermaid-themed bath bombs can be a great way to add a magical touch to your bath time routine! With just a few simple ingredients and a little bit of creativity, you can make your own unique and luxurious mermaid-themed bath bombs. | eng | 7e7f745f-df2e-46e3-b21f-1a52b851dc08 | https://www.kidsartncraft.com/mermaid-bath-bomb-craft-ideas/ |
A photography blog on how to become a photographer: including reviews of the latest and best online photography courses, photography schools, and photography workshops, camera deals and reviews, and photography accessories.
Why the Mirrorless Camera Revolution Matters
I remember being one of the first to hear about the Micro 4/3 mirrorless camera technology in 2008 when another editor returned from a confidential press trip to Japan with Panasonic. He informed us about the mirrorless technology, and was excited about what it meant for photographers. Although it wasn't the first mirrorless camera in existence (Epson R-D1 and the Leica M8), the secrecy behind the Panasonic G1 was being touted as the technology of the future by photography journalists once it was revealed by Olympus and Panasonic.
At the time it seemed unimportant, especially when affordable Full-Frame DSLRs were gaining traction like the Nikon D700. Point-and-shoots were selling hand over fist, DSLRs had finally reached the sub-$1,000 mark, and there wasn't much space for cameras without mirrors.
The mirrorless technology was supposed to be cheaper, have the same image quality as a DSLR, and most importantly—smaller. When I finally got my hands on a G1 it felt cheap and plasticky, and it had an EVF (electric viewfinder) that was sluggish, low-res, and awkward. Mirrorless seemed like a novelty or fad to me, but obviously I was wrong.
Mirrorless Camera Revolution
One of the other major issues that plagued early iterations of mirrorless cameras was the slowness of contrast-based autofocus, unlike the speed of Phase Detection AF in DSLRs. This made it harder to find focus in low-light and contrasty scenes.
So you might be wondering why mirrorless cameras haven taken off in the last few years, and what set up the conditions for them to make such a dent in the marketplace?
Collapse of Point-and-Shoot Market
Smartphones have made photography and digital imaging more accessible than ever before, and the rise of our digital and social network lifestyles like Facebook, Instagram, Pinterest, Snapchat, Tumblr and photo-sharing sites made using a smartphone, well, smarter.
The accessibility of 4G and internet-enabled devices made sharing images as easy as taking a photo and uploading it to the Internet. Convenience is always king, and nothing is easier than having a high-resolution image captured on your phone that connects to the internet with a push of the 'share' button.
This made camera companies play catch up to Olympus and Panasonic to release their own mirrorless cameras. When the point-and-shoot profits dwindled in the market, there was an increase in the interchangeable-lens camera market from 12.9 million units shipped in 2010 to 17.1 million in 2013. The revenue for both DSLRs and mirrorless cameras combined went from $4.6 billion to $6.2 billion, according to CIPA.
Improvements Across the Board
The camera companies adaptation to the market demand help paved the way to improve mirrorless technology and design; allowing for more affordable, smaller, classic-looking cameras reminiscent of the film days.
The major difference between a DSLR and mirrorless is the presence of a mirror in the DSLR. So what's the big difference anyhow? Well, a mirror gives you an exact optical view of what your camera is seeing, while the mirrorless uses an EVF or a live-view LCD screen on the back of the camera to frame.
But as stands today, the EVF viewfinders on higher-end models like the Fujifilm X-T1, Sony Alpha A7II, and the Olympus OM-D E-M5 Mark II, are far superior then first versions. I shoot with the Fujifilm X-E2, and the EVF is leaps and bounds better than the original Panasonic G1. The gap on mirrorless EVF's and optical viewfinders on DSLRs is starting to narrow, and it only seems to be getting better with every new camera release.
Autofocus Adjustments
Another contention a few years ago was the poor performance of Autofocus in mirrorless cameras. Most had a contrast-based AF, while DSLRs had the laser fast phase-detection AF which is better for predicting the movement of a subject and focusing in low light. That is no longer the case with some mirrorless cameras, because companies like Fujifilm, Nikon, Olympus, and Samsung all have a phase-detection AF mirrorless cameras in their offerings today.
That distinction between DSLR and the mirrorless camera seems to be shoring up, but it doesn't necessarily mean that they are as good or better than a DSLR. But as technology always does, they will one day equal each other until the law of diminishing returns puts a kibosh on improvements that can be made. We can see that the arguments that mirrorless cameras aren't as good as a DSLR is getting smaller, but DSLRs are still the better camera for the professional today.
Mirrorless Retro Designs
I think that one thing that helped mirrorless cameras succeed from 2011 on was the aesthetic quality of their design and build. When you see them you are instantly struck that people were using a film camera with all the dials, classic body style, rangefinder looks, and nice metallic/aluminum finishes, but instead it was just a smaller digital camera with a bigger image sensor.
We now have modern day rangefinders and sleek retro cameras that take us back to the days of film, but give us the convenience of lightweight, compact, and digital images that are a joy to capture; and that render beautiful images. They sort of fall into the Leica threshold, as in they look like rangefinders and capture images with great image quality.
Will Mirrorless Become a Pros Camera?
Yes and No…maybe. We already see a lot of professionals latching on to the Sony Alpha A7 cameras because of their low-light performance and full-frame sensor. Another reason early on that mirrorless got a bad wrap is because their sensors were too small with a Micro 4/3 or APS-C size, and the argument was that they couldn't achieve the same depth of field as a DSLR. Well, that is an antiquated argument, especially since there are now full-frame sensors on mirrorless cameras.
However, there is still no equating a high-end professional DSLR to the high-end mirrorless cameras today, but that may change with advancements in technology. Professionals still need the pro performance of speed and versatility found in elite DSLRs, and they also need the low-light performance and high-end glass. Although we see this gap shortening, I am not sure if DSLRs will always be around, or if mirrorless will carry on with its trend.
So where a mirrorless camera for professionals will increase is in the photojournalist or street photographer who needs a lightweight, discrete camera, or the nature photographer that captures wide-angle landscapes. If mirrorless cameras are built more weatherproof and rugged, I predict that many different photographers will start considering mirrorless cameras in these subsets of photography.
Mirrorless, Mirrorless On the Wall
What does this mean for the future of photography? Everything evolves, and camera companies ahead of the curve and daring enough to put out new technologies and great cameras will survive.
The point-and-shoot market collapse was an evolution of Apple iPhones and Android smartphones releasing cameras that capture pictures good enough pictures for the consumer market. It's hard to beat the convenience of having your camera and phone all in one place, and that's why we saw the mass decline in the point-and-shoot market.
Point-and-shoot cameras were the main source of profit for camera companies, and that trend is over. Just like when film was overtaken by digital, the market dipped but gave way to the consumer point-and-shoot camera. Camera companies will evolve and continue to make what the market demands, or if they don't, they may have a slow death like Kodak.
To read the perspective of a professional photographer and filmmaker on the death of the point-and-shoot, check out what Vincent Laforet's blog post here. | eng | affaa713-5043-430a-b79b-32e753cc55fe | https://precise-moment.com/2015/02/25/why-the-mirrorless-camera-revolution-matters/ |
The nationality and statelessness of nomads under international law: With a comprehensive examination of the nationality and statelessness of former Bedouin in Kuwait, Tuareg in Mali and Sama Dilaut (Bajau Laut) in Malaysia
title = "The nationality and statelessness of nomads under international law: With a comprehensive examination of the nationality and statelessness of former Bedouin in Kuwait, Tuareg in Mali and Sama Dilaut (Bajau Laut) in Malaysia", | eng | 5f545e7b-c88b-4aec-908f-90a69b477f0d | https://research.tilburguniversity.edu/en/publications/the-nationality-and-statelessness-of-nomads-under-international-l |
Why is zinc oxide used in rubber?
If you are looking for high-quality products, please feel free to contact us and send an inquiry, email: brad@ihpa.net
Zinc oxide is an oxidized form of zinc. The oxide of zinc is insoluble and can dissolve in water. It can also be easily dissolved by acids or strong bases. Zincoxide is a widely-used chemical additive. This is often used to make plastics.
A large band gap, exciton binding energy and high luminescence are all characteristics of zinc oxide at roomtemperature. It's found in liquid crystal displays, thin-film transistors and light emission diodes.
A type of nanomaterial called zinc oxide of microparticles was also introduced.
Where is Zinc oxide employed?
1. It is used principally as a filler and coloring agent for white rubber.
2. In the chemical fertilizer business, fine desulfurization occurs.
3. They are used mostly as white colors, rubber vulcanizing aktiv agents, organic cats, desulfurizing substances, for xerographic prints, pharmaceuticals, etc.
4. For the desulphurization NH3, oil, natural gas and chemical material gas.
5. As a matrix for reference reagents as well as analytic and analytical reagents. It can also be used to identify fluorescent agents or photosensitive data.
6. For electrostatic wetcopying, dry-transfer printing, laser facsimile communications, electrostatic recording electronic computers, and file making.
7. This product is suitable for use in plastic occupations such as sunscreen cosmetics products, specialty ceramic products, functional coatings, textile cleaning and processing.
8. It is extensively used in deep desulfurization or purification processes of industrial materials, such as NH3, methanol and hydrogen production.
9. As an astringent for plastering or making ointment.
10. As a white pigment, it is widely used in papermaking, matching, pharmaceutical, and printing. This is used in the production of rubber latex, natural rubber rubber's vulcanizing activ agent, reinforcing agents, and coloring agents. It can be used in producing pigments like zinc chromium-yellow, zinc acetate and zinc carbonate. This is in addition to its use in electronic laser information, phosphors feed additives catalysts magnetic materials and feed additives.
11. Inorganic white pigment. Although the color is stronger than Lithopone, titanium dioxide has less power. Used extensively in ABS resin, C8H8, n, epoxy, phenolic, PVC, and paint and ink colors.
Why Zinc oxide is important in Coatings
1. As a White Pigment. Zinc oxide can also be known as zinc white. Due to its lower color power than titanium dioxide's, zinc oxide can be used for more purposes. For example, when titanium dioxide varnish is being adjusted in color. It is convenient for the color paste to be weighed, it can easily and precisely adjusted, and its color sensitivity has been reduced.
2. Used to increase the property of epoxy resins containing carboxyl functional systs.
Zinc oxide, an alkaline substance that reacts with carboxyl functional chains in epoxy resin, can form zinc soap. The zinc soap will increase the coating's flexibility and strength. Also, the coating is less toxic than titanium dioxide.
3. Promote the reaction between epoxy and carboxyl groups.
Epoxy resin / polyamide, polyester / TGIC and other electrostatic products have an enhanced carboxyl reaction, which results in a dry-solid catalytic process and a faster solidification.
4. A conductive functional filler.
For some industrial applications it is essential to stop static electricity from being generated. Zink oxide is often used to fill in the architectural coatings with conductive materials because of its electrical conductivity.
5. It also protects against ultraviolet light.
Zinc is a cheap and affordable white light stabiliser. It offers good protection against ultraviolet radiation in the range of 240-380nm. It's often used to protect outdoor powders and electrostatic powders.
6. They have the ability to kill bacteria and mildew.
Zinc dioxide has strong germicidal capabilities and can also be used as a sterilizer in water-soluble coatings. You can use it to make electrostatic powder for mildew-removal.
Why do you use zinc oxide in rubber?
Zink oxide, although it makes up only a fraction of the rubber-making process's formula, plays a crucial role. Zinc oxide is combined with stearic, creating zinc soap. The zinc soap improves the zinc oxide solubility in rubber compound. This complex, which reacts with sulfur to create a strong agent for vulcanization, is especially active. A key component of rubber production is zinc oxide, which plays a significant role in encouraging vulcanization.
The cross-linking of rubber can be increased by zinc oxide during rubber production. Zinc oxide reacts well with stearic and creates Zn2++. You can also increase the aging, thermal and mechanical properties of vulcanizates.
Zincoxide Price
Many factors affect the price of a product, such as market demand, trends in the industry, economic activity, sentiment within the market, unexpected events, and the availability and cost of the item.
To get a quote on the latest ZnO pricing send us an email. (brad@ihpa.net)
Zinc OxideSource
Luoyang Tongrun Nano Technology Co. Ltd. Our products can be shipped anywhere in the world.
Please contact us to receive a quote if high-quality ZnO o powder. (brad@ihpa.net)
Inquiry us | eng | 8d489e81-02a9-4656-ab18-d972b9d86ca8 | https://www.lanzatesolo.com.ar/why-is-zinc-oxide-used-in-rubber.html |
Hire a Nurse Attorney When You Receive a Complaint From November 30, 2016, through December 14, 2016, the said LVN submitted falsified nursing documentation for a patient, which indicated she made multiple visits to the patient's home. Records indicate the visits were not conducted due to the patient being hospitalized during this time period. The LVN's conduct created an inaccurate medical record and was likely to defraud the facility of money paid to the LVN for hours not actually worked. Additionally, her conduct was likely to injure the patient in that subsequent caregivers would rely on his documentation for further care.
During the hearing, the LVN states she made a nurse visit to the patient's home on December 1, 2016, and spoke with the care provider, who gave him information on the patient. She states the patient was said to be asleep and therefore no direct patient contact was made. The LVN further states she conducted a follow-up phone visit and attempted another home visit, but she was unable to establish contact with the patient, with the care provider. She states she documented the second attempted visit in a nurse visit form instead of a supplemental interdisciplinary form, and she called in the visit and entered the code "017 Nurse Visit" rather than "011 attempted visit at patient home."
However | eng | 6afb407f-599a-4462-8ac8-83ab6b41a5a6 | https://www.fortworthnurseattorney.com/hire-nurse-attorney-receive-complaint-work-3/ |
We Are Part Of The Solution!
Share
"The passenger pigeon, the golden toad, the Caspian tiger: they are all gone, and other species hang by a thread. Our actions are not merely driving other species to extinction: we threaten our own survival, too, by destabilizing ecosystems and destroying biodiversity."- Owen Jones
Biodiversity refers to the variation of life at all levels which bounds the evolutionary, ecological, and cultural processes that support life on earth. Biological diversity resources are the pillars that fabricate human advancements.
The need for the preservation of biodiversity is typically overlooked by individuals. The fundamental reason for this is that humans consider only themselves and not about the environmental factors and climate. Unfortunately, by now, several species have disappeared and if we did not take any mandatory step to save them then, more of them will become extinct in near future. The fruitfulness of biodiversity is indispensable for the correct functioning of the world's current circumstance.
Throughout the most recent century, people have dominated the planet, causing biological system change and monstrous loss of biodiversity across the planet. While the Earth has consistently experienced changes and eradications, today, they are happening at an uncommon rate.
Habitat loss and fragmentation, invasive species, pollution, unsustainable use of resources and global climate change are major threats to biological diversity now. The growing human population and overconsumption of natural resources are often complex interrelated factors that contribute to this major issue.
Loss of biodiversity compromises all, including our health. It has been proven that biodiversity loss could expand zoonoses; diseases transmitted from animals to humans, while, on the other hand, if we keep biodiversity unblemished, it offers an excellent atmosphere to fight against pandemics like what caused by coronaviruses. This is much needed nowadays because the whole world is under a risk and fighting for saving human lives.
Fortunately, is that it is within our power to help guarantee the endurance of species and the wellbeing and integrity of biological frameworks. By understanding threats to biodiversity, and how they work out in our setting, we can be ready to oversee conservation challenges.
We have to fix our relationship to the natural world, simply because, despite all our innovative technological advances we are completely dependent on healthy and lively ecosystems for our water, food, medicines, clothes, fuel, energy and many more.
Today, The International Day for Biological Diversity is a day to speak out and make amendments to build sustainable development strategies which can address the current issue of losing the value of biodiversity. But, this should be a fruitful discussion that comes up with solutions that provide advantages to nature and should not confine for today. This is a global issue and we as humans should embrace the power of nature and help to rebuild the vanished aspects of nature.
This year 2021 the theme is "We're part of the solution". Indeed humans have the answer for this huge disaster. Building momentum to increase awareness on the significance of biodiversity is much important for this. We might not be able to eradicate all negative anthropogenic impacts on biodiversity, but with knowledge, we can work together to change the direction and shape of effects of human activities, on the remaining life on Earth.
It is a must to reunite for conserving biodiversity because the human community is just a strand of the web of nature.
So on this valuable day,
"We must convince and empower people to adopt the conservation and sustainable use of biodiversity as their guiding principle" – Klaus Topfer | eng | 128b8341-86fc-468d-89da-bf81440f6250 | https://fos.cmb.ac.lk/blog/we-are-part-of-the-solution/ |
CJ203: Reactions to Juvenile Cases
he following outline serves as a valuable preparation tool and reference for this week's assignment. Complete it prior to starting your W7 Assignment.
Outline
I. Name and discuss the the various causes of crime using criminal justice theories and practices (Note: The textbook indicates what the various causes of crime are.
II. Compare and contrast historical and contemporary police functions and issues (Note: How has the function of police changed? What issues caused/contributed to the changes?)
III. Describe the nature and function of corrections, its services, and practices (Note-Answer these questions: What is the purpose of corrections? What services does corrections provide? What are practices in corrections?
IV. Apply fundamental concepts of the administration of justice processes to a multicultural populations responses to crime institutions (Note-Answer these questions: What is the process in which justice is administered? How is justice administered amongst different cultures? We talked about disparities so how is justice administered to crime institutions in various cultures/populations? Is it equal?)
V. Select and discuss at least two (2) specific crimes (homicide, rape, burglary, etc.) using criminal justice theories to discuss possible causes of these crimes, as well as the past and present police responses to these crimes. a. Have the courts and corrections been adequate in the prosecution and punishment of persons accused and/or convicted of these crimes? b. Do you feel that legislation needs to be changed to influence either the capture or punishment of suspects accused of these crimes? c. Do you feel the victims or their families would support the legislation changes you are proposing? Why or why not? d. Do you feel that this legislature targets one group more than others (i.e. gang populations, lower socio-economic status populations, etc.)? Note: Answer this in detail
Note: Be sure to apply fundamental concepts of the administration of justice processes in your reflection.
VI. Summarize what you have learned in the course and end with your own personal opinion about what you have learned.
Reactions to Juvenile Cases
In this assignment, you will select and discuss two juvenile cases, the crimes committed, and the various criminal justice theories that may help explain the crime (hint- you will have to draw on knowledge gained from your CJ102 course, Intro to Criminology). Have the juvenile courts and juvenile corrections been adequate in the prosecution and punishment of minor persons accused and/or convicted of these crimes? Do you feel that legislation needs to be changed to influence either the punishment or rehabilitation of suspects accused of these crimes? Do you feel the victims or their families would support the legislation changes you are proposing? Why or why not? Do you feel that this legislature targets one group more than others (i.e. gang populations, lower socio-economic status populations, etc.)? Be sure to apply fundamental concepts of the administration of the justice process in your reflection. Your paper must be formatted in the following manner:
The paper must be approximately 1000 words in length (roughly 3-4 pages). Only the body of the paper counts towards the length. No information that belongs on a title page or reference page will count towards the length requirement.
Please follow APA guidelines; provide a title and reference page, use Times New Roman 12 point font, and double-space your lines. You need to have 5 references, two of which can be your texts from this class. Their references can be found in the course syllabus | eng | 5d5a2c5f-57ff-4406-882b-ba1715d30638 | https://theanswersbay.com/cj203-reactions-to-juvenile-cases/ |
Unleashing the Power of American Bully XL: Understanding Size and Weight
How to Determine the Size and Weight of an American Bully XL: A Step-by-Step Guide
The American Bully XL is a robust and muscular dog breed that has gained immense popularity in recent years. Their imposing appearance coupled with their friendly demeanor has made them an ideal pet for those looking to add a faithful companion to their family.
As the name suggests, American Bully XLs are large breeds that require special care and attention when it comes to determining their size and weight. Therefore, it is crucial to understand the steps involved in measuring the size and weight of this gentle giant.
Here is a step-by-step guide on how to determine the size and weight of an American Bully XL:
Step 1: Measuring Height
The height of your American Bully XL should be measured from the ground up to their withers (the highest point on their shoulder blades). Stand your dog upright against a wall or flat surface while ensuring that they are standing squarely on all fours. Carefully take measurements using a tape measure or ruler.
Step 2: Determining Weight
To effectively determine your bully's weight, use a digital scale that can weigh up to at least 100 pounds. Place your dog on top of the scale and ensure they stand firmly with all paws on the scale platform. Record measurement readings from two separate occasions (morning and evening) as dogs tend to weigh more during different times of the day due to feeding habits.
Step 3: Measuring Chest Size
One way you could determine chest size appropriately would be using flexible tape measurements like tailor's tape tools. Measure approximately one inch behind your bully's front legs across its widest area around its chest area where it appears most muscled out - not too tight but firm enough such that you get accurate readings.
Step 4: Analyzing Data
Once you've measured all data correctly, refer individual recorded values into appropriate Index Ratio Chart Table for analysis purposes. This helpful chart allows dog owners consistency in tracking health status as well as establishing ideal ranges of measurements and weight for the breed.
Ultimately, determining the size and weight of an American Bully XL is important in understanding their nutritional needs, proper exercise requirements, and overall health. These steps should help you get accurate readings so that you can provide your beloved pet with a healthy lifestyle suitable to their specific needs.
In Conclusion
Measuring your American Bully XL's height, weight, and chest size using these simple steps will provide a wealth of information on its ideal range as well as maintaining optimal health. With correct feeding habits and exercise routine established from the assessment carried out using these steps; you can be assured a healthy, happy furry companion.
Common FAQs about American Bully XL Size and Weight Answered
American Bully XL size and weight are some of the most commonly asked questions about this dog breed. While the American Bully is a relatively new breed, it has skyrocketed in popularity amongst dog lovers due to its unique characteristics and temperament. In this blog post, we will tackle some of the frequently asked questions about their size and weight.
What is an American Bully XL?
An American Bully XL is a large-sized version of the American Bully breed. It weighs between 80-150lbs for males and 70-120lbs for females. These dogs have huge muscles, broad chests, wide shoulders, and thick necks making them look extremely powerful.
Why are they larger than other bully breeds?
The American Bully breeds were created through crossbreeding selective breeding processes between the American Pitbull Terrier (APBT) and the Staffordshire Bull Terrier (SBT). Later on other bulldog types also came into mix such as French Bulldog or Boxer which grew its size exponentially resulting in a larger bulldog-type dog known today as an American Bully XL.
Are there any health concerns associated with their size?
Yes. Heavier weight can put more strain on joints leading to conditions like hip dysplasia or arthritis that can shorten your Bullies lifespan. Proper exercise programs during growth stages are important for managing healthy musculoskeletal development without not overdoing it which could cause structural problems because of their muscular build up.
How do I know if my dog is overweight?
The best way to determine whether your dog's muscle mass compares against his fat mass rather relying solely on scale weight as every individual may possess varying fat percentage among body composition demographics so knowing their BMI helps make sure you're not overfeeding or underfeeding without consulting veterinarian help whenever unsure what steps to take next.
How can I keep my American Bully XL at a healthy weight?
American bully owners should monitor food intake by providing the recommended amount of food based on their dog's individual needs. Exercising daily with appropriate activities and frequency to match the physical traits of this breed can also maintain healthy development, so he doesn't end up overweight or underweight when feeding is concerned.
In conclusion, American Bully XL size and weight questions are common amongst owners and those interested in purchasing one. While their build may give the impression of being bulky, maintenance involves cautioning against weight gain while promoting proper exercise habits that benefit their musculoskeletal health at all stages of life. Be sure to ask for assistance from professionals should you have any concerns that arise about dieting adjustments as puppies grow up into adulthood – it's better safe than sorry!
Top 5 Facts You Should Know About American Bully XLs and Their Size/Weight
American Bully XLs are a breed of dog that have become increasingly popular over recent years, and it's easy to see why. These dogs are big, powerful, and incredibly loyal pets that can make fantastic companions for those who understand their unique needs.
But what exactly should you know about American Bully XLs and their size/weight? Here are the top 5 facts:
1. American Bully XLs are one of the largest breeds of dog
While there may be some debate over which breed is actually the largest, it's safe to say that American Bully XLs are up there. On average, these dogs can grow to be anywhere from 19-23 inches tall at the shoulder and weigh between 70-120 pounds.
2. Their weight can vary depending on genetics
Despite their name, American Bully XLs don't always grow to be extra-large in size. Like with any breed of dog, their weight can vary greatly depending on factors like genetics and individual health.
3. They require a lot of space
Because of their large size, American Bully XLs require plenty of space both indoors and outdoors. This means that they may not be suitable for apartment living unless you're willing to provide them with ample exercise and playtime outside.
4. Proper exercise is crucial for maintaining a healthy weight
As with any breed of dog, maintaining a healthy weight is crucial for an American Bully XL's overall health and wellbeing. Regular exercise is key — these dogs need at least an hour or two each day to run around and burn off excess energy.
5. Nutrition plays a big role in their health
In addition to exercise, proper nutrition is also critical for keeping American Bully XLs healthy and happy. Feeding them a high-quality diet rich in protein will help keep them energized throughout the day while supporting muscle growth and maintenance.
Overall, if you're considering bringing an American Bully XL into your home, it's important to understand their unique needs as a large breed of dog. With proper care and attention, these dogs can make beautiful, loyal companions for years to come.
The Importance of Proper Nutrition for Maintaining Healthy Size and Weight in an American Bully XL
As American Bully XL owners, we know how important it is to maintain the health and well-being of our pets. These dogs are known for their impressive size and muscular build, which can be achieved through a combination of genetics and proper nutrition. However, as with all breeds, the nutritional needs of an American Bully XL are unique and require careful attention to ensure they maintain a healthy size and weight.
To understand why proper nutrition is so important for maintaining a healthy weight in an American Bully XL, we need to first look at their anatomy. These dogs have a broad chest, powerful shoulders, a muscular neck, and well-defined hindquarters – all of which require significant amounts of energy to develop and maintain. Additionally, they have active metabolisms that enable them to burn calories more quickly than other breeds. This means that if they do not receive proper nutrition, they may struggle to gain or maintain weight.
So what does "proper nutrition" mean for American Bully XLs? It starts with understanding their specific dietary requirements. As carnivorous animals, these dogs require high-quality protein sources such as beef or chicken, as well as complex carbohydrates from vegetables like sweet potatoes or peas. They also need plenty of essential fats from sources like fish oil or flaxseed oil to support heart health and provide energy.
When designing a diet for your American Bully XL, it's important to consider factors like age, activity level, and overall health status. Puppies will need more frequent meals than adults because they're growing rapidly; similarly, an adult dog who lives an active lifestyle will require more calories than one who spends most days lounging around the house.
Another essential component of maintaining the right weight in an American Bully XL is portion control - feeding too much (or too little) can lead to obesity or malnutrition respectively. So how do you determine how much food your pup needs? It's best practice to consult with a veterinarian or a professional dog nutritionist to create the right diet plan. They could take into account various factors like age, exercise routine, and other medical conditions to ensure your dog gets the right balance of protein, fat, and carbohydrates.
In addition to feeding your American Bully XL the right amount of nutritious food, it's critical to limit high-calorie treats and table scraps. You'd be amazed at how quickly those little bits of cheese or bacon can add up in terms of calories! We encourage owners to switch these indulgences with low calorie options that stimulate their taste buds without adding unwanted pounds.
Finally, providing enough water is equally important when keeping an American Bully XL healthy. These dogs have a larger surface area relative to their weight than many other breeds; this means they require more water per pound of body weight than smaller dogs. Always make sure there's clean drinking water within reach for your pup.
We all want our American Bully XLs to live long lives full of energy and vitality. By providing them with proper nutrition that provides the building blocks for strong muscles and joints, we can help ensure they stay in peak physical condition throughout their lives. In conclusion, feed them well, educate yourself about their specific nutritional needs before starting any diet regimen; always consult a veterinarian if you have questions or concerns – after all, they deserve nothing less!
Key Differences Between the Male and Female American Bully XL When It Comes to Size and Weight
There's no doubt that the American Bully XL is a formidable breed of dog. With their muscular build and confident demeanor, they make for excellent companions and protectors. However, when it comes to males and females of this breed, there are some key differences in size and weight that are worth exploring.
Firstly, let's talk about height. Male American Bully XLs typically reach heights between 19-23 inches at the withers (the highest point on their shoulders). Females tend to be slightly smaller, measuring in at 18-22 inches tall. While this may not seem like a huge difference, it can be fairly noticeable when comparing two dogs side-by-side.
When it comes to weight, male American Bully XLs are generally larger than females. It's not uncommon for a fully grown male to weigh anywhere between 70-120 pounds. Females usually weigh between 60-100 pounds when fully matured. Again, these numbers may not seem vastly different, but they can add up when considering the overall size of these dogs.
Another notable distinction between male and female American Bully XLs is their body shape. Males tend to have larger heads and broader chests than females, giving them a more imposing appearance overall. Females tend to have more slender frames and narrower heads, which can lend itself to a more athletic appearance.
One factor that many people consider when choosing between male and female dogs is temperament. While each individual dog has its own unique personality regardless of sex or breed, there are some general characteristics associated with males versus females. Male American Bully XLs are often described as being more dominant and territorial than females. This can manifest in behaviors such as marking territory or displaying aggression towards other animals or people they perceive as threats. Females are typically described as being more nurturing and affectionate towards their owners.
Overall, when deciding whether to get a male or female American Bully XL, it's important to consider your own lifestyle and preferences. Are you looking for a larger dog that makes a statement? Or do you want a more athletic and affectionate companion that can easily fit into your home? No matter what you choose, these dogs are sure to bring love, protection, and joy into your life.
Tips for Keeping Your American Bully XL Safe and Comfortable as They Grow in Size and Weight
As an American Bully owner, it is crucial to ensure that your dog grows up healthy and happy. The American Bully XL breed can grow up to 23 inches in height and weigh as much as 120 pounds. So, if you have decided to bring this hulking giant into your home, it is advisable that you know how to keep them safe and comfortable as they undergo rapid growth spurts.
Here are some tips that can help you take care of your American Bully XL puppy:
1. Give Them the Right Nutrition
The first step towards ensuring a healthy growth for your bully requires providing them with high-quality nutrition. Poor nutritional intake can cause nutritional deficiencies which, in turn, may lead to stunted growth or malformation of bones and joints.
Ensure that their daily diet consists of protein-rich foods such as chicken, beef or fish in appropriate proportions. Additionally, make sure that they have adequate water supply throughout the day.
Avoid feeding them human scraps or food containing onions, garlic or other toxic ingredients because Bully XLs are prone to specific dietary allergies.
2. Regular Exercise Helps Build Muscle Mass
Like any other breed of dog; their overall health depends on a healthy lifestyle infused with regular exercise workout routines. Engage them at least twice a day in high-intensity activities like running outdoors or indoor games such as chewing-proof toys(kong) mentioned on various pet reviews online sources.
Exercising helps build muscle mass that will support their growing body weight by preventing unusual bone deformity issues caused by excessive wear and tear damages through strain from supporting extra weight loads.
3. Schedule Routine Vet Checkups
Routine veterinary visits ensure preventive health care minimizing severe conditions concerning discomfort due to infections and internal organ disorders from developing at an early stage under professional medical attention.
Their increasing size during adolescence stages necessitates regular checkups throughout each developmental phase primarily for genetic disorders such as hip dysplasia et al., which are common in large dog breeds.
4. Ensure Comfortable Sleeping and Rest Space
Bully XLs enjoy comfortable rest periods to rejuvenate their bodies health-wise before engaging more physical activities. As much as a crate or dog bed is necessary, opt for heavy-duty materials such as ultradurable chew-proof kisiwaa blankets that enable them to lay down comfortably without wear and tear in minimal time.
5. Identify Safety Vulnerabilities
Ensure the puppy confinement area provides adequate protection from environmental stressors keeping them away from harmful chemicals, electrical hazards, and sharp objects et al., which can lead to injuries resulting in lifelong damages that reduce the quality of life standard for your beloved pet.
In conclusion, taking care of an American Bully XL requires responsible ownership through proper nutrition, regular exercise workouts and identifying potential safety vulnerabilities essential for your valuable pet's stable healthy growth spurts until adulthood. With practical guidelines, you will provide a happy and fulfilling experience with your furry companion for years to come. | eng | 4da77c24-6225-43d7-8362-0155c16d3815 | https://kingbuiltbullies.com/unleashing-the-power-of-american-bully-xl-understanding-size-and-weight/ |
Nurture Tute –
Online Tuitions
Nurture Tute is a platform for students to take academic courses online. Nurture Tute covers all subjects and all primary languages and help the students gain more scores in their academics. The courses are rigorous and guarantee success in academics and provide parents peace of mind knowing they're being safely progressed through their studies.
Help your child grow in all aspects of their education
Nurture your child's physical and social skills
Feel confident that you're providing the best possible education for your child
Ease of communication - From parents through dedicated coordinators to help them.
Testimonial
Charles Vimal
My Son C.Eric Noel has been attending Nurture Academy for about last 3 months. He has
drastically improved his game. He is attending weekly 2 classes which helps him a lot. Recently he attended District School level chess competition in which he won 5 games out
of 7, felt so happy to share with u all and the teacher guides him a lot with patience .Thank u
Teacher and the Nurture Academy.
- Charles Vimal (Chess)
Mary Joseph
I have really enjoyed this course, it has been really inspiring and I would love to take the full
course,very good instructor really love the way she teaches Thank you
- Mary Joseph
Manns Mansoor
I had been looking for a social skills, art and craft coaching for my kids for a while and found
Nurture Art Academy. They are not just like any other academy they are team of parents with
deep understanding on a holistic development of child. Their programs are designed to
nurture child's skills and abilities from an early age. My kids attended a couple of their
programs and I have enrolled them to art and crafts, painting, social skills (highly
recommended) and kids yoga. I highly recommend Nurture academy to all parents.
- Manns Mansoor (Social Skills, Art & crafts, painting, Kids Yoga)
Aravind and Arjun
Thank you so much for being a wonderful teacher, Varsha mam. Kids enjoy the class a lot. And they are learning great. You are one of the big reasons for me to continue this classes. Earlier Aravindh had two chess teachers but he did not enjoy the class. Now he likes to play chess. Thanks to you. I dont mind you being strict with them. Im very satisfied with the way you teach. Keep it up mam.
- Priya Vijaykumar, Mother of Aravind and Arjun (Chess)
Book a Course Today!
Nurture Academy offers online courses that are tailored to help you grow academically and professionally. With courses in a variety of subjects, we have something for everyone. Our programs are designed by experts in the field, so you can be sure you're getting the best out of our courses everytime. | eng | 0a5bc4da-e8e1-40cd-b4b5-5bc0c8cab3cb | https://nurtureacademy.co.in/ |
DoLE Region 4-A develops QR code for health checklist
A regional office of the Department of Labor and Employment (DoLE) has developed a Quick Response (QR) code that takes the body temperature and writes down contact details that allows quick tracing of visitors and employees.
The DoLE Region 4-A rolled out the online health checklist which functions like a digital logbook using QR codes to contain the spread of COVID-19.
"These online services aim to respond to the challenges brought about by the COVID-19 pandemic by developing new systems, innovating ways to communicate, and in bringing fresh approaches in providing better client experience while observing health and safety protocols," DoLE Region 4-A Regional Director Henry John S. Jalbuena said.
He said contact tracing is an essential step that is why the online health checklist is being implemented to speed up the current manual processes and to make it quicker to stop the spread of the virus, particularly if the restrictions are eased.
All employees and clients of DoLE Region 4-A who would like to enter the office are required to scan the QR code using their mobile phones or tablets before they can enter the premises.
Those with Apple gadgets can use their cameras to scan the QR code while Android users are encouraged to download the Google Lens application to scan the codes.
Users will be required to provide their contact information after scanning the code posted at the office entrance and answer a quick health symptoms questionnaire which will be evaluated by the safety officers on duty.
"It is only for one purpose, to help in contact tracing. It aims to assist public health authorities to find those persons who are at risk more quickly, thus helping them to get early diagnosis and treatment and to ensure that everyone is protected at all times," Jalbuena said.
He assured the public that all their personal information can only be accessed by the administrator of the system and is protected by the Republic Act No. 10173 or the Data Privacy Act of 2012. | eng | 844cb5c7-0812-4182-abfd-25b5c0293e73 | https://mb.com.ph/2020/08/30/dole-region-4-a-develops-qr-code-for-health-checklist/ |
Abstract
title = "Entrepreneurial success of cottage-based women entrepreneurs in Oman",
abstract = ".",
N2AB | eng | 120d0a39-d122-4096-bfe8-1378b07f95c9 | https://pure.hud.ac.uk/en/publications/entrepreneurial-success-of-cottage-based-women-entrepreneurs-in-o |
***** Space Probe *****
Author: drHirudo@Amigascne.org (Ventzislav Tzvetkov)
Platform: Apple ][
Version: 1.1
Stripped for the 1K minigame compo 2003.
Intro:
Pass through the five deadly zones with the help of your Lazer beam.
Use the following keys to control your ship:
A - Up
Z - Down
- Fire.
- Speed up.
or alternatively use Joystick 1.
It works on every Apple ][ series computer, so CAPS DOWN on any model
greater than Apple II+, for example IIe or IIc or GS. Currently the disk
image is with DOS derivative, but it was tested under ProDOS and worked fine.
Good Luck.
If you think that the game is easy or too slow, then try to compile it with
TASC for example (Available on White Disk 09B) and it will become much faster
and harder (And bigger too). Or alternatively the easiest way is to set
faster speed on your emulator,or use accelerated Apple ][ or AppleIIGS.
P.S. Currently because of the size optimization BASIC hacks used in the game,
it wouldn't compile with any compiler.
Developed entirely on C= Amiga 4000/040 computer.
Emulators:
For Amiga - Apple2000e. Available from:
For Macintosh - Catakig, MacE, Stop The Madness, XGS-Mac etc..
For PC - Appler, ApplePC, AppleWin, AppleEmu etc..
Visit my site for the Apple ][ version of AntiISDA Warrior!
_ ` _ _
| \ C) |_| | C) | | | \ / \
|_/ |\ | | | |\ |_| |_/ \_/
| eng | 4641c9af-c0fc-43f8-8f12-8b2e014b1a38 | https://mirrors.apple2.org.za/ftp.apple.asimov.net/images/games/action/Space_Probe.txt |
HIRAKI - Stumbling Through The Walls 12" LP
Regular price
€18,00
Tax included.
"The album's contorting but just about relentlessly driving rhythms feel jarring, like fiery psychological unease. Overall, HIRAKI expertly interweave punk abrasion with disorienting synths, and the group's venomous onslaught could be compared to groups like the avant-garde metal powerhouse The Body, who HIRAKI cite as an inspiration." - New Noise Magazine | eng | 202e97ef-ac79-4f02-94e8-b997e945b95e | https://www.icorruptrecords.com/products/hiraki-stumbling-through-the-walls-12-lp |
Case Studies
Putting a Manufacturing Business Back on the Map: Our Journey with Allied Systems How do you handle the scarcity of materials and products? How do you ensure efficient production times? How do you overcome the talent shortage? Manufacturing plants need to resolve these tough questions regularly and get the answer right. The pandemic only
Manufacturing companies put a significant amount of resources into hiring and training the right employees, and this is no different for temporary positions. Therefore, why is it so difficult to retain contingent workers? The Problem Jubilant HollisterStier (JHS) is a contract manufacturer of sterile and non-sterile injectables and liquids. They employ a large contingent […] | eng | 5c86cc5b-c73c-465b-9424-5d44f1a4b0a2 | https://atlasstaffing.com/category/case-studies/ |
Looking to craft a glass bottle in Minecraft? I've got you covered! In this guide, I'll walk you through the step-by-step process of creating your very own glass bottle within the game.
How to Make a Glass Bottle in Minecraft
To make a glass bottle in Minecraft, the first material you'll need is sand. Sand is a common resource found in abundance near beaches, deserts, and riverbanks. When mined with any tool (even bare hands), it drops as a block that can be collected.
Finding and Smelting the Necessary Minerals
Once you have enough sand, it's time to find the necessary minerals to turn it into glass. You'll need to locate both coal and silica (also known as quartz). Coal is usually found deep underground while silica can be obtained by mining quartz blocks in the Nether.
Coal: Coal is an essential fuel for smelting items in Minecraft. It can be found within caves or by mining coal ore blocks using a pickaxe.
Silica (Quartz): Quartz can be obtained by mining nether quartz ore blocks commonly found within the Nether dimension. Once mined, smelt them into quartz using a furnace.
Crafting a Furnace to Melt the Glass
With your materials gathered, it's time to craft a furnace – an essential block used for smelting various items including glass.
Placing and Using: Place the furnace on solid ground and open its interface by right-clicking on it. Then place your sand inside along with some fuel like coal or charcoal to start melting process.
Crafting a Furnace
In Minecraft, crafting a glass bottle requires the use of a furnace. The furnace is an essential tool for smelting ores, cooking food, and turning sand into glass. Let's go through the steps to create a furnace and gather the necessary materials.
Collecting Fuel for the Furnace
Coal: Mined from underground, coal is a common and efficient fuel source.
Charcoal: Created by smelting wood logs or stripped logs in a furnace, charcoal serves as an alternative to coal.
Wood Logs: A sustainable option for fuel; chop down trees and convert the logs into planks before smelting them.
Dried Kelp: Found underwater, dried kelp blocks can be used as an eco-friendly fuel source.
Keep in mind that different fuels have varying burn times, affecting how long your furnace will operate. Experiment with different options to find what works best for you.
Finding Sand for Glass
Mining and Collecting Sand in Minecraft
When it comes to making a glass bottle in Minecraft, one of the essential resources you'll need is sand. Sand can be found abundantly in various biomes within the game world, making it relatively easy to acquire once you know where to look.
To start your search for sand, head towards beaches or deserts. These biomes are known for their vast stretches of sandy terrain, which makes them prime locations for gathering this crucial material. Simply equip a shovel and start mining the sand blocks by right-clicking on them. Each block will drop 1-4 pieces of sand that you can collect.
Understanding the Importance of Sand for Glass Making
Sand plays a vital role in crafting glass bottles as it serves as the main ingredient required in its creation process. In real life, glass is made by melting silica-rich materials like sand at high temperatures and then rapidly cooling them down. Similarly, Minecraft mimics this process by using sand as a key component when crafting glass bottles.
Once you have gathered enough sand from your mining efforts, you can smelt it into glass blocks by placing the sand into a furnace and providing fuel such as coal or charcoal. The furnace will convert each block of sand into one block of glass.
Congratulations! You have successfully gathered all the necessary resources to create a glass bottle in Minecraft. In the next section, we'll explore how to use your glass bottles | eng | 8f50cbaf-8699-4e8a-88af-b135bd4af5c4 | https://theportablegamer.com/2023/09/12/how-to-craft-a-glass-bottle-in-minecraft-and-find-the-neccessary-minerals/ |
How Does Coronavirus Spread?
With cases of the new coronavirus reported across the globe, health officials are focused on slowing the spread. By understanding how coronavirus spreads, you can take the right steps so you don't get sick and infect others.
Person-to-Person Transmission
Experts believe the virus that causes COVID-19 spreads mainly from person to person. There are several ways this can happen:
Droplets or aerosols. This is the most common transmission. WhenAirborne transmission. ResearchSurface transmission. A less common method is when you touch surfaces that someone who has the virus has coughed or sneezed on. You may touch a countertop or doorknob that's contaminated and then touch your nose, mouth, or eyes. The virus can live on surfaces like plastic and stainless steel for 2 to 3 days. To stop it, clean and disinfect all counters, knobs, and other surfaces you and your family touch several times a day.
Fecal-oral. Studies also suggest that virus particles can be found in infected people's poop. But experts aren't sure whether the infection can spread through contact with an infected person's stool. If that person uses the bathroom and doesn't wash their hands, they could infect things and people that they touch.
The virus most often spreads through people who have symptoms. But it is possible to pass it on without showing any signs. Some people who don't know they've been infected can give it to others. This is called asymptomatic spread. You can also pass it on before you notice any signs of infection, called presymptomatic spread | eng | 99c8047d-2245-4764-9d93-a182c89dc082 | http://en.tcet886.com/how-does-coronavirus-spread/ |
How do you regulate your emotions to manage anxiety?
The stress that causes anxiety does not go away, but we do have the capacity to "optimize" our response to it. Researchers including Alia Crum, a Stanford psychology professor, have shown that it is possible to approach stress as a challenge and an opportunity for performance and growth.
On a neurobiological level, what Crum and others suggest is part of a broader area of research and framing of the brain known as emotion regulation – processes that help us manage all emotional responses, particularly anxiety.
What does emotion regulation mean?
James J. Gross, an expert in emotion regulation and another professor of psychology at Stanford University, defines emotion regulation as "processes that affect the emotions individuals have, when they have them, and how they experience and express them." He also points out that regulation is a set of processes that exist on "a continuum from conscious, laborious and controlled regulation to unconscious, effortless and automatic regulation".
What does this mean in practice? The bottom line is this: While anxiety may appear as a kind of attention-grabbing signal to avoid danger, it need not necessarily cause discomfort, distraction, or otherwise interfere with our natural urge for well-being and balance. We can learn to use mindfulness to reframe a situation, remove the perception of danger, and re-evaluate it as an opportunity to overcome a challenge and create new responses. We have multiple options for managing both the attention to the signal and the anxiety (emotions) and, if it gets to that point, the reaction itself. Our brain is a wonderful thing!
[Related: Stress and anxiety wear down your brain. Here's how to fight back.]
Our brain-body systems are in constant motion towards homeostasis, the state of balance between arousal and relaxation. Every system, from the nervous system to the digestive system, interacts and exchanges signals to respond to a stressor and then regain homeostasis. This also applies to our emotional system. Our negative emotions arise to draw our attention to something that could be dangerous, and then some sort of change or adjustment to make ourselves feel better. In other words, it has a positive purpose. It's the same with anxiety: it's the brain-body's way of telling us to pay attention. Our inbuilt system for managing our negative emotions, especially processing, responding to, and coping with negative emotions, so that we can maintain or return to balance is called emotion regulation.
How emotions are regulated
Anxiety is a bundle of emotions that overwhelms our ability to regulate emotionally. And they mean it because they aim to draw our attention to an area where not everything is as it should be. However, our ability to regulate our emotions is not always predictable. In fact, the degree of a person's capacity for emotion regulation varies depending on a number of factors, such as how we were brought up, our lifestyle, and even our genetic profile. The good news is that we can learn to regulate our emotions more effectively. According to Gross' model of emotion regulation, we have five types of anxiety management strategies that can help manage anxiety and other negative emotions. These; situation selection, situation change, distraction, cognitive change, and response modulation. The first four can cut anxiety before it escalates into an extreme state or a chronic one. Fifth, it is a regulation technique applied after anxiety (or other negative emotion) has occurred.
Let's look at how emotion regulation plays out in real life. Let's say you're waiting for an important job interview after you left your old position six months ago. You feel pressure, self-doubt and fear – fear of rejection, fear of failure, fear of not being measured. It's four days until the meeting, but you're already feeling nervous. Even when you imagine walking through the door of the building, your hands begin to sweat, your heart begins to pound, and your breathing becomes a little shallow. Then you start imagining every possible thing that could go wrong: you might forget to bring your resume, you might wear mismatched socks, or you might forget everything you knew about why you applied for the job in the first place.
Status selection
One option is to avoid a situation that you think will make you uncomfortable or exacerbate your anxiety. Avoiding the situation (skipping the job interview) can reduce fear and stress in the near term; however, it is clear that this will not help you if you want or need that job in the long run. Gross calls this strategy case selection.
Status change
Another option is to change the current situation in a way that makes the expectation or anxiety more bearable or bearable. For example, if you are worried about the call waiting, you can change the situation by asking to hold the call by phone or video conference. This allows you to exert some control over your anxiety and makes you more responsible for feeling that it is bigger than you. Gross calls this a change. I call this the transition from bad anxiety to good anxiety. Your tension has not disappeared; it is under your control and directed only.
distraction
A third option is called distraction, which involves several ways to divert your attention from the worrying situation to something else that attracts your attention. Parents often use this technique with their babies and toddlers. For example, if the young child is afraid of dogs, a parent might shift the child's focus to a funny face as the scary dog moves away. This is a kind of deliberate distraction.
[Related: How to keep your anxiety from spiraling out of control]
cognitive change
The fourth and possibly the most complex of emotion regulation strategies is called cognitive change. In this case, you're actively and consciously re-evaluating or reframing your mindset or attitude: instead of thinking of the job interview as a terrible way to spend your Friday morning, you reframe it as an opportunity to show yourself and your potential employer how much you know. about the role and company or organization; It also improves your self-confidence. The reframe acts as a mental suggestion that reshapes the feeling of anxiety from feeling fear and overwhelmed to feeling excited and challenged.
response modulation
Once you've managed to get yourself through the front door and seated in the interview, anxiety is likely to set in despite the strategies you've used so far to alleviate it. In this case, you are actively trying to suppress or alleviate anxious feelings. Maybe you do some breathing work (for example, deep breathing, which is one of the quickest and most effective ways to calm the entire nervous system) or drink some water. If it was a date and not a job interview that got you excited, you could have a beer or a glass of wine to quench your excitement. These are just a few of the many coping strategies you can use after experiencing anxiety.
You can learn to manage your anxiety
Current research on the interaction between anxiety and emotion regulation points to strong evidence that interventional strategies such as reappraisal can improve one's emotion regulation capacity and positively affect anxiety; these studies were conducted in the context of anxiety disorders. Specifically, neuroimaging studies have shown that negative feelings of anxiety or fear are reduced in response to emotion regulation strategies. Moreover, neuroimaging studies have also shown that negative emotions such as anxiety or fear occur in different neural regions of the brain where emotion regulation occurs. This area of research is in its infancy, but the good news is: We can update our emotional responses. We can learn to regulate emotionally. We can become better at managing and then channeling our anxiety.
I like to think of this approach to anxiety as a way to build our resilience to stress. Consider this: We need to both feel emotions and update our response to those emotions. It starts with awareness. When you notice that you are bothered by any of the signs of anxiety, you need to stop and think about what you are doing with those feelings. We all need to constantly just sit with our emotions and not immediately try to mask, deny, escape, or distract ourselves. By sitting comfortably, you do two things: you get used to the feeling and realize that you can truly "survive", and you give yourself time and space in your brain to make a more informed decision about how to act or react. . This is exactly how a new, more positive neural pathway is established.
Dr. Wendy Suzuki is an award-winning professor of neural science and psychology at the Center for Neuroscience at New York University and Seryl Kushner dean of the NYU College of Arts and Sciences. She is a renowned international authority on neuroplasticity, recently named by Good Housekeeping as one of the top 10 women to change the way we see the world, and serves as a regularly sought-after expert. The Wall Street Gazette, Shapeand Health. His TED talk has been viewed more than 55 million times. he is the author Good Anxiety and Healthy Brain, Happy Life.
Purchasing Good Anxiety: Using the Power of the Most Misunderstood Emotion Here. | eng | 6252ad18-12db-446c-985d-7525ab092e87 | https://nayemleon.com/how-do-you-regulate-your-emotions-to-manage-anxiety/ |
RESEARCH - Youth Drinking and Driving
A new Road Safety Monitor (RSM) poll by the Traffic Injury Research Foundation (TIRF) shows that a majority of young Canadian drivers are aware that they cannot drive safely after drinking alcohol. The public opinion poll conducted in September 2010 investigated a variety of drinking and driving behaviours and attitudes among youth including riding with a drinking driver.
Young drivers aged 16-24 were asked to rate whether they could drive safely after drinking alcohol. An overwhelming 86% of those surveyed agreed or strongly agreed to not being able to drive safely after consuming alcohol. This percentage was also significantly higher than drivers aged 25 and older, of which only 77% agreed to not being able to drive safely after drinking.
This self-reported attitude towards drinking and driving appears to be reinforced by their actions, with only 12% of 16-24 year olds surveyed admitting to driving in the last 30 days after drinking any amount of alcohol. In comparison, 25% of adult drivers aged 25 and older admitted to driving after drinking in the last 30 days. This information is encouraging to researchers as youth are generally at a higher risk of being in a collision, even when sober, than older drivers.
"While youth are still over-represented in road collisions, they make up a smaller proportion of the drinking driving problem," notes TIRF President and CEO, Robyn Robertson. "However some research has shown that when young drivers do drink and drive, they are more likely than older adults to experience an alcohol-related crash. Hence, strategies such as zero-tolerance for new and novice drivers in GDL programs and for drivers 21 or under have been put in place in recent years to reduce this risk." | eng | 97d9038f-5549-493b-8ef1-5494df0e4feb | https://www.drivesmartbc.ca/impaired-driving/research-youth-drinking-and-driving |
Link to comment
tbh I was surprised to see Mood Ring there! I thought Liar would be on there... No shade to Mood Ring but it's literally a Japanese Bonus track like I don't think it's going to happen (no matter how much the song slays my existence )On a more serious note-- I've heard flip sides of both, to answer your question. I've heard of limited singles for other artists/bands, and my friend's two favorite bands release singles regularly with their releases.... but there are only one or two music videos. She was telling me about it the other day when I mentioned being afraid Glory might not have a second single. I've also been hearing a lot of chatter about record labels in general flopping these days. (Moreso than usual, I mean.) | eng | 578a5a75-b736-4045-b95a-7dbdca65f5de | https://exhale.breatheheavy.com/forums/topic/730258-britney-choosing-the-second-single-like-screaming/ |
Successful email confirmation!
Welcome! You are now registered for our webinar series and there is no need to register again in the future. Simply login to the site and participate in the discussions. | eng | bf9f846b-a63b-462b-bc00-7cb7723a180b | https://spedpac.org/successful-email-confirmation/ |
At Battelle, we're not just committed to creating a safer, healthier and more secure world through science and technology. We also must nurture future generations of science, technology, engineering and math (STEM) leaders in a rapidly changing world.
Through our Regional Community STEM Grant Program, we're able to reach talented students who may not have access to STEM programs that support their interests. Directly investing in quality STEM education programs creates more equitable and accessible opportunities for students in underserved communities.
By building relationships with community organizations where Battelle employees live and work, we can equip the next generation of innovative and diverse thinkers, creators and inventors with the skills they need for future careers in STEM disciplines. This year, 47 grants were awarded totaling $1.2 million and reaching more than 71,000 students.
Following are features on some of the regional organizations that have received funding through this year's Regional Community STEM Grant Program.
The Foundation for Glacier & Environmental Research
The Foundation for Glacier & Environmental Research has received funding to engage students in creating a virtual Antarctica. Students will collect 3D video and photography across three sites in West Antarctica, including Mount Waesche, Pine Island Glacier and Hercules Dome, key locations that are host to major research programs.
The collected imagery will be used to develop interactive experiences during field and classroom programs with students from across the U.S. who are part of the Upward Bound Program. Students will develop an Antarctica virtual world, integrating current research from Antarctica into an educational video game while learning how to use the virtual reality equipment, software and hardware. Ultimately, the virtual reality suite created is expected to be used by approximately 150 students during the spring and summer of 2023, and is expected to reach a total of 500 Alaskan students over the next three years. The acquired equipment will also be made available to other programs supporting low-income and first-generation students and will be shared with the Toolik Field Station research station in Alaska through a collaborative partnership as they develop field Arctic educational programs.
iSchool for the Future
Through a scaled professional development program aimed at supporting and developing educators in Title 1 schools, iSchool for the Future will support 3,000 students and 50 teachers by incorporating social and emotional skills development and arts into project-based STEM learning for students in grades K-8.
With Battelle funding, this teacher development program will help educators integrate social and emotional learning into existing curriculum through hands-on, project-based activities designed to increase student engagement with STEM subjects. This program also strengthens teachers' confidence and effectiveness in teaching and modeling social emotional learning skills, which are critical to success, while providing best practices, organized resources and learning experiences to ensure teachers are equipped with practical and innovative strategies to engage students in STEM learning experiences.
Wolf Trap Foundation
The Wolf Trap Institute for Early Learning Through the Arts (Wolf Trap Institute), a signature program of the Wolf Trap Foundation, introduces professional performing artists, musicians, dancers, actors and puppeteers into early childhood programs. This arts-based education model supports young children's cognitive, social, emotional and physical development and enhances teachers' instructional skills through professional development.
With Battelle's support, the Wolf Trap Foundation will provide services to early childhood educators across three school districts with the aim of establishing long-term, sustainable partnerships at the district level. This program aims to increase school readiness for children living in economically disadvantaged circumstances and provide teachers with access to effective professional development.
Search
INSIDE BATTELLE
Inside Battelle gives you an all-access pass to the incredible work of Battelle researchers. | eng | de97277b-a131-4a00-b41d-2f444807a373 | https://inside.battelle.org/blog-details/battelle-reaching-students-across-the-u.s.-through-regional-community-stem-grants |
Crowdfunding
Crowdfunding is a technology-enabled financial service often used by start-up companies or growing businesses as a way of accessing alternative financing. It is an innovative way of sourcing funding for new projects, businesses or ideas. The main types of crowdfunding include peer-to-peer lending, equity crowdfunding and rewards-based crowdfunding.
Until recently, EU crowdfunding markets for business finance were largely underdeveloped compared to other major economies. Most importantly, this market was unable to properly operate across borders. Hence, in an effort to provide legal certainty for investors and promote common rules across the EU for crowdfunding, the European Commission presented a proposal to regulate crowdfunding and peer-to-peer lending in Europe during the first quarter of 2018.
On 5 October 2020, the European Parliament adopted the final text for a "Regulation on European Crowdfunding Service Providers for Business" (Regulation 2020/1503 or the "Regulation"). The Regulation comes into effect on 10 November 2021 and aims to help crowdfunding service providers function smoothly in the internal market and foster cross-border business funding in the EU.
The scope of the Regulation captures both lending-based crowdfunding and investment-based crowdfunding. In relation to investment-based crowdfunding, the Regulation will apply to all crowdfunding service providers facilitating investment-based platforms dealing with transferable securities for campaigns raising up to €5 million. Larger operations will be regulated by the Markets in Financial Instruments Directive (MiFID II Directive) and the Prospectus Regulation (EU) 2017/1129.
In relation to lending-based crowdfunding, the Regulation applies to crowdfunding services that consist of the facilitation of granting of loans, including services such as presenting crowdfunding offers to clients and pricing or assessing the credit risk of crowdfunding projects or project owners.
Lending to consumers is not covered by the Regulation, while reward and donation-based crowdfunding also fall outside the scope of the rules.
The European Parliament also adopted Directive (EU) 2020/1504 (the "Directive"), which amends Directive 2014/65/EU (MiFID II Directive). The Directive exempts crowdfunding service providers falling under the scope of the aforementioned Regulation from the application of the MiFID II Directive. Member states are required to transpose the provisions of the relevant Directive into their national laws by 10 May 2021.
The Regulation has two specific objectives. First, it seeks to enable platforms to scale cross-border, namely by creating a licensing regime that can be used across the EU without requiring further authorisation in each EU country. Second, the Regulation wants to guarantee that these platforms are subject to a framework that ensures their proper management and the protection of fund providers by focusing on sound risk management and adequate information disclosure.
A prospective crowdfunding platform will need to request authorisation from the national competent authority of the member state where it is established. Further to notification, the crowdfunding service provider can also offer its services in other EU member states. Supervision will be carried out by the national competent authorities with ESMA facilitating cooperation between the member states.
The Regulation introduces certain provisions that:
Impose a total threshold for crowdfunding offers made by a particular project owner. This threshold is set at €5 million and calculated over a period of 12 months.
Provide for authorisation and supervision of crowdfunding service providers.
Provide for the establishment of a European register to include all crowdfunding service providers.
Include organisational, operational requirements and provisions for the effective and prudent management of crowdfunding service providers.
Include provisions for complaints handling, conflicts of interest and outsourcing.
Impose requirements applicable to the marketing communications of crowdfunding service providers.
Set rules for investor projection, including:
Distinguishing between sophisticated and non-sophisticated investors and introducing different levels of investor protection safeguards applicable to each of those categories.
Requiring crowdfunding service providers to provide clients (i.e. investors or project owners) fair, clear and non-misleading information concerning the company, associated costs, financial risks and charges related to crowdfunding services or investments. This also applies to the provision of information on the crowdfunding project selection criteria and the nature of and risks associated with their crowdfunding services.
Laying down the content of the Key Investment Information Sheet (KIIS) that will be provided by the crowdfunding service providers to prospective investors for every crowdfunding offer in order for them to make an informed investment decision.
Introduce due diligence requirements for project owners that would like their projects to be funded via the crowdfunding service provider's platform.
Furthermore, the Regulation takes into consideration the fact that several member states have already introduced domestic bespoke regimes on crowdfunding and national rules that diverge across the European Union (see for example, CySEC Directive DI87-10 and Policy Statement PS-01-2020). The Regulation provides for transitional arrangements so that persons providing crowdfunding services in accordance with their national law now fall within the scope of the Regulation and can adapt their business activities to comply with the Regulation. Such persons may continue to provide crowdfunding services that are included within the scope of the Regulation in accordance with the applicable national law until 10 November 2022.
Our multidisciplinary team can provide comprehensive and holistic solutions tailored to each client's individual needs, including:
Assistance in obtaining relevant authorisation for entities that wish to act as crowdfunding service providers under the scope of the Regulation;
Consulting services and guidance to authorised crowdfunding service providers in order to comply with the provisions of the Regulation;
Money laundering compliance;
Assistance in preparing the Key Investment Information Sheet (KIIS);
Provision of answers to relevant queries (Q&A Service), and;
Regulatory updates.
MAP S.Platis closely monitors this industry and our team of seasoned specialists is always standing ready to assist clients adapt and prepare for any new and challenging regulatory developments experience | eng | e6c0dd2b-9bc9-4626-b45e-9c16583c946e | https://www.mapsplatis.com/sectors/crowdfunding/ |
Body, Part 2: Body Neutrality in Youth Ministry Spaces it here!
Over the last few years, the Body Positivity movement has been making waves on social media, marketing campaigns, and more. For many, it has been a long-awaited moment where individuals of all sizes, abilities, and backgrounds are being represented. But, body positivity can feel like a reach for many people, especially youth, as their bodies are constantly changing, feeling foreign and awkward. From this reality, another movement has emerged: Body Neutrality. But you may be wondering, what exactly is Body Neutrality?
Body Neutrality is the space between disliking our bodies and loving them – for so many of us, loving our body can feel like a stretch. One of the main goals of neutrality is to lessen negative self-talk about your body and living authentically and peacefully in the body you have. It focuses on everything your body does for you every day, rather than what it cannot do. This can feel immensely more attainable than loving your body every day because the reality is putting on a smile doesn't change our relationship to our bodies. Body positivity can be a beautiful thing, but it often takes years of work to truly learn to accept then love the bodies we exist in. Body Neutrality takes away what can feel false about the body positive movement and allows us to be real with ourselves about how we're feeling while still recognizing our body as good.
But why does this matter, and what does it have to do with Youth Ministry? How does seeing our bodies as strong and capable affect our and our youth's mentality? I want you to sit for a moment and think about any ads you may have seen today for diets, or if you've heard anyone speaking recently about getting that "summer body" they've been desiring… When we are wrapped up in negative body image and body talk, we are more likely to feel bad about who we are. The obsession with our bodies and how they are received in the world starts from a young age and impacts self-esteem, which directly affects mental health, suicidal ideation, substance abuse, and more. And let's not forget that the diet industry is using this negative self-talk against you to make money; currently, it is a 50 billion dollar industry.
So how do we, as women in youth ministry, create spaces for all youth, leaders, volunteers, and more to come and feel comfortable to be themselves and not fear body talk?
Here are some actionable steps you can take to make your youth ministry space more body neutral:
Focus on Language | Our society is a little obsessed with commenting on other people's bodies, diets, and food choices. The first big step we can take is to just try talking about something else, anything else. When you want to compliment someone, talk about their intelligence, their humor, their joy – and avoid talking about their body. Words such as healthy, fat, diet, heavy, thin, etc., can be incredibly damaging to youth. Even more sublte coded language ("you're looking so fit!" "I'm just trying to eat clean") can have an impact.
Redirect Body Talk/Shaming | This may take some practice and patience. We have become so accustomed to body talk, we may not even recognize it happening in our groups, but it most likely is. If you hear someone say, "I'm on a diet so…" or, "I feel fat…" or, "This shirt looks stupid on me…", or anything else like this, offer a redirection. Attempt to change the conversation away from body talk and towards something else.
Speak 1-on-1 with Youth | If certain youth have a habit of chronic body talk about themselves or others, pull them aside and check in with them. Youth may feel pressure to look or act a certain way. By walking alongside them to change these behaviors, you can offer support. If you feel there is something larger happening that is beyond your purview, it may be time to speak with their guardian. Most eating disorders and body dysmorphic disorders start in adolescence, and some studies suggest that 10 in every 100 youth struggle with an eating disorder. As you check in with youth, resist the urge to fight fire with fire. Responding to negative body talk with positive body talk ("But you're so skinny!" "You shouldn't feel bad about your body, I would never fit in those clothes") doesn't do the crucial work of unrooting the issue itself.
Watch "Diet" Talk at Meals/Snacks | Youth Ministry is centered around food at times; we bring snacks, eat meals together around the table, but this time can be when most negative talk around food occurs. Avoid speaking about "healthy" options or certain diets – diet culture is rooted in our society, so breaking this habit can be difficult. Remember, no food is truly good or bad, food is just food, it doesn't have moral value. Also, keep mental notes if you notice youth who avoid eating around the group or eat just a few bites and play with their food. If this becomes a concern, make sure you connect with them and possibly their guardian.
Ask Questions and Keep Learning | Creating body-neutral spaces takes time; if you feel like it's not working, keep at it. Also, keep learning! There are so many people working in this space to promote body neutrality. Below is a list of Authors to Read, Documentaries to Watch, and Podcasts to Listen to!
Remember, youth pastors are not therapists, so if you see behavior around food that you think might be dangerous to a young person's well-being, help parents find the resources they need to get professional help for their children. For many people suffering from eating disorders, there is a gap of years between when the disordered behavior begins and when they receive treatments. This is particularly true for those who don't fit our mental image of eating disorders, such as men, non-white folks, LGBTQ+ people, and people in larger bodies.
This work isn't easy, but it is important, and it is worth the effort we put into it. Every time we make our youth spaces safer and more inclusive, we invite more people into experiencing the love of Christ.
This blog is the second-positive spaces in youth ministry and how to help our young people see their bodies as what they are: not problems to be solved, but gifts from their Creator, given endless worth, and made in the very image of God.
Christina Rees-Fletcher | eng | 1b81310b-bfac-46ee-8383-a3f9d3f62f07 | https://womeninyouthministry.com/2021/05/11/body-part-2-body-neutrality-in-youth-ministry-spaces/ |
Linking Cubes
Tub of 1000
73423
£20Approx. 100 of each colour
Multifunctional connecting cubes
Teaches various maths skills encouraging learning through play
Full Description
Linking Cubes provide mathematical learning experiences in developing children's skills in counting, classifying, addition, subtraction, place value, measuring and patterning. A versatile tool for all levels of learning and proficiency in the classroom. 10 bright colours. | eng | ae4d6731-8394-4118-975b-716c9b423008 | https://www.espo.org/linking-cubes-73423.html |
Message from Primary Head Mistress
I am honoured to serve as the Primary Headmistress of this wonderful institution, where
we are dedicated to nurturing young minds and shaping the leaders of tomorrow. Our
school is not just a place of learning; it is a vibrant community where dreams are
cultivated and a lifelong love for learning is ignited.
At the primary level, education forms the cornerstone of a child's development. It is
during these formative years that they acquire fundamental knowledge, values and skills
that will shape their future. Our primary department is not just a place of learning; it is a
vibrant and nurturing environment where curiosity is celebrated, creativity is encouraged
and every child is valued.
Our dedicated team of educators is not only highly qualified but also deeply passionate
about fostering a love for learning. We recognize that each child is unique and we tailor
our teaching methods to suit the individual needs, ensuring that every child reaches
his/her full potential. We believe in the holistic development of our students, not only
academically but also socially and emotionally. We instill in them the values of respect,
empathy and integrity, which will serve as their guiding principles throughout life.
Our curriculum is designed to be engaging and relevant, providing a strong foundation in
core subjects while also promoting critical thinking, problem-solving and a lifelong love
for learning. We embrace modern educational approaches and technology to enhance the
learning experience, preparing our students for the challenges of the 21st century. Using
technology effectively, improving the infrastructure continuously, enhancing skills &
competencies regularly, educating & assessing with the latest norms of NEP seamlessly,
training & empowering the teachers efficaciously are few of our initiatives.
As primary headmistress, I am committed to fostering a strong partnership between the
school, parents and the community. We believe that the education of a child is a shared
responsibility, and together, we can create a supportive and nurturing environment in
which our students can thrive.
I invite you to explore our school website to learn more about our primary department
programmes, extracurricular activities and the various initiatives we undertake to ensure
a well-rounded education for our children. Please feel free to reach out to us with any
question or concern you may have. Your involvement and feedback are essential in
shaping the educational journey of our young learners.
I look forward to a fruitful and enriching partnership with all of you as we embark on
this exciting journey of educating and nurturing the primary department children of
Gyan Mandir Public School. Together, we will help them grow into confident,
compassionate and capable individuals who will make a positive impact on the world.
Thank you for entrusting us with the privilege of educating your precious ones.
Warm regards
Neeru Sachdeva
Primary Headmistress | eng | 9377122f-d189-477e-9252-ec62f51f5020 | http://gyanmandirpublicschool.com/head-mistress.html |
Additionally, our measurements make considerations for different body types. If you're looking for something slightly less snug, please order a size up | eng | e5b37850-0b92-4322-9542-3207c445f7eb | https://www.iwoot.us/clothing/daddy-bear-women-s-t-shirt-grey/11612799.html |
What Is Passwordless Authentication and How Does It Work?
At first, passwordless authentication seems like a radical idea. But you're probably already using it. Whenever you receive a "magic link" to your email address or a "one-time passcode" to your phone, you're actually engaging in passwordless authentication.
Passwordless authentication is being adopted as an ultra-secure method of protecting accounts. Let's take a look at the basics of passwordless authentication, its benefits, and why your organization should consider making the switch.
The Basics of Passwordless Authentication
So, how does passwordless authentication work? Passwordless authentication is a method of authenticating users without the use of passwords. Instead of a password, the user is authenticated using another factor, such as a one-time code sent to their mobile phone or email address, a USB security key, or biometrics.
While it may sound complicated, the reality is that passwordless authentication is simple—in fact, one of its most important advantages is how simple it truly is. Thus, passwordless authentication can simplify an organization's security while also improving productivity for employees.
Passwordless Authentication vs. Multi-Factor Authentication
Passwordless authentication is often confused with multi-factor authentication (MFA). MFA also relies on additional factors for authentication, but it always includes a password as one of those factors.
With MFA, even if an attacker were to gain access to one of the other authentication factors, they would still need the password to log in.
Passwordless authentication, on the other hand, removes the password entirely. Even if an attacker were to gain access to one of the other authentication factors, they would not be able to log in without also having possession of the user's phone or email account.
Both passwordless authentication and multi-factor authentication can be great security systems.
The Benefits of Passwordless Authentication
There are several important benefits of passwordless authentication"
Increased security. It's is more secure than traditional password-based authentication. Without a password, attackers have one less factor they can try to brute force.
Improved usability. Passwordless authentication is also more user-friendly than traditional password-based authentication. Users no longer need to remember a password, and they can authenticate using a device they already have with them, such as their phone.
Reduced support costs. Going passwordless can also reduce support costs. With traditional password-based authentication, users often forget their passwords and need to reset them. With passwordless authentication, there are no passwords to forget, so users are less likely to need help from support.
In short, passwordless authentication increases security without increasing work for employees and administration. But it also does require that organizations make a switch to an entirely different password infrastructure.
Passwordless Authentication and Zero Trust
Passwordless authentication is a key component of zero-trust security. Zero trust is a security model that does not rely on predefined trust levels.
In other words, with zero trust security, every user and every device is treated as untrusted. This might sound like it would make security more difficult, but in reality, it makes it much more effective.
By treating all users and devices as untrusted, you remove the need for predefined trust levels. This makes it much more difficult for attackers to exploit trust relationships.
In a zero-trust environment, passwordless authentication can be used to verify the identity of users before granting them access to resources. Zero trust is primarily a philosophy, but passwordless authentication runs parallel to it because users are not trusted merely because they know authentication credentials.
Single Sign-On and Passwordless Authentication
Single sign-on (SSO) is an authentication method that allows users to access multiple applications with a single set of credentials.
SSO is often used in conjunction with passwordless authentication. With SSO, users can authenticate using a one-time code or biometrics, and then they are given access to all of the applications they are authorized to use.
SSO can be used with any type of authentication, but it is particularly well-suited for passwordless authentication. When combined, these two technologies provide a high level of security and convenience for users.
How to Implement Passwordless Authentication
There are many different ways to implement passwordless authentication. The most common methods are via one-time codes, biometrics, and FIDO tokens.
One-time codes can be delivered to the user's email address or phone number. The user then enters the code to authenticate.
Biometrics can be used for authentication on devices that have a biometric sensor, such as a fingerprint sensor.
FIDO tokens are physical devices that generate one-time codes. The user plugs the token into their computer and enters the code to authenticate.
Organizations can also use a combination of these methods. For example, they might require a one-time code for initial authentication and then use biometrics for subsequent authentications.
Choosing the methods that will work for an organization is essential. The easier a security system is to use for employees, the more likely it will be to be used the right way.
That being said, a transition to passwordless authentication systems isn't always simple. It requires a change in mindset for many organizations, as well as a commitment to educating employees about the new system. Employees may at first find the idea of passwordless authentication foreign.
Switch to Passwordless Authentication with Axiad
Now you know how passwordless authentication works and its advantages. If you're still using traditional password-based authentication, you should consider making the switch to passwordless authentication or, at the very least, multi-factor authentication.
Passwordless authentication is more secure and user-friendly than traditional authentication… and it can also reduce support costs.
Axiad can help. Axiad's Unified Credential Service makes it easier for organizations to control their login services — within a single, easy-to-use platform. Contact Axiad today to find out more about the benefits of passwordless authentication and how it can help protect your organization.
Axiad Demo
See How Axiad Works
See a comprehensive demo of Axiad and envision how it will revolutionize authentication for you! | eng | 1adde392-4fcc-4b87-bd3c-f4893b411273 | https://www.axiad.com/blog/what-is-passwordless-authentication-and-how-does-it-work/ |
Featuring mechanical switches, this keyboard offers a satisfying tactile feedback and quick response time, ensuring every keystroke is registered accurately. The mechanical switches are also durable, with a lifespan of up to 50 million keystrokes, making this keyboard a long-lasting investment. Additionally, the floating keycap design not only adds a touch of elegance but also makes cleaning a breeze, as dust and debris can be easily removed.
Designed with gamers in mind, the Logitech K835 comes with customizable RGB lighting, allowing you to personalize your gaming experience. You can choose from a wide range of vibrant colors and effects to match your gaming setup or create your own unique lighting profiles. The keyboard also features anti-ghosting technology, ensuring that every command is executed flawlessly, even during intense gaming sessions. Whether you're a professional gamer or simply enjoy gaming in your free time, the Logitech K835 Mechanical Keyboard is an excellent choice for a seamless and immersive gaming experience.
Introducing the Xiaomi Youpin Miiiw Wireless Mute Mouse, a cutting-edge peripheral that combines functionality, convenience, and silence. With its sleek design and advanced features, this mouse is a must-have for any tech-savvy individual. The wirele...
Introducing the DK61 60 Percent Hot Swap Wireless RGB Backlit Mechanical Keyboard Kit, the ultimate companion for any PC gamer or keyboard enthusiast. This compact and sleek keyboard offers a modern design with a 60% layout, providing a more immersiv... | eng | 4d25af5b-5c34-4817-bfc7-67db71337ea5 | https://www.cheap-android.net/products/computer---office/computer-peripherals/logitech-k835-mechanical-keyboard-wired-gaming-keyboard-tkl-84-key-floating-keycap-for-desktop-laptop-pc-office-gamer-keyboard/99033.html |
Legal Assistant vs Paralegal: What's the Difference?
I went to dinner with a friend last week, and she told me she's looking for a new career. She's interested in work that involves the law but doesn't want to go to law school due to both the time commitment and the cost. We tossed around a few ideas, including the possibility of working in a law firm in a different role that didn't require a JD law degree. As we talked about the different roles in a law firm, we kept coming back to legal support roles, which led to the question: what is the difference between a legal assistant versus a paralegal?
Both legal assistants and paralegals work in law firms to support licensed attorneys in their legal work. Neither of these positions requires a law degree, and they both support the law firm in different ways. Neither position is allowed to give legal advice.
A legal assistant performs primarily administrative work, such as scheduling appointments, organizing files, answering the phone, and transcribing audio dictation. Paralegals can draft legal correspondence, documents, and forms that will be reviewed by the attorney, perform legal research, and interview clients.
What Is a Legal Assistant?
A legal assistant position focuses primarily on administrative work that supports the lawyer and the firm at large. Some firms may also refer to this role as a Legal Secretary. However, that terminology has become less frequent as it's easier to bill clients for the services of an assistant instead of a secretary Duties Does a Legal Assistant Perform?
Legal assistants will perform a variety of administrative tasks. Although many of their functions may be client-facing, they typically serve as a conduit between the client and their attorney. They can answer phone calls and take messages or questions to pass on to the attorney, but they will not answer any questions themselves.
Legal assistants are often responsible for an attorney's calendar, including scheduling meetings or phone calls with their clients. They may also work with other legal assistants at the law firm when scheduling time with other attorneys or at firm events and activities. People who need to reach an attorney will likely get their legal assistant first, as this role can act as a gatekeeper to the attorney's availability.
Legal cases often generate a lot of documents, and legal assistants are usually responsible for keeping everything organized. They can receive, sort, label, and file documents from clients, often using a law firm's content management system. This organization is essential as it ensures the paperwork and other documents are easily found by paralegals and lawyers when it comes time to work on a case.
Lawyers often use dictation to write their legal briefings, create memos, or write other correspondence. These dictations are often given to legal assistants to transcribe into written documents that are subsequently returned to the attorney for review and signature.
Common skills for legal assistants include:
Word processing
Transcription
Scheduling
Data entry
Organizing and filing
General client communication
What Duties Can Legal Assistants Not Perform?
Legal assistants are not permitted to give clients legal advice or legal assistance as they are not licensed to practice law. In most cases, legal assistants do not specialize in a specific area of law as their duties and responsibilities are more general and can be performed within any law firm.
What Level of Education Do Legal Assistants Need?
Most legal assistant positions require a high school diploma as the minimum level of education. Some legal assistants will have a bachelor's degree, but they don't need any specific degree or course specialization.
Most law firms will provide additional on-the-job training for their legal assistants. This training will focus on the needs of that specific law firm, including any specialized document management systems that they use. In-house legal assistant training will also cover the law firm's preferences for interacting with clients and attorneys, such as email systems, phone responses, and written correspondence.
How Much Money Do Legal Assistants Make?
The salary for a legal assistant can vary greatly depending on their geographical location within the United States. Legal assistants in large cities tend to have higher wages than those in smaller towns or rural areas.
According to the employment website Indeed, the average annual salary for a legal assistant in the US is $42,573. They also indicate that the highest-paying cities for legal assistants are:
New York, New York
Los Angeles, California
Philadelphia, Pennsylvania
Chicago, Illinois
Tampa, Florida
What Is a Paralegal?
A paralegal position focuses on directly supporting a lawyer as they work on a case and prepare to go to trial. Although some of their work may be administrative, they perform more detailed tasks and typically have a deeper understanding of the law than a legal assistant.
What Duties Does a Paralegal Perform?
Unlike legal assistants, paralegals often specialize in one specific type of law, such as immigration, personal injury, criminal defense, or family law. Their level of knowledge within their chosen sub-area won't be as high as a licensed attorney's, but they will be able to perform highly specialized work.
One of the primary responsibilities of a paralegal is performing legal research, often using specialized legal services and databases. They must know how to accurately find relevant information within these systems and summarize the critical points into a shorter document for the attorney to review. Paralegals must be able to perform this research work quickly and accurately.
Paralegals are also responsible for drafting many different legal documents for their supervising attorneys. These documents can be client-facing, such as memos, emails, and letters. They can only sign their own name to correspondence that doesn't include any legal advice or recommendations; otherwise, the paralegal must draft everything before it is reviewed and signed by the attorney.
Paralegals will also draft legal documents that they may file with a government agency or with the court. Again, these documents must be reviewed and signed by the attorney before submitting them.
Some types of legal documents that a paralegal may work on include:
Legal forms
Formal petitions
Cover letters
Legal summaries
Lists of evidence or documentation
Paralegals will also play a role in submitting some court documents or filing documents with the opposing counsel. They are again supervised when performing these duties, but filing these documents doesn't usually require a separate sign-off by their attorney.
Some common skills for paralegals include:
Legal research
Legal documentation
Familiarity with government agencies
Writing and editing
Communication and correspondence
What Duties Can Paralegals Not Perform?
Since they are not licensed attorneys, paralegals are not permitted to give legal counsel or advice. They can take information from their attorney and give it to the client. They can also answer questions about what certain legal forms or documents mean.
What Level of Education Do Paralegals Need?
All paralegal positions require at least a high school diploma, and most will need either a specialized associate's degree or a bachelor's degree. There are no required fields of study for paralegals with a bachelor's degree. Still, many paralegals will have a background in an area that is writing-intensive or that focuses on the legal system, such as criminal justice.
Many, but not all, paralegals will also choose to attend a post-secondary certificate course to become certified. These courses are typically one semester in length and can be either in-person or online. Some schools will also offer a hybrid option.
These certificate programs will teach new paralegals the specific legal research and writing skills they'll need to perform their jobs. They'll also cover many standard tools and systems used in law offices, such as internal document management systems and court filing systems. Some paralegal certificate programs will focus on specific areas of law to better prepare paralegals for those niche fields.
How Much Money Do Paralegals Make?
The amount of money that a paralegal makes can vary depending on their level of education if they've received any specialized training and their geographical location within the US. Paralegals with a bachelor's degree or a paralegal certificate tend to make more money than those with an associate's degree. Also, some specialized areas of law will pay more for paralegals who are trained or experienced in that area.
Employment company Indeed lists the average salary for a paralegal in the United States as $49,177. This is just under $5,000 more per year than the average salary for a legal assistant. They also indicate that these are the highest paying cities for paralegals:
Washington, DC
Dallas, Texas
Chicago, Illinois
Phoenix, Arizona
New York, New York
Is It Better to Be a Legal Assistant or a Paralegal?
Both legal assistants and paralegals play an essential role in the legal field. Both positions allow people to work for a law firm or independently practicing attorney without needing to attend law school.
A legal assistant position is a better fit for those interested in performing more administrative work. These employees must be highly organized and adept at juggling projects with many moving parts. Legal assistant jobs usually require less formal education than a paralegal.
The paralegal role is a better fit for those who want to do legal research and writing. These positions need someone who can consume, understand, and summarize dense legal information. They also need to be able to write coherent and highly structured legal documents and communicate accurate information to clients. Paralegal jobs usually require post-high school education, such as an associate's or bachelor's degree. | eng | a724c934-3f13-4f8e-afb3-9cf0225753da | https://elawtalk.com/legal-assistant-vs-paralegal/ |
done? Check your answers below! Scroll Down to get the ANSWERS TO THE QUIZ
1) How long did the Hundred Years War last? 116 years 2) Which country makes Panama hats? Ecuador 3) From which animal do we get catgut? Sheep and Horses 4) In which month do Russians celebrate the October Revolution? November 5) What is a camel's hair brush made of? Squirrel fur 6) The Canary Islands in the Pacific are named after what animal? Dogs 7) What was King George VI's first name? Albert 8) What color is a purple finch? Crimson 9) Where are Chinese gooseberries from? New Zealand 10) What is the color of the black box in a commercial airplane? Orange , of course. | eng | 6d7c7e53-7b4f-4893-b89f-5f92bc4f99fd | http://www.geekstogo.com/forum/topic/184896-easy-test-4-correct-is-a-pass/ |
Mark Strong, Votes: It was the first sci-fi film to show humans traveling faster than light in their own spacecraft, and it was also the first film to take place entirely on another planet, outside the Solar System, in another part of the galaxy with not a single glimpse of Earth.
124 min
Téa Leoni, Let's hope that the common thread of all of them -- humanity getting off the Earth and finding its way to the stars and beyond -- someday comes true. | | | A group of intergalactic criminals must pull together to stop a fanatical warrior with plans to purge the universe. | 121 min I've always loved the idea of Stargate -- using an ancient alien portal to leap across the galaxy in seconds -- and it put a different spin on the concept of space exploration (namely, no ship necessary). Stars: But for its first half, Prometheus not only channels the vague sense of dread that also permeated director Ridley Scott's Alien, but something else -- a sense of mystery and awe as the crew of the title ship goes in search of the origins of humanity itself. | | $54.81M, TV-14 Stars: |
Jamie Bamber, Mary McDonnell, Votes: John Hurt, The movie suffers from budgetary and script problems, but it has some effective twists and the idea of the effects of deep space travel on human beings is a compelling one. Stars: | | |
Two exploratory ships are lured to an uncharted planet by a distress signal, where an unknown force causes them to crash land and possesses the corpses of one crew, reanimating them. The legendary debut of director John Carpenter, Dark Star started out as a student film by Carpenter and writer Dan O'Bannon (who later penned Alien and stars here as Pinback). Ryan Reynolds, Mel Brooks An astronaut becomes stranded on Mars after his team assume him dead, and must rely on his ingenuity to find a way to signal to Earth that he is alive. Red Planet, on the other hand, starts off in similar fashion but overloads its voyage to Mars with a killer robot, alien insects and more plot contrivances. Paul Reiser, Votes:
This indie production is a relative rarity in that it is a sci-fi story told through the "found footage" format, in this case the video logs of a six-person mission to the Jovian moon Europa in search of signs of life.
Exploring the many ways space, space travels, and extraterrestrial life are depicted in film, TV, literature, anime, comics, and technology. Stars: A list of the best space movies. Tommy Lee Jones, Stars: | Zoe Saldana, | 44 min $426.83M, PG It's a haunting premise driven home by the loneliness and squalor of the space station in both films that orbits Solaris, suggesting that no matter how far we voyage into the cosmos, we'll never leave our psychological and emotional baggage behind. Oscar Isaac, 115 min 2001: A Space Odyssey is known for its hyper-realistic depiction of space, so it's hard to believe that this Stanley Kubrick-directed film was released back in 1968. The first Disney movie to be rated something other than "G" (it was given a "PG" by the MPAA ratings board), The Black Hole was the Mouse House's first conscious effort to break out of the "family film" box and create more adult fare. |
Ben Mendelsohn, It's all pure fantasy, of course, but it is our first cinematic voyage off this rock, and a movie we should all love dearly. | | Gross: Daisy Ridley,
The surviving members of the resistance face the First Order once again, and the legendary conflict between the Jedi and the Sith reaches its peak bringing the Skywalker saga to its end. The Most Memorable Quotes From 'Ad Astra'. After the devastating events of Avengers: Infinity War (2018), the universe is in ruins. Director: Stanislaw Lem's novel provided the basis for two theatrical film adaptations; the 1972 version directed by Andrei Tarkovsky remains a landmark of Russian cinema while the 2002 American take from Steven Soderbergh was modestly ambitious but now largely forgotten. Stars: Ron Howard Adventure, Sci-Fi. 137 min A team of explorers travel through a wormhole in space in an attempt to ensure humanity's survival. A look at the life of the astronaut, Neil Armstrong, and the legendary space mission that led him to become the first man to walk on the Moon on July 20, 1969.
96 min | | Conquest of Space (1955) Early space travel films tended naturally to stay local: the Moon and Mars were favored destinations. 151 min Dave Filoni, Action, Adventure, Sci-Fi. Taika Waititi Gary Oldman, On one hand, The Martian is primarily a story of survival and rescue, but it's also a celebration of the spirit of the space program and of the idea of space exploration itself.
141 min Josh Duhamel,
572,996 Michael Bay Space travel movies can be fantastical, but the top films about the galaxy feature realistic stories about astronauts. James Gunn |
The plot concerns a starship that accidentally opens a wormhole to another dimension which may or may not be hell -- a problem caused by the ship's gravity drive that they may want to fix next time out (not that there is one). Vote up the greatest realistic films about exploring the cosmos.
| Gross: Billy Dee Williams, Votes:
Director: | Director: 124,754 Vin Diesel, Francesca Annis, A box office bust when it first came out, this sci-fantasy from director Joe Dante (Gremlins) is a Spielberg-lite story of three young friends who are mysteriously compelled by the same dream to build a spacecraft and launch it, eventually getting into space and meeting the aliens who planted the dream in their heads in the first place. | $623.28M, PG-13
A team of explorers travel through a wormhole in space in an attempt to ensure humanity's survival.
| Following clues to the origin of mankind, a team finds a structure on a distant moon, but they soon realize they are not alone. | A secret military team, SG-1, is formed to explore other planets through the recently discovered Stargates. | eng | 52231a56-b3d5-42f6-a5ba-5f3a180b2323 | https://www.doctordanger.com/journal/bjuau0.php?tag=space-travel-movies-d88552 |
Otherwise . . .
So many creatures like to eat grasshoppers and crickets that these jumping insects have to lay lots of eggs and have lots of babies in order to keep their numbers strong.
The many predators, or natural enemies, of grasshoppers and crickets include spiders, hundreds of kinds, or species, of birds, snakes and even rodents such as mice and rats.
It is a good thing that so many creatures like to eat grasshoppers. If they didn't, grasshoppers would become so numerous they would eat much of the world's edible plant life, and people and other animals would starve.
What do grasshoppers eat? A grasshopper eats a variety of leafy plant material | eng | 17a61d94-81b9-4136-a008-a7d219adda43 | https://earthlife.net/what-eats-a-grasshopper/ |
A new study by Rand Corp. has found a much smaller number of veterans, compared to the general population, say they believe in white supremacy, the QAnon conspiracy theory and left-wing extremism. But nearly 18% of veterans did agree with…[Continue Reading...] | eng | 93ebf5a2-8e45-4ce7-92b8-8675dcb7953d | https://cultnews.net/tag/american-military-veterans/ |
Yo, Which Country Blazes the Most Weed? (Straight Up Exposed)
Yo yo yo, what's good, it's your boy Dan back at it again with some dope knowledge about the countries that love to puff puff pass the most. Now, when you think of places where weed is lit, you probably think of Amsterdam or even Colorado and Cali in the US. But just because a place has some chill cannabis laws doesn't mean everyone's getting lit. So which country tokes the most? Brace yourself, cause it ain't Amsterdam like we thought.
You know how everyone thinks Dutch people are lighting up spliffs all day every day? Turns out only about 8% of them actually smoke weed. So who's at the top of the list? Portugal? Uruguay? Nah fam, it's Iceland. That's right, according to the United Nations Office on Drugs and Crime (UNODC), about 18.3% of Icelanders ages 15-64 use marijuana. The US came in second with 16.2% and Nigeria was third with 14.3%. Surprising right?
What makes Iceland stand out even more is that weed isn't even legal for recreational use there. Y'all heard me right, it's straight up illegal. You can't grow it, sell it, possess it, or use it unless you're a chronically ill patient. But despite all that, the peeps in Iceland still love to light up. Reports say smoking in public or having a small amount usually just results in a fine and smoking weed is widely accepted socially.
But hold up, how do they get away with it if it's illegal? Well, the police can't just search anyone without reasonable suspicion so people can lowkey get away with it. The fine for getting caught with one gram of weed is around $500 but you gotta have at least half a kilo to get jail time for three months. Even though people still get arrested for it, there's definitely a culture of consuming and producing weed in Iceland.
Now you might be wondering why Iceland loves weed so much. Well, the answer isn't crystal clear but some say it's because they didn't have legal beer until 1989. Alcohol was completely prohibited there back in 1915 and although wine and liquor became legal, beer only got the green light around 30 years ago. So for those college kids looking to party back in the day, weed may have been a safer alternative to dodgy beer.
So what about the other countries on the list? The US is second with 16.2% of the population using marijuana. And even though it's still illegal on a federal level, 33 states and Washington DC have legalized it in some form or another. As people get more open-minded about cannabis, the culture is only gonna grow.
Nigeria's third on the list with an estimated 20.8 million users each year even though sales are illegal there too. Canada comes in fourth with 12.7% of the population admitting to using weed. And down in 20th place is Amsterdam, which proves that just because it's legal doesn't mean everyone's lighting up.
But let's be real fam, we all know that when you walk into a coffee shop (not your regular cafe) in Amsterdam, most of the customers are tourists anyway. Locals aren't really into that scene like that and you won't find many coffee shops where they're at either. So next time you think everyone's blazing up all over Amsterdam, remember that it's mostly just us outsiders enjoying what they got going on over there.
In conclusion, just cause a country has some chill cannabis laws doesn't mean everyone's lighting up all day every day. Iceland takes the cake when it comes to smoking the most weed even though it's illegal there. So if you're ever in Reykjavik, don't be surprised if you catch a whiff of that sticky icky in the air. | eng | 6a1fd0e1-f5f9-49a9-84c5-3ff7df4cdedb | https://roscommonacres.com/yo-which-country-blazes-the-most-weed-straight-up-exposed/ |
Recommended Posts
Am trying to change something in one of my sites. Since when clients sent me money there is a confirmation code which is given to them. I want them to be able to use this code to pay for their goods instead of the normal credit card and paypay payment systems. Is there a way to achieve this and by using which component. I am new to osCommerce. | eng | 29aefdc8-54e8-4a4b-95e2-042ad1f4505a | https://www.oscommerce.com/forums/topic/362104-payment-by-a-code-or-coupon/ |
Mosaic® Hops Pellets
Product details
A new dual-purpose hop, bred from Simcoe®, formerly known as HBC 369, this is the latest experimental variety to "go pro" and become a full commercial variety. Bred by the same people who brought you Citra®, Warrior®, Ahtanum, Palisade®, and Simcoe®.
Mosaic is so named because it gives a range of aromas and flavors which are complementary to other common hop aromas. Everything from citrus, pine, earth, herbal, mint, bubblegum, blueberry, lime peel, black pepper - the character seems to change depending upon the conditions it is in. | eng | 52d24060-85ed-43c5-8e11-3ce0bbb9b5be | https://www.northernbrewer.com/products/mosaic-hop-pellets |
How to Help Your Electronics Last Longer and Preserve Important Data
Table of Contents Show
We all know how important technology and electronics are in our daily lives. They play such an important role in everything that we do. We use electronics every day, so it's important to make sure they work right. The key is to come up with ways to make your electronics last longer, and preserve important data.
The good news is there are a fair few ways you can go about doing this. It's just a case of making the right decisions. Whether it's on a business or personal level, you've got to make sure your electronics work well. Have a look at these suggestions for helping them to last a little longer.
Treat Them With Respect
As with anything you own, it's important to treat things with respect. You need to understand how delicate and fragile electronics can be. So it's important to take the time to make sure you look after them. That means you have to handle with care and be sensible when you use your computer, etc. The more respect you show, the better you'll treat them. And this means they will last much longer than they would otherwise.
Clean Often
Of course, it might not always be possible to clean your electronics. But, where it is possible, you need to make sure you do so. There are certain specific ways of cleaning particular items. For instance, you can buy a special DVD for your DVD player that will help to clean the inside. You can also use a duster to keep your TV and computer dust free. But, you also have to remember that you can clean the system as well. So, for example, you might choose to run a system clean on your laptop. Or you could delete apps to clear some space on your smartphone. Either way, physical and virtual cleaning are important.
Make Sure You Use the Right Materials
It's vital that you make sure your electronics are well made. And this means checking that they have the right materials. For instance, a lot of micro-electronics make use of thermal material. This is essential for helping with the longevity and durability. So, you need to make sure that your electronics have the right sorts of materials. They need to be well made and have a good reputation. Make sure you do a bit of research before you buy so you know what you're buying.
Back Up Information
Okay, so, it may not necessarily help you to make your technology last longer, but it will help you preserve data. These days we are so reliant on technology, but there are still risks of problems with the machines. So, it's important to back up information as much as possible. This means that you'll always be able to access it even if you have problems with your machines. This is a vital part of modern technology, so make sure you take it seriously.
You already know how important electronics and technology are in your life. So you need to make sure you look after them as much as you can. Make use of the ideas on this post to get the best out of your technology. | eng | 3361e63a-1937-4b52-b2b3-5c4f4dd245dc | https://techdaring.com/how-to-help-your-electronics-last-longer-and-preserve-important-data/ |
Studying Online vs On Campus
Classroom versus online study
The benefits of online education
The topic of online education has never been discussed as much as it has this year, mainly because of the COVID pandemic. With campuses closed during lockdowns, many students had no option but to continue their studies online; and it seems many students are discovering that online education is just as good – if not better – than classroom education. There are several reasons why students prefer online courses, and it's important to note that online education is equivalent to on-campus education. Benefits of online education include:
No geographical restrictions
The internet gives us immediate access to the world, meaning students have options to study online through universities and institutions that they wouldn't be able to attend geographically. This is a great option for international students whose preferred school would otherwise be out of reach. It also means that students have more choice in terms of courses, not just universities. Through the offering of online education, students can search for the exact course that suits their needs, rather than settling for a course that is not ideal simply because they attend a particular campus.
Better engagement between educators and students
Online learning doesn't mean you are isolating yourself and having to self-teach. These days you can attend an online classroom through your computer's camera, and have face to face interaction with educators and other students. Technology means there is now plenty of opportunity for engagement while online learning, which enhances the learning experience.
Recent evidence has shown that many students find online classrooms far less intimidating than actual classrooms, and studying in the comfort of their own home provides a relaxing and non-confrontational environment for them to learn. For this reason, online learning is actually better for many students than classroom learning.
Faster responses from lecturers
Educators face the challenge of being time-poor and often unable to give one-on-one attention to each student. Online education is a time saver for educators as well, and students find they actually receive more attention from their educators through their online classes. Online students are generally asked to complete more-frequent assessments so that professors can monitor their competence, and educators are able to respond quickly due to the content being all submitted online. This is a huge benefit because learning gaps can be identified and addressed faster, and students progress through their courses more successfully.
Lower fees and related expenses
The biggest "plus" for online education is the cost. A traditional university will cost a student far more in fees than an online course due to the cost of attending a campus. Online education is far more affordable and therefore is more accessible for those wanting to study. Students will also save money in transport to and from campus, accommodation if they live far away, and general expenses while on campus.
Less debt through student loans
Many online programs have been structured for students who are already working, and who are choosing to study part-time. This means you can still have the best of both worlds. Remaining employed while you are completing your studies means you will be financially better off, and therefore may not need to take out a student loan, or at least you may have a smaller loan than a full-time student. Smaller student loans mean you'll be paying less interest, which is highly beneficial in the long run.
Flexibility in schedules
While some online programs may require students to be online at certain times and participate in classrooms; many do not. Generally, most classes are archived for future reference, so if you do miss the scheduled timeslot, you can catch up later. This allows for flexibility in your own day-to-day schedule – rather than having to be on campus to attend a class, you can take the class at a time that suits you online. This flexibility also allows students to work through the course material at their own pace, taking their time where they need to or moving ahead if that suits. Flexibility in schedules also benefits students who are studying around work commitments or children.
Working with modern technology
Technology is updated so frequently that sometimes it's hard to keep up. However, students undertaking online study will find they are adapting to changing technology as their courses require it. Video conferencing, messaging, sourcing information and presenting findings are all skills that online students are developing while they undertake their studies. The ability to work with technology, learn to use it and adapt to changes enables a student to be better equipped to enter the workforce. This is especially important when studying information technology, as the industry is constantly changing.
Online learning is getting a good reputation
Online education is being recognised the same as classroom education, particularly when the student has studied through an accredited institution. The myth that getting a degree online was somehow less important than a campus-based degree is no longer a factor for most employers. Online college options will continue their increase in popularity as costs get lower (thanks to better technology) and quality of instruction evolves and improves. The future will see more and more students opt for getting their degrees online, and it won't be too long before online studying is seen as the norm.
If you are considering entering into online education, contact our Student Services for information on course options.
TrainSmart Australia acknowledges and pays respect to the past, present and future Traditional Custodians and Elders of this nation and the continuation of cultural, spiritual and educational practices of Aboriginal and Torres Strait Islander peoples. | eng | ced4697d-088c-4bab-a644-1f7913d20826 | https://tsa.edu.au/studying-online-vs-on-campus/ |
Valthirian Arc: Hero School Story - a game that combines a huge number of different genres, which created a unique gaming experience. There are features taken from role-playing adventures, city-planning strategies in real time, as well as from action games and simulators of managing a particular community or business. Gamers will have to take the reins over the magical school of wizardry and bring it to a new level of development. The developers have added an extensive list of achievements, adding to the interest in researching the available mechanics.
Developing the infrastructure of his unusual educational institution, the gamer attracts new people who want to learn the basics of magic and become powerful wizards through long training. The player needs to organize the process of obtaining the fullest possible education for their own students. Due to the full control of the educational process, the user carefully monitors the students and corrects their classes.
The school is located on the territory of the world called Valtiria. The inhabitants of this area require help from the magicians, which is why there are tasks to which it is necessary to send wizards to gain experience and useful things presented as a gift from satisfied inhabitants. There are 5 kingdoms on the map with different rulers. The player will have to choose who to help and which side he takes. Elections affect the ending of story missions. To confront the bandits of Valtiria, gamers need to create weapons and armor that will be used by the sorcerers. | eng | 458fbb83-f084-4e73-af85-fc47e65be9d8 | https://utorrent9.org/1140-valthirian-arc-hero-school-story.html |
Specifications
Specifications include,
but are not limited to: Qualified vendors to
furnish all material, equipment, labor, and any
incidentals needed to complete the project. This
project will include:(1)
Mobilization/Demobilization.(2) Provide all
pumps/piping/fittings, etc. necessary to move
liquid material from the digester tanks to the
dewatering equipment (Belt press or Centrifuge)
provided by the Awarded Bidder.(3) Provide staff
to run a 12-hour per day, six day per week
operations (the operation days/hours are
estimated, and the contractor shall be flexible
to change based on the plant operations
needs).(4) Provide trailers and road tractors to
transport and dispose the material (dewatered
sludge) to an approved disposal site. The
Contractor is responsible for the tipping fee
and other related fees | eng | f7cb3fe0-7943-46cc-a18f-3ae903b394ad | https://www.bidnet.com/closed-government-contracts/wastewater-reclamation-facility-digester-cleaning?itemId=744078299 |
✅ 3 WAYS TO PROPERLY CHECK CONCRETE QUALITY
Proper concrete quality check is one of the topics that attracts contractors' attention, because the quality of concrete directly affects the project/ works. In fact, there are many methods, commonly used are concrete sampling and core drilling. Let's learn how to do this with Dufago:
Concrete Check Before Pouring
Before pouring, it is necessary to check whether the concrete satisfies the given requirements, it is necessary to follow the procedures:
Conduct concrete slump tests.
Conduct concrete cylinder testing to verify its quality.
Check the amount of water in mixed concrete to ensure its original quality.
Mix concrete well so that when discharging the concrete sample, the most accurate slump can be achieved.
Proceed to remove the mold, soak the specimen for a while for later quality evaluation.
Kiểm tra chất lượng bê tông trước khi đổ
Proceed To Conduct Compressive Strength Test on Concrete Specimens
After a 28-day curing period, the 15cm x 15cm x 15cm concrete cube specimen is placed in the machine to determine concrete compressive strength in units of MPa (N/mm²) or daN/cm² (kG/cm²).
Hình ảnh kiểm tra độ sụt bê tông
Steps of Conducting Core Drilling Tests
The core drilling is a method used by many people to check the quality of concrete. In order to complete the procedure, it is necessary to follow these steps:
Step 1: Locate the drilling core to be tested, usually customers will locate it beforehand.
Step 2: Use a core drill to extract a concrete specimen to be tested.
Step 3: Mark on the concrete specimens and flat cut 2 ends of the specimen along the mark with a cutter. | eng | 9e37ce9c-0496-4a7e-8097-2e4e98a316b7 | https://dufago.com.vn/en/check-the-quality-of-concrete-properly/ |
Do you have a valid medical malpractice claim?
Healthcare providers can be held responsible for the harm suffered by a patient due to medical negligence. Therefore, you may wonder if you have a valid claim after a medical procedure yielded unexpected results or if you experienced an adverse outcome after receiving treatment.
The answer may not be straightforward, since every situation is unique and different. That being said, there are a few key factors to consider when determining if you have a valid medical malpractice claim.
What are the elements of a medical malpractice claim?
First, there must have been a patient-doctor relationship. In other words, the healthcare provider must have been attending to you as their patient when the alleged malpractice occurred. A patient-doctor relationship is essential in medical malpractice claims since it places a duty of care on the healthcare provider.
Second, there must have been negligence or a breach of the duty of care owed to you. Would you have ended up in a similar situation in the hands of an equally qualified healthcare provider? Establishing negligence can be a complex determination and often requires expert testimony.
Third, the healthcare provider's negligence must have directly caused you harm. If a doctor made a simple error that did not affect you in any way, you may not have a case. However, you may have a valid malpractice claim if you suffered harm due to the doctor's mistake.
Get an informed assessment and guidance
Determining whether or not you have a valid medical malpractice claim can be difficult. As such, seeking qualified assistance to help evaluate your case and guide you through the legal process is advisable.
While it can be daunting to consider pursuing legal action as a victim of medical malpractice, it's important to hold healthcare providers accountable for their actions and ensure you receive the compensation you deserve. | eng | 388e153b-e5d0-4fab-8588-d34d4290018c | https://www.marshalllawnm.com/blog/2023/04/do-you-have-a-valid-medical-malpractice-claim/ |
Lizmap is designed to show spatial data in the main map, and you can propose users to see an object data through the "popup"
feature (a small popup containing the objects data is shown whenever the user clicks on the map ). See Popup
Sometimes this is not enough, and as a map editor, you would like the user to see all the data of a specific layer,
as you can do in QGIS by opening the attribute table.
Note
When activating the attibute table on a layer, it also activates graphical selection tools.
In the Tools tab of Lizmap plugin dialog, there is a group called "Attribute layers" which shows a table and some form fields dedicated to add or remove vector layers.
To add a layer to this tool:
Click the button.
Layer: Choose one of the vector layers (spatial or not). This can be any vector layer format : GeoJSON, Shapefile, PostGIS, CSV, etc.
Unique ID: The attribute table tool needs to be able to defined each feature as unique. We strongly advise you to add such a field if your layer has not one yet. Usually the unique ID field contains integers. If the layer do not have this kind of field, you can easily create it with the Field calculator. Choose the correct field with the combo box.
Fields to hide: You have 2 ways of hiding fields in the published attribute table.
In the vector layer properties dialog of the QGIS vector layer, in the Fields tab, you can uncheck the checkbox of the column WFS for the fields to unpublish. This means this fields will not be published via the WFS protocol. This is the simplest and safiest way to restrict the publication to some fields (for example to get rid of sensitive fields)
You can use this Fields to hide option to hide the given fields in the attribute table display. The hidden fields won't be visible for the end user, but will still be available for Lizmap Web Client. You must use this option to hide the Unique ID field. If you use the first way (uncheck WFS column), the unique ID won't be usable by Lizmap, and some of the attribute table features will not work properly.
Let the Pivot table unchecked. This is for relation.
Let the Hide in child subpanels unchecked. This is for relation.
Let the Hide layer in the list unchecked. This is for relation.
To edit a layer in the table, select it and click the button or double click on the row.
To remove a layer from the table, select it and click the button.
To move a layer up or down, select it and click or buttons. The order will change in Lizmap too.
By adding the layer in the attribute table tool, this will also enable the download capability on that layer.
By default, only GeoJSON and GML are available. By installing the WfsOutputExtension plugin on QGIS Server,
you can add more formats such as Geopackage, ODS, CSV, XLSX …. Read
In QGIS, you can configure relations between layers, in the project properties dialog. If you publish in Lizmap more than one layers in the attribute layers tool, and if some layers are part of a relation, the end user will be able to see child tables under the parent layer table, and a click on one line in the parent table will trigger the filter of the child tables content.
For example, you could have a layer of cities, and a child layer of public building. Clicking on one city in the attribute table will make the public building child table refresh its content with only the public buildings of the clicked city.
You can use "many-to-many" (N:M) relations in QGIS since version 2.14 but Lizmap doesn't use it automatically yet. You have to define the relations in the Lizmap plugin too. In many cases, N:M relations are very handy. For example, you can have the three following vector layers in your project:
Tramway lines: this layers name "Lines" contains one feature per tram line, and has a unique ID field tram_id
Tramway stops: this layer named "Stops" contains one feature per tram stop, with a unique ID field called stop_id
Correspondance table between lines and stops: this layer named "Pivot" is a pivot table between tram lines and stops, since a stop can be used for more than one line, and a line serves many stops. It has the following fields: tram_id, stop_id and order which defines the order of the stop in the line.
You can add 2 relations in QGIS project properties : one between Lines and Pivot using the tram_id field, and one between Stops and Pivot using the stop_id field.
In Lizmap, we added a simple way to configure the N:M relation. You can simply
Create the two relations described above in QGIS project properties dialog, tab Relations
Add the Lines and Stops layers in the attribute layers tool
Add the Pivot layer in the attribute layers tool with the option Pivot table checked
Lizmap Web Client will then handle the relation as a N:M relation:
The pivot table will be displayed under each parent attribute table and show only the corresponding children.
The filter feature based on the attribute layers will trigger the cascading filter of the pivot and the other parent. For example, if the user uses the filter to show only one tramway line, Lizmap will also only show the corresponding stops in the map and in the Stops attribute tables | eng | ef97b3a5-58cd-4673-b648-47c55b66988c | https://docs.lizmap.com/next/en/publish/lizmap_plugin/attribute_table.html |