title
stringlengths
6
301
keywords
listlengths
0
23
messages
listlengths
2
2
Food traceability technologies and foodborne outbreak occurrences
[ "Food supply chain", "Traceability technologies", "Foodborne outbreaks", "Health surveillance" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Foodborne diseases affect 1 in 10 people and kill around 420,000 people every year worldwide, with one-third of deaths occurring in children aged up to five years. These cases are consequences of the contamination of food by bacteria, viruses, parasites, chemicals and/or toxins (World Health Organization, 2015a). According to World Health Organization (2015b), when \"two or more people develop a similar illness after ingesting the same contaminated food or drink,\" the case is considered to be a foodborne outbreak (FBO). Hussain and Dawson (2013) stated that the United States Department of Agriculture reported losses of almost $7bn in FBO-related issues only in the early 2000s. Therefore, FBOs does not only affect public heath, but it can also have economic impacts. According to Center for Disease Control and Prevention (CDC, 2018a, b), this can be controlled and avoided through regular occurrence reports of FBOs, which help identify emerging threats from foodborne illnesses.\n2. Approach: To reach the proposed objective, a systematic literature review was used to select the FSCT material published in the literature. In the analytical phase of the material, the content analysis is used to classify the FSCT technologies and list the most cited one. Finally, the relationship between the frequency of publication and the occurrence of food outbreaks in the five countries studied is determined through a correlation analysis.\n3. FSCT technologies and tools: In the group of papers, 69 technologies and tools used in FSCT were found. Figure 2 shows the number of articles that cite each of them. To simplify, only those with ten or more citations were considered. It is observed that radiofrequency identification (RFID) technology is the most searched, followed by deoxyribonucleic acid (DNA) technology, wireless sensor network (WSN), hazard analysis and critical control points (HACCP) and Internet of Things (IoT). Therefore, the first five tools were considered the most relevant for FSCT and were the subject of an in-depth analysis in the research studies that followed.\n4. FBOs and FSCT correlation: There is a growing trend in the number of publications over the years, which correlates with the evolution of the FSCT technologies. Figure 6 shows the frequency of the 1,029 FSCT publications without restriction of technology in the period from 2000 to July 2018.\n5. Conclusion: The literature on FSCT has evolved over the last 18 years, especially with the emergence of intelligent technologies. However, FSC need ways to increase the security and transparency of traceability data and require further integration and standardization of systems for complex food export and distribution chains.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to identify the relationship between the frequency of publication on food supply chain (FSC) traceability and the occurrence of foodborne diseases outbreaks.\nDesign/methodology/approach: A systematic review of the literature was carried out to locate the main articles published in the literature, followed by a content analysis in order to list the main food traceability technologies and their evolutions. Finally, a Spearman's r correlation analysis between the frequency of publications on FSC traceability and the annual occurrence of foodborne outbreaks in the five largest food exporting countries in the world was performed.\nFindings: In these analyses, the tools of radiofrequency, deoxyribonucleic acid, wireless sensor network, hazard analysis and critical control points and Internet of Things are the most researched technologies, and they are relevant in the evolution of traceability in the FSC. With correlation coefficients above 0.700 at 0.01 significance levels, this evolution of food traceability technologies has been one of the factors reducing the number of food outbreaks in the USA and Germany, countries with greater development of the health system and food control.\nOriginality/value: This paper provides an evaluation of the food traceability technologies and the effects of their evolutions in the occurrence of food outbreaks. This may help in the proposal of public policies related to food and outbreak control.\n", "role": "assistant" } ]
Parallels and distinctions between health care quality improvement and health care disparities reduction
[ "Health services", "Quality improvement", "United States of America" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: For many years, the US health care system has focused on its failure to deliver appropriate care at the appropriate time to the appropriate patient with appropriate satisfaction. System approaches including total quality management (Crosby, 1979; Juran, 1986), continuous quality improvement (Deming, 1986; Donabedian, 1993), and various individual physician behavior change strategies (Bero et al., 1998; Grol, 1997; Smith, 2000, 2003), including the now widely touted concept of \"pay-for-performance\" (Pham et al., 2007), have all been proposed or implemented to improve care for all in the health care system. These approaches have met with some success. Still, the US health care system is being roundly criticized for its dysfunction (Poses, 2003; Poses, 2004; Kleinke, 2001; Ludmerer, 1999) And, sensing a loss of professionalism in physician behavior, specialty societies are newly challenging US physicians to aspire to high ethical and professional standards (Brennan et al., 2002)Nonetheless, health care quality improvement systems have largely been institutionalized and accepted. Most if not all hospital and health care systems now have quality monitoring and improvement teams, which produce an abundance of reports, consume the time of practitioners and administrators serving on various quality committees, and represent a good chunk of hospital budgets. Appropriately, nationally sponsored and promoted publications continue to focus attention on quality improvement, including the years of life lost due to failures of the health care system (Committee on Quality of Health Care in America, 2001).In contrast, the US health care system has only recently admitted its failure to deliver health care with equity, when the care of whites is compared with blacks, Hispanics, Native Americans, and other ethnic and racial minorities. For many years, disparate health outcomes of racial minorities were considered merely a product of factors outside the health care system: disparate access to care, incomes, housing, employment, and cultural practices and preferences. However, a plethora of stark reports in the literature (Harris et al., 1997; Peterson et al., 1997; Ayanian and Epstein, 1991; Schulman et al., 1999; Johnson et al., 1993; Todd et al., 1993; Bernabei et al., 1998; Todd et al., 2000; Bach et al., 1999; Ayanian et al., 1999a, b; Gornick et al., 1996), leading up to a compilation by the US Institute of Medicine, have highlighted the fact that racial and ethnic minorities - even with equivalent access to the health care system - receive lower quality care than white patients for many medical conditions (Smedley et al., 2002).Together, these have sparked a national debate and have heightened national attention on what the health care system can and should do to eliminate health disparities that are due to health care disparities (Lavizzo-Mourey and Lumpkin, 2004; King and Wheeler, 2004). Both the US Secretary of Health and Human Services and the Senate Majority Leader (Agency for Healthcare Research and Quality, 2004), as well as medical specialty societies (American College of Physicians, 2004) have each identified health care disparities as a national problem. A new textbook on the topic includes as an author the former US Surgeon General (Woolf et al., 2004). Further, a researcher calculated that 886,202 excess deaths were attributable to health disparities in African-Americans vs. whites from 1991 to 2000, while at the same time only 176,633 lives were saved by medical advances (Computer Retrieval Projects, n.d.). A search of the National Institutes of Health's Computer Retrieval of Information on Scientific Projects (CRISP, n.d.) database using the term \"disparities\" yielded 993 active grants funded by the National Institutes of Health, Agency for Healthcare Research and Quality, or Centers for Disease Control (US Census 2000, 2002)This attention could not come at a more notable time in US demographic history. Led by a few highly populous states, the US is becoming a nation dominated by populations who suffer from health disparities. Demographers predict that racial and ethnic minorities will compose 50 percent of the US population by the year 2050 (Donabedian, 1980). Hispanics already outnumber Blacks as the largest racial/ethnic minority in the USA, and as of the 2000 Census, fully 18 percent of US residents aged five and older spoke a language other than English at home.\nThe parallels: Health care disparities are themselves a subset of poor quality care. Health care disparities can be thought of as resulting from structures of care and processes of care that are somehow flawed, resulting in outcomes of care that are flawed. Since the quality of structures and of processes of care are viewed as the major influences on outcome quality (National Healthcare Quality Report, 2006), repairing structures and processes of care will be necessary to reduce health disparities.It follows that, since health care disparities (structures and processes) may result in health disparities (outcomes), they should be owned by all as part of the quality problem. Like overall quality improvement, reducing health care disparities requires the attention and the will of all providers of care, rather than a few. Similarly, like overall quality improvement solutions, solutions for health care disparities require the participation of all providers, not just those already sensitized to the problem, or those coming from a personal experience of disparities. Last, to ensure overall system quality improvement, the topic of health care disparities should be given special focus by all hospital and health care systems, not just those treating predominately disparity populations.Some leadership in this merger has already been undertaken by the US Agency for Healthcare Quality and Research (AHRQ). They have published, simultaneously and linked for several years, the National Healthcare Quality Report (NHQR) and the National Healthcare Disparities Report (NHDR). The 2006 NHQR is:[...] a comprehensive national overview of quality of health care in the United States. It is organized around four dimensions of quality of care - effectiveness, patient safety, timeliness, and patient centeredness. Those dimensions cover the four stages of care from the patient perspective (staying healthy, getting better, living with illness and disability, and coping with the end of life). The 2006 NHQR includes 211 performance measures that can be used to monitor the Nation's progress toward improved health care quality for all Americans. This year's report focuses on a group of 42 \"core\" measures representing the most important and scientifically sound measures of quality (National Healthcare Disparties Report, 2006).The 2006 NHDR:... tracks disparities in both quality of and access to health care in the United States for both the general population and for AHRQ's congressionally designated priority populations. The 2006 NHDR uses the same measures of quality as its companion National Healthcare Quality Report (NHQR) to monitor the Nation's annual progress toward eliminating disparities in health care. Racial/ethnic group comparisons focus on 22 core measures of quality and six core measures of access that support reliable estimates for Whites, Blacks, Asians, American Indians/Alaska Natives, and Hispanics; income group comparisons highlight 17 core quality measures and six core access measures (Betancourt et al., 2007).As can be seen from the above descriptions of NHQR and NHDR, the reports contain language linking health care disparities to health care quality. Both reports allow a \"drill-down\" feature meant for use by those seeking information specific to their localities. Users can also use the reports to monitor changes in specific performance measures over time in specific groups.However, to date, anecdotes suggest the new focus on health care disparities has not yet reached the level of institutionalization that health care quality improvement reached in the 1980s and 1990s. For example, a brand new Commonwealth Fund report indicates that many post-graduate medical residents in training feel unprepared to treat minority patients, based on surveys of 2,047 residents specializing in internal medicine, surgery, pediatrics, obstetrics/gynecology, emergency medicine, psychiatry and family medicine:While nearly all respondents thought it was important to consider patients' cultures when providing care, most felt unprepared to treat patients with certain cultural characteristics. Specifically, residents reported feeling unease when faced with patients who mistrust the US health care system, believe in practices at odds with Western medicine or have limited English skills ... . Some residents attributed their lack of preparedness to the dearth of instruction on cross-cultural care delivery and infrequent evaluation of residents' cultural competence (National Survey of Physicians, 2002).Further, US physicians surveyed in 2002 said that the health care system \"rarely\" (55 percent) or \"never\" (14 percent) treats people unfairly based on their race or ethnicity (Taylor et al., 2006). A similar survey of cardiovascular surgeons in 2004 revealed that 44 percent thought that, among patients with cardiac risk factors, black patients were not as likely as White patients to receive cardiac diagnostic tests and procedures. Additionally, 30 percent thought that black patients were not as likely as White patients to receive therapeutic tests and procedures. However:... only 13 percent agreed that cardiac care disparities occur 'often' or 'somewhat often' based on patients' race/ethnicity, independent of their insurance and education. Only 3 percent thought disparities were likely to occur in their clinical setting. Respondents appeared more likely to endorse patient factors (e.g., health behaviors or treatment adherence) than system or provider (e.g., miscommunication or continuity of care) factors as reasons for disparities (Holmboe and Hawkins, 1998)Thus, to overcome this lack of dissemination of appropriate messages regarding elimination or reduction of health care disparities, health system governance personnel must now create a monitoring, reporting, and committee structure for facing and improving of this aspect of quality. As mentioned earlier, a similar infrastructure has already been created for other aspects of quality, proving it can be done in willing health care systems. Similarly, individual practitioners must now consider whether their practice patterns are part of the problem of health care disparities, rather than part of the solution.\nThe distinctions: Despite the significant parallels between health care quality improvement and reduction of health care disparities in general, a few distinctions deserve mention, because they suggest unique interventions to reduce health care disparities.Attitudes\nConclusions and implications: In the USA, the cry to reduce racial and ethnic health disparities by improving the quality of care has risen to a perhaps fevered pitch. Meanwhile, quality improvement professionals face newly recognized and perhaps unexpected challenges to reducing disparities, namely the health care system itself and the attitudes of its workers. Interventions to combat these challenges, including attitudinal approaches, have begun, but are by no means in full swing. A quality improvement system exists within which to place efforts in this domain, but that system must be adapted. New quality improvement techniques, those that address attitudes and not merely knowledge and skills, must be implemented. Interventions must be customized for the specific targeted disparity problems and populations of a given health care system.The urgency to attack this problem could not be greater. The US population is becoming increasingly racially and ethnically diverse. Simultaneously, thousands of US lives are being lost due to racial and ethnic health inequities each year. Failure to make improvements in the quality domain of health care disparities could result in an increasing segment of the US population receiving substandard care, with continued substantial loss of life.\n", "role": "user" }, { "content": "Purpose: - This paper aims to compare and contrast quality improvement in the domain of health care disparities with quality improvement in other domains.\nDesign/methodology/approach: - The author provides a descriptive essay and review to put forward the findings of their research.\nFindings: - In the USA, health care quality improvement systems have largely been accepted and institutionalized. Most if not all hospital and health care systems now have quality monitoring and improvement teams. In contrast, despite a plethora of stark reports in the literature showing that the US health care system has failed to deliver health care with equity when the care of Whites is compared with that of racial and ethnic minorities, there is not a parallel health care disparities improvement system in most health care settings.\nPractical implications: - Paralleling many steps that have been taken to improve quality in general, health care workers and health systems must take steps to improve structures and processes of care to reduce health care disparities.\nOriginality/value: - Pinpoints some important distinctions between improving structures and processes of care related to health care disparities, and those related to other aspects of quality improvement. Doing so will save lives, and in the process improve overall quality.\n", "role": "assistant" } ]
Evaluating course design principles for multimedia learning materials
[ "Multimedia", "Electronic resources", "Armed forces", "Junior managers", "Career development" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: This paper reports on evaluation studies of principles of course design for interactive multimedia learning materials. The evaluation studies included interviews, observation studies and the use of online questionnaires. At the Defence Academy of the UK, Cranfield University has worked with military colleagues to produce multimedia learning materials for courses on \"Military Knowledge\" (MK). The courses are part of Officer Career Development (OCD) training for junior officers in the army. The courses constitute more than 200 hours of stand-alone learning materials. The courses were developed following clearly defined course design principles and a pedagogic model believed to best suited for the delivery of resource based learning materials. Key features of the model include: clearly stated aims and learning outcomes, knowledge maps to support flexible navigation and a wealth of activities to support effective learning. The lead author was responsible for formulating the course design principles and for specifying the pedagogic model. The principles draw on several sources in the literature for what is considered as good practice. The paper reports in-depth evaluation studies concerned with validating the pedagogic model and associated course design principles. The authors believe this is the first time that such a large-scale set of studies has been carried out. The overall findings suggest that, in general, the course design principles are indeed regarded by students as useful for effective learning. However, there are also interesting individual differences in the learning strategies adopted, with some students making more use of some of the course design features than other students. Some conclusions are drawn about what does indeed constitute good course design, together with suggestions for further research.Section 2 briefly sets the scene concerning the MK courses.Section 3 overviews the course design and development model and key features of the learning designs used for the MK.Section 4 describes the research methods used, with information about research designs, sampling and data gathering.Section 5 presents summaries of the main findings of the research. Section 6 has some concluding comments, together with descriptions about how the research programme, of which the reported studies are a part, is being taken forward at the Defence Academy - College of Management and Technology.\n2. About the MK courses: Cranfield University supports military colleagues at the UK Defence Academy in the delivery of a wide range of educational courses. We have been engaged in developing quality web-delivered distance learning courses for junior officers in the British Army. The courses are known as Military Knowledge 1 and Military Knowledge 2 (MK1 and MK2)[1] and provide in total some 200 hours of self-directed learning. The courses are self-contained and designed to be worked through by learners working individually. Informal support groups do exist and, as noted in the evaluation findings below, many students would welcome more opportunity to interact with peers and tutors. For more about the background to the MK courses, see Mackain-Bremner and Scott (2006) and Potts (2009).The MK courses are divided into parts, modules, sections and lessons. Each lesson is further divided into as many as five topics, which may in turn have subtopics. The MK courses are structured in such a way as to give total freedom to the student in terms of the order in which each lesson, and indeed each topic or subtopic is accessed. To this end, students are provided with a \"Knowledge Map\", a visual representation of the course with a mechanism for browsing and launching lessons. Students are permitted to browse lesson content or to go directly to any given topic within the lesson via the lesson's topic menu internal navigation. Students are also provided with a \"Lesson Map\" which allows them to navigate to a given topic or subtopic. The learning design used for lessons follows the principle that for each topic there is:* a clearly defined learning outcome;* expository text supported by multimedia resources (graphics, animations, video clips);* one or more interactive learning activities with formative feedback;* self-assessment activities with formative feedback; and* summaries that can be downloaded for review purposes.Summative assessments are delivered online using a dedicated assessment engine.\n3. The course design and development model: This section briefly describes a principled approach to course design for interactive multimedia learning environments (See Scott and Cong, 2006). The principled approach is based on conversation theory (CT), a theory of learning and teaching.CT originated in the 1960s from a cybernetics approach to explaining learning in both living organisms and machines (Pask and Scott, 1973; Pask, 1975, 1976; Scott, 2001). The fundamental idea of CT is that humans are dynamic self-organising systems with a need to learn and that learning is mediated by external and internal conversations about a topic or subject matter. In an external learning conversation very often one participant has the role of teacher and subject matter expert. The teacher aims to develop a learner's understandings. Part of this process requires the learner to externalise his/her understandings as explanations and models for purposes of formative assessment. In CT, this process is referred to as \"teachback\". When engaged in a productive learning conversation, the participants typically share access to external representations of the subject matter in the form of texts, diagrams, concept maps, models, examples and so on. Figure 1 shows the form of a learning conversation.Pask (1975) refers to this model as the 'skeleton of a conversation'. It shows a \"snapshot\" view of two participants (learner and teacher) in conversation about a topic. All such exchanges have, as a minimum, two logical levels (\"how\" and \"why\"). The \"how\" level is concerned with how to \"do\" a topic: how to recognise it, construct it, maintain it and so on; the \"why\" level is concerned with explaining or justifying what a topic means in terms of other topics. External representations mediate these conversations and assist the participants to negotiate agreements, including agreements to disagree. Apart from these external conversations, each participant has a continual internal conversation or \"inner dialogue\" where concepts are constructed and assessed for logical coherence and consistency of naming with other concepts within a participant's conceptual system. Pask (1975) also defines a generic term, \"modelling facility\", to describe the resources that enable the teacher to demonstrate and exemplify the topic. Here, the teacher can use non-verbal demonstrations. Modelling facilities can be laboratories, computer based micro-worlds and simulations or parts of the real world.To support conversational learning as characterised by CT, we believe a multimedia learning environment should have four major components, as exemplified in Pask and Scott's (1973) seminal Course Assembly System and Tutorial Environment (CASTE). The four components are:1. Learning outcomes that are clearly specified. These may be conceptual, procedural or attitudinal.2. Course content should be analysed in terms of knowledge and tasks to ensure logical coherence and consistency of terminology.3. Learning designs (also referred to as \"tutorial strategies\") need to be specified to ensure effective learning takes place.4. Assessment procedures need to be specified, both formative and, where relevant, summative. Formative assessments play a key role in eliciting \"teachback\" activities within a learning design.A key principle to be followed is that all items of each the four components should map onto corresponding items of the other components. This is illustrated in Figure 2.In similar spirit, John Biggs (1999) has coined the term \"constructive alignment\" for the idea that course components should work together coherently and consistently.A key contribution from CT is that learning designs should be viewed as ways of conversing with the learner to support effective learning. Rowntree (1990) refers to this approach to learning design as providing a \"tutorial in print\". The conversations that occur should encourage the learner to:* reflect on his/her current knowledge, skills and attitudes;* carry out tasks that ensure effective learning takes place; and* engage in \"teachback\" of knowledge, skills and attitudes and apply these to novel problems.These ideas are summarised in Figure 3.We will now look at the whole process of course design, development and delivery.To help ensure that the fundamental principles of course design are followed when designing, developing and delivering courses, a ten-step process model has been developed (Ryan et al., 2000; Scott, 2002). The steps in the model are summarised in Figure 4.Although the model follows a particular sequence to ensure that all the components are designed to fit with each other, it is understood that they interact with each other. This means that the decisions of later steps may influence the earlier steps that may then need to be revised. This is indicated by anti-clockwise arrows at the centre of the diagram in Figure 4.There are some features of this model of course design which contribute to and account for its effectiveness in the design of interactive multimedia learning environments. They are:* All components are designed to work together.* Components are analysed and developed in a planned sequence, although each is reviewed again as new components are specified.* The entire design process is orderly but flexible. There is both \"feedback\" and \"feed forward\" in iterative cycles of work.* The key step of knowledge and task analysis is supported in CT by a well-defined methodology (Scott and Cong, 2007).* The design process includes formative evaluation using pilot studies and technical testing.Summative evaluation of a course once it has been delivered is seen as a key step in the design, development and delivery process. This step is one which, in practice, often skimped on or even omitted. Summative evaluation may use summative assessments to provide data on how well the course as a whole is doing its job of helping learners achieve learning outcomes. As below, it may also look to see how relevant and effective are the various components built into the course design.\n4. Research methods: To elicit in-depth information about students' perceptions of the course design methods, semi-standardised, open-ended oral interviews were conducted, together with deployment of an online questionnaire. Four students thus far have served as interviewees. Each student was interviewed by the researchers for about one and half hours. After the student interview, the dialogues between students and researcher were transcribed into texts. The \"constant comparative method\" of grounded theory (Bogdan and Biklen, 1992) was then used to analyse all the written scripts and recorded transcripts in order to ascertain the similarities and differences in perceptions among these participants.The students' responses were useful in developing an online questionnaire. The questionnaire contains Likert scale questions and four open-ended short essay questions. Table I shows the major components within the framework of questions used in the interviews and questionnaire.The survey lasted for four month starting from October 2008 to February 2009. Data was collected from 162 respondents.\n5. Summaries of the main findings of the research: For each question, the respondents indicated their level of opinion using four-point Likert scales. Four points were chosen, rather than three or five, in order to \"force\" a choice so it would be clearly positive or negative. The means of the responses to each question were calculated. The results are presented next with brief commentaries and with samples of the comments made by respondents as free text entries. Here we present data per question only. Further analysis, not yet carried out, will look for correlations amongst the responses, for example, to see if a student's preferred way of working correlates with their use of the several available navigation aids.5.1. Respondents' perceptions of expected study time, introductions and desired learning outcomes\n6. Concluding comments: This paper has reviewed a principled approach to the design of multimedia interactive learning materials based on conversation theory and deployed in the form of a ten-step model for course design, development and delivery. A case study was described using courses on Military Knowledge as the context for ongoing detailed evaluation studies aimed at validating the principled approach to course design. We are not aware of any other large scale evaluations of course design practice and so consider our work to be a valuable contribution to the research literature.With respect to the evaluation of the course design model, the questionnaire data show that a large majority of respondents consider the courses to have been well designed. By way of summary, we have carried out a SWOTS analysis to illustrate and interpret the data (see Table VII). A SWOT analysis is a common method of strategic analysis for strengths (S), weaknesses (W), opportunities (O), and threats (T). We have added a fifth component of \"strategies\" (S) to capture suggestions for how the MK courses could be improved.Data from the evaluation studies show that:1. Overall, learners find the course design satisfactory and do make use of the many features built into the course with the aim of making it pedagogically effective. More specifically, the learning design features employed (statements of learning outcomes, use of multimedia assets and interactive activities, use of self-assessment questions, use of a course knowledge map and lesson maps and so on) are all seen to be working together constructively and effectively.2. There are many, still to be explored, interesting individual differences amongst the learner population.3. Many learners' would value more opportunities to work collaboratively with peers.We are feeding back our findings to our MOD colleagues with the view to discussing how the course might be improved. Topics being addressed include:* how to ensure all learners make best advantage of the course features;* how to implement and manage elements of collaborative learning;* how to best support a range of different learning styles; and* how to investigate a possible role for adaptive teaching where access to lessons and topics within lessons is made contingent on successful completion of lessons and topics at subordinate levels within the Knowledge Map and Lesson Maps.\n", "role": "user" }, { "content": "Purpose: - This paper aims to report on evaluation studies of principles of course design for interactive multimedia learning materials.\nDesign/methodology/approach: - At the Defence Academy of the UK, Cranfield University has worked with military colleagues to produce multimedia learning materials for courses on \"Military Knowledge\". The courses are part of Officer Career Development training for junior officers in the army. The courses were developed following clearly defined course design principles believed to be best suited for the delivery of resource-based learning materials. Evaluation studies included focus groups, interviews, questionnaires and observation studies.\nFindings: - The overall findings suggest that the course design principles are indeed regarded by students as supportive of effective learning.\nOriginality/value: - The findings of this research are being fed back to MOD colleagues with the view to discussing how the course might be improved.\n", "role": "assistant" } ]
Exploring dualities of service innovation: implications for service research
[ "New service development", "Service innovation", "Failure" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Through a process of creative destruction, innovation can explain and provide a way forward for economic growth (Schumpeter, 1934). Service innovation is not only dynamic, evolutionary, non-linear and interactive but also risky - and without risk, no sustainable business can be created. Recent studies highlighted that service innovation occurs in service systems that intersect and overlap at individual (micro), organization (meso) and societal (macro) levels (Chandler and Vargo, 2011). Perhaps, most important, service innovation emphasizes the value-creating processes that occur for the benefitting party. Service innovation creates new opportunities and influences almost everything in our daily lives, but it also causes problems for individuals, organizations and society.\nWhat is service innovation?: Service innovation has become not only a buzzword but also a key theoretical concept. Jon Sundbo, Ian Miles and Faiz Gallouj sparked the research interest on service innovation and have continued to develop our understanding of the theoretical concept. Early on, Gallouj and Weinstein (1997) defined service innovation as \"Any change affecting one or more terms of one or more vectors of characteristics\". From an often technical- or characteristic-based view of service innovation, it now has been complemented with broader perspectives (Coombs and Miles, 2000). Lusch and Nambisan (2015) suggest that service innovation is \"rebundling of diverse resources that create novel resources that are beneficial [...] to some actors in a given context\". Breaking free from the provider-customer dichotomy enables theories on service innovation to cover new types of services such as social networks (e.g. Twitter) and public services (Khan academy) that, by a Schumpeterian view, are not considered as innovations.\nIntroducing level and outcome to understand service innovation: Service innovation can be understood at the individual, organization and societal levels. Newness can be understood as \"new for the individual\", \"new to the firm\" or \"new to the world\" (Alam, 2012). This understanding is consistent with viewing service innovation at the micro, meso or macro levels (Vargo and Lusch, 2014) and gives different perspectives or meanings to the concept. From this perspective, the need emerges for a more in-depth investigation of the effects that an innovation may have not only on the individual or the organization but also on society.\nExploring dualities of service innovation: To move beyond the traditional view of service innovation and study the topic on different levels of abstraction, we can use the two dimensions introduced in this essay,\nReflections: Service research should find new ways to build knowledge regarding service innovation. Theory development must include general, mid-range and theories currently in use (Brodie and Storbacka, 2014). Theory development in service research appears on several levels, but the translation process in both directions is failing, which limits the results to either a general theory that does not work in practice or a theory that is far too context-specific. Perhaps if the dualities of service innovation were more often considered, we could move service innovation research forward.\n", "role": "user" }, { "content": "Purpose: The purpose of this study is to explore three paradoxes of service innovation and provide a way forward for fresh thinking on the topic.\nDesign/methodology/approach: Through a conceptual model of service innovation research, the authors challenge the \"pro-change\" bias and explore what can be learnt from the duality of service innovation.\nFindings: This paper suggests that research moves beyond a firm perspective to study service innovation on multiple levels of abstraction. A conceptual model based on two dimensions, level (individual, organization and society) and outcome (success, failure), is used to pinpoint and explore three dualities of service innovation: adopt-reject, change-static and good-bad.\nOriginality/value: By challenging the traditional perspective on service innovation, the authors present new avenues for fresh thinking in research on service innovation. In this paper, the authors encourage researchers and managers to learn from failures and to acknowledge the negative effects of service innovation.\n", "role": "assistant" } ]
Institutional investors and voluntary disclosure and transparency: the case of Egypt
[ "Egypt", "Corporate governance", "Institutional investors", "Agency theory", "Disclosure checklist", "Voluntary disclosure and transparency" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Corporate governance is an important element in enhancing both economic efficiency and investor confidence. Among the main principles of corporate governance are disclosure and transparency (OECD, 2004). Improvements of disclosure and transparency are among the main corporate governance priorities in emerging markets. This is because emerging markets in general are known for high information asymmetry (Uyar and Kilic, 2012). For instance, in 2001, the World Bank reviewed corporate governance practices in Egypt using the Organisation for Economic Co-operation and Development (OECD) guidelines as a template and revealed that the best scores achieved were in relation to stakeholders' role in corporate governance, and the worst scores achieved were in relation to disclosure and transparency category (McGee, 2009). The OECD (2004) guidelines consist of the following corporate governance principles: ensuring the basis for an effective corporate governance framework; the rights of shareholders and key ownership functions; the equitable treatment of shareholders; the role of stakeholders in corporate governance; disclosure and transparency; and the responsibilities of the board.\nBackground context on Egypt: Egypt has gone through many phases of economic reform in the past years. Starting from President Anwar El Sadat's era, an Open Door Policy was adopted to encourage a greater participation of the private sector. By the 1990s, the share of the public sector declined to about 68 per cent (Salah, 2013). But up until today, the size of the public sector still signifies a substantial share of the Egyptian economy.\nTheoretical framework: Among the main theories that are used in explaining disclosure practices are agency theory (Mallin, 2007; Rouf, 2011; Uyar et al., 2013) and institutional theory (Deegan, 2006). Agency theory highlights the relationship between two different parties, usually a principal and an agent. This relationship results in what is known as the principal-agent problem. This principal-agent problem results from information asymmetry, where both parties have access to different levels of information, leading to agency conflicts (Ross, 1973; Mallin, 2007). The characteristics of corporate governance have been perceived as proxies for independence and as ways to minimize agency conflicts (Alhazaimeh et al., 2014). Agency theory highlights the issue of ownership structure, and how institutional ownership can be used as a monitoring device (Bowden, 2000) to mitigate the agency problem by enhancing levels of corporate voluntary disclosure, and hence transparency.\nLiterature review and development of hypotheses: In this study, and following Shin-Ping and Tsung-Hsien (2009); Azzam (2010) and Jara-Bertin et al. (2012), institutional investors are defined as a long series of owners such as banks, investment companies, mutual funds, insurance companies, governmental agencies, incorporated and holding companies and other institutional investors. They basically consist of any type of ownership other than individual ownership and free float.\nMethodology: Data\nEmpirical results and discussion: Descriptive statistics and bivariate analysis\nConclusion: This study has examined the impact of different types of institutional ownership on corporate voluntary disclosure and transparency and is the first to have explored the impact of six categories of ownership on levels of voluntary disclosure and transparency in Egypt. This makes this study unique. The study controlled for the impact of four important firm characteristics: size, leverage, profitability and age.\n", "role": "user" }, { "content": "Purpose: Disclosure and transparency are major pillars of corporate governance which need to be greatly promoted in Egypt. This research aims to understand how different kinds of institutional investors affect levels of voluntary disclosure and transparency.\nDesign/methodology/approach: The research was conducted on the most active Egyptian companies over a period of five years. A voluntary disclosure checklist was first developed to assess levels of voluntary disclosure and transparency.\nFindings: Empirical results support significant positive impacts of both bank ownership and foreign ownership on voluntary disclosure and transparency. Among the four firm characteristics controlled for in the research, firm size was the only one with a highly significant positive impact on voluntary disclosure and transparency.\nResearch limitations/implications: The results of this research may not be generalized to all companies, as it was only conducted on the most active firms on the Egyptian Exchange. Therefore, it is recommended that future researches integrate a more diversified sample.\nPractical implications: The research provides empirical evidence that institutional investors are not a homogeneous group and that different kinds of institutional ownership impact differently on voluntary disclosure and transparency. As such, some institutional investors are more influential than others when it comes to increasing corporate voluntary disclosure and transparency and in reducing agency problems.\nOriginality/value: This research offers assistance to policy makers interested in enhancing corporate disclosure and transparency. It is particularly important during any adjustment to ownership policies in Egypt.\n", "role": "assistant" } ]
Reflective vicarious learning (RVL) as an enhancement for action learning
[ "Learning", "Education", "Action learning", "General management", "Executive training", "Vicarious learning", "Goldfish bowl exercise" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The fundamental assumption within the action learning is that \"there is no learning without action and no action without learning\" Revans (2011, p. 85). The premise is that learning can only take place through participation, reflection and action (McGill and Beaty, 1992). Revans maintains that individuals will fail to understand the \"how\" of an experience until they have experienced it for themselves, i.e. \"learned by doing\". Aristotle's wisdom that \"For the things we have to learn before we do them, we learn by doing them\", written around 350 BCE, in the Nichomachean Ethics (Broadie, 1991, p. 104) is both obvious and common sense, but is largely ignored within conventional learning and education generally. Arguably, \"learned or learning by doing\" is a natural activity, part of growing up during which we learn through both a mixture of simulation and emulation and the subsequent reflections upon the process (Felicia, 2011). There is, however, a seeming disparity between Revans' concept of \"learning by doing\", (in which he emphasised that the individual should actually engage in action themselves in order to learn from their own experiences), and one particular aspect of social learning theory - vicarious learning or learning from the behaviour of others. The concept derives from the work of Bandura (1977) and has been defined as: \"a process that allows an observer of another individual's model behaviour to change his (sic) behaviour without directly experiencing the consequences\" (Voit and Drury, 2006, p. 1054).\nDiscussion: It is well understood that declaratory knowledge, which is largely in the cognitive domain, builds on what has gone before (on the shoulders of giants) but, what about learning in other domains, for example: affective and behavioural? Anecdotally, we are certainly less willing here to use the experience of others to influence others. How many teenagers take any notice of their parents, for example? How often are societies destined to repeat the mistakes of their predecessors? Why are members of action learning sets able to learn to modify their emotions and behaviours?\nConclusion: As stated in the introduction, there is no wish to contradict Revans' views, but simply to add to his work by exploring how vicarious learning experiences can enrich the work in this field. The paper simply seeks to provide a valuable insight into how vicarious learning can be beneficial in situations when it is not possible to actually participate in the experience itself.\n", "role": "user" }, { "content": "Purpose: A fundamental assumption within action learning is that learning only occurs through participation, reflection and action. Revans maintains that individuals will fail to understand the \"how\" of an experience until they have experienced it for themselves, i.e. \"learned by doing\". The purpose of this paper is to postulate that Revans' second phase of action learning, the reflection phase can, in situations where participation is not possible, be enhanced through reflective vicarious learning (RVL) or learning from the behaviour of others.\nDesign/methodology/approach: This paper adopts a desk research approach review of the literature.\nFindings: The authors maintain that Revans' concept of \"learning by doing\" in the context of the goldfish bowl exercise can enhance an individual's insight through RVL or learning from the behaviour of others.\nResearch limitations/implications: The paper is limited in some respects as it focusses on the viewpoint of the author coupled with the literature. Future research could explore participant voices to add an extra dimension to the work.\nPractical implications: In terms of utility for others, this paper is useful for developing an understanding of the differing learning opportunities that RVL and action learning combined can offer. As such, it has meaning for action learning facilitators, set members, academics and educational consultants.\nOriginality/value: This papers originality is that it seeks to enhance Revans' proposition by illustrating how RVL in the second phase of action learning can enhance an individual's learning in situations where participation is not possible.\n", "role": "assistant" } ]
Exploring current enablers and barriers for sustainable proximity manufacturing
[ "Sustainability", "Reshoring", "Supplier selection", "Sustainable supply chain", "Local sourcing", "Manufacturing location decision" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Over the past few decades, companies have moved manufacturing from local high-cost countries to foreign low-cost countries, giving rise to the offshoring phenomenon. However, there seems to be a trend of bringing foreign manufacturing back home, known as back-shoring, or reshoring (Fratocchi et al., 2016), due to operations and control problems at foreign locations (Kinkel and Maloca, 2009; Gray et al., 2017). Reshoring is particularly common in the USA and Europe (Fratocchi et al., 2016). Companies reshoring manufacturing mainly have characteristics suitable for manufacturing in high-cost countries, e.g. with high technology intensity and automation (Arlbjorn and Mikkelsen, 2014). Nevertheless, some companies manufacture their labour-intensive products, such as garments, at their home high-cost countries in the USA and Europe in order to gain benefits from proximity to their customers and headquarters (Forney et al., 1990; Bolisani and Scarso, 1996; Macchion et al., 2015). Labour-intensive products are sometimes manufactured in neighbouring lower-cost countries in a practice known as nearshoring in order to gain both lower labour costs and proximity to their customers and headquarters. (Tate et al., 2014; Sirilertsuwan et al., 2018). In this paper, proximity manufacturing includes domestic or local manufacturing, nearshoring and reshoring because they provide companies benefits of proximity to their customers and headquarters. Moreover, proximity manufacturing includes both a company's own manufacturing and outsourcing manufacturing provided by suppliers.\n2. Literature review: 2.1 Background on the TBL concept\n3. Methodology: The paper is exploratory in nature and primarily based on interviews with key managers from 12 different clothing companies. Interviewing key managers gains insight into companies' own experience and reasoning regarding the positive and negative factors that are enablers and barriers of proximity manufacturing under TBL, thus, revealing the reasons they choose proximity manufacturing. This paper adopts a qualitative approach based on interviews (Miles and Huberman, 1994; Yin, 2003) and uses RQs, the TBL concept and existing literature of proximity manufacturing towards TBL sustainability as guidelines for collecting data, creating an initial coding scheme, coding data and analysing data. TBL dimensions and four business aspects proposed by Sirilertsuwan et al. (2018) were used as the predetermined categories in the initial coding scheme. The four business aspects are profits (cost-price-sales), service and delivery, product or process innovation and development and product quality.\n4. Results from within and cross-case analysis: 4.1 Enablers\n5. Discussion: The findings are consistent with existing literature on proximity manufacturing in that most enablers and barriers relate to the business dimension, especially the profit, as well as service and delivery, aspects, rather than the environmental and social dimensions (Sirilertsuwan et al., 2018). The most-mentioned enablers in Table VI that are in common with prior research (Table I) are short lead-time, fast replenishment, short transportation times, meeting product specifications and accessing high product quality. Moreover, the findings show that a relative reduction in transportation costs, production problems and risks from currency fluctuations and delays possibly leads companies to have reshoring to be in proximity to their main markets and headquarters. An interview study on other industries by Hofmann and Gregory (2018) reveals the same benefits of proximity-to-market and proximity-to-headquarter manufacturing as this paper in terms of innovation, rapid prototyping and fitting, customization and timely delivery for high service level to customers.\n6. Conclusions: This paper investigates factors under TBL enabling and hindering labour-intensive companies in high-cost countries to pursue proximity manufacturing, based on within-case and cross-case analysis of data gathered through 12 interviews of different Swedish clothing companies. Common enablers of proximity manufacturing among the companies are supplier specialisation, short lead time and fast replenishment. Common barriers are expensive European production costs and lack of industrial structure and seamstresses in Europe.\n7. Implications, limitations and future research: It is important for all supply chain management practitioners and scholars to take into account that short distance from proximity manufacturing benefits not only time-to-market, fast replenishment, decreased transportation cost and time and low carbon emissions but also ease of visiting proximity suppliers. The ease of visiting suppliers for fast sample development and controlling the whole supply chain ensures on-time delivery, high-quality and non-harmful products, as well as good factory compliances leading to TBL enhancement.\n", "role": "user" }, { "content": "Purpose: Amidst offshoring and reshoring trends, the purpose of this paper is to explore why business practitioners, especially from the labour-intensive clothing industry, choose to manufacture some products in proximity to the high-cost European market. Moreover, the rise of sustainability concerns led us to further explore whether these reasons relate to triple bottom line (TBL): business, environment and society.\nDesign/methodology/approach: The content analysis was adopted for within-case and cross-case analysis of data from semi-structured interviews of managers from 12 clothing companies.\nFindings: Within-case analysis showed enablers and barriers (factors) of proximity manufacturing within each company's characteristics under TBL. Cross-case analysis showed the most-mentioned enablers (high-quality suppliers, short lead-time and fast replenishment) and barriers (expensive production cost and lack of industrial set-up and seamstresses). The findings revealed both common and different factors from existing studies.\nResearch limitations/implications: Besides being motives for companies to bring manufacturing back to Europe, the results can be used by researchers and companies to develop criteria and performance measures of manufacturing locations for enhancing the TBL sustainability. Future research may explore different locations and industries for possibilities of proximity-manufacturing generalisation.\nSocial implications: Findings show that governments could focus on eliminating barriers of proximity manufacturing and creating favourable institutional infrastructure for the European clothing industry and sustainability.\nOriginality/value: This paper highlights updated proximity-manufacturing factors from practices in relation to TBL sustainability, including support for proximity manufacturing as a practice for TBL enhancement.\n", "role": "assistant" } ]
An investigation into manufacturing performance achievements accrued by Indian manufacturing organization through strategic 5S practices
[ "Manufacturing industries", "5S implementation", "Manufacturing performance achievements" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The manufacturing organizations are working under extreme pressures because of high competitive manufacturing environment in the national and international market. The extreme competitive pressures on the manufacturing organizations are build up as result of globalization, high customer expectations, intense competition, international trade, deregulation of markets and evolution of innovative management practices (Singh and Ahuja, 2012; Randhawa and Ahuja, 2017a). The global scenario wants that manufacturers acquire best capabilities in terms of quality, production, delivery, customer demands and cost optimizations. Organizations need to improve the quality of their products, services and meet the ever increasing customer demands. The proficient and viable administration of resources and operations has been envisioned as a basic component of manufacturing industries, to ensure competitiveness of organizations (Mathur et al., 2012; Singh and Ahuja, 2014). Thus for realizing overall organizational profitability, manufacturing organizations have been consistently focusing upon productivity and quality enhancement for achieving their sustainability endeavors (Singh and Ahuja, 2017).\n2. Literature review: Japan was the birth place where 5S had been first utilized as a part of the manufacturing division in the mid-1950s Toyota Production System (TPS) was a pioneer in development of 5S principles all through this period, from its beginning as 3S (seiri, seiton and seiso) to its advancement into today's 5S (seiri, seiton, seiso, seiketsu and shitsuke) (Ohno, 1988; Sawada, 1995; Gapp et al., 2008). 5S is the significant element of the DNA of TPS for establishing plant visual control and automation (JIDOKA). 5S principles have been implemented in Japanese organizations for many years and considered as integral part of other Japanese approaches (Ohno, 1988; Ho and Cicmil, 1996). Takashi Osada (1989, 1991), the creator who built up the framework for applying 5S in the organizations in the mid-1980s, expressed that with a specific end goal to comprehend 5S, it is important to see how this procedure is incorporated into individual's daily routines. While Hirano (1995, 1996), proposed another framework of 5S in practical term, he considered 5S as best industrial formula to compete their other industrial competitors in the manufacturing sector. Traditionally 5S is based on the both socio-historical and philosophical roots of original Japanese concept. The essential elements for the starting of operation in TPS are to establish the first S- Seiri and S-Seiton, which are the base for the elimination of muda- waste in any organization (Monden, 1998). As indicated by the review directed in 1999 by the Japan Industrial Safety and Health Association, the following two Ss - seiso (cleaning) and seiketsu (standardize) - were evolved later on, during 1956 to 1972. The term shitsuke (discipline) was incorporated into the 5S umbrella during 1990s (Japan Industrial Safety and Health Association, 1999). According to Nishibori (1985), this S is the most critical component of all, as it endeavors to change the proactive behavior of the employees.\n3. Research methodology: The present study is aimed at conducting a detailed study of literature about the 5S, and also elaborates upon the an exhaustive case study on 5S implementation in an Indian automotive parts manufacturing organization to investigate the quantitative and qualitative benefits accrued by the 5S implementation program. The case study has been undertaken in a leading and rapidly growing manufacturing organization dealing with manufacturing of automotive suspension and fastening components for the various types of vehicles such as commercial vehicles, utility vehicles, multi axle vehicles and special purpose vehicles. The Company provides manufacturing facilities in major industrial area of Northern and Eastern India. The company has various manufacturing facilities in forging, drawing, coating, heat treatment, surface treatment and testing facilities. The hardcore market competition in the last two decades had been putting consistent pressures on the organization to maintain its competitive position in the market by achieving/maintaining highest quality levels acceptable to valued customers through stringent control on repeatability of manufacturing processes.\n4. 5S implementation methodology: The organization adopted following steps for 5S implementation in industry.\n5. Evaluation of 5S audits in various zones of organization: The purpose of this procedure was to carry out review of 5S principles implementation on a continuous basis in the factory location. The 5S facilitator in consultation with process owner, workplace committee members and concerned executives ensured the implementation of 5S principles in the factory location. Basic in-house training/refresher training and introduction to model companies has been carrying out for successful implementation of 5S practices on a continual basis. Based on audits, the points have been assigned to each and every section of 5S and the rating of the each section has been decided accordingly. Records of 5S implementation program have been maintained by following steps:\n6. Achievements accrued through strategic 5S implementation: The organization-wide implementation of 5S program was divided into the different zones areas that were effectively operated/managed under the supervision of zone leaders. The outcomes in regards to different parameters were appropriately investigated periodically at the end of every month. The 5S program was continued for a year and various manufacturing performance achievements accrued through 5S implementation program were monitored, recorded and analyzed to investigate realization of coveted quantitative and qualitative organizational goals and objectives, which were set in regards to the melioration of quality, safety, delivery, production, employee's morale values and cost optimization.\n7. Conclusions: The study clearly depicts that 5S program has facilitated manufacturing organization to achieve considerable improvements in the plant performance as compared to its previous condition. The practice of 5S technique started yielding improvements in the early phases of implementation, and once it was successfully implemented at the plant, it led to significant improvement in the working environment of manufacturing organization. 5S technique has been practiced for approximately one year which has resulted into highly improvement of quantitative and qualitative benefits of the industry.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to evaluate the quantitative and qualitative benefits accrued by an Indian automotive parts industry through strategic 5S implementation initiatives.\nDesign/methodology/approach: The study involves evaluation of steps involved in systematic implementation of 5S program methodology in an automotive manufacturing organization and investigation of achievements accrued by the industry through the successful implementation of 5S program.\nFindings: The empirical results of the study have revealed that effective practice of 5S program brings considerable level of improvements in the quality, production, cost optimizations, employee's morale values and work culture in the manufacturing industry. The industry accrued both tangible and non-tangible benefits through the holistic adoption of 5S principals. 5S principals have been envisioned to further support other quality improvement programs like lean manufacturing initiatives of the organizations.\nResearch limitations/implications: The limitation of the study is that this research has been carried out in only manufacturing industry while similar study will be conducted in the service industry also.\nOriginality/value: Global competition in the manufacturing sector has provided necessary impetus for manufacturing organizations for affecting continuous improvements in manufacturing performance for achieving sustainability and profitability in the competitive market. 5S implementation is the fundamental tool for the overall achievements in both quantitative and qualitative performance enhancements in the manufacturing as well as service organizations.\n", "role": "assistant" } ]
Is learning anytime, anywhere a good strategy for success? Identifying successful spatial-temporal patterns of on-the-job and full-time students
[ "Online learning", "Student performance", "Entropy analysis", "Full-time students", "On-the-job students", "Spatial-temporal patterns" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: With the rapid development of information and communication technology, online learning has flourished in recent years due to the flexibility of engaging anytime and anywhere. Sixty-five per cent of higher education institutions in the USA have claimed that online learning is critical to their long-term strategy (Allen and Seaman, 2013). One report indicated that more than 6.3 million American students took at least one online course in fall 2016 (Seaman et al., 2018). However, due to the lack of face-to-face interactions, online learning is significantly different from traditional in-classroom learning. Online education often requires specific management strategies and instructional designs. Therefore, it is crucial for higher education institutions to understand online learners' behavioral patterns to design and implement better online education programs.\nLiterature review: The benefits of anytime and anywhere learning\nResearch method: Data collection and pre-processing\nResults: Learner segmentations\nDiscussion: On-the-job learners have higher anytime and anywhere tendencies\nConclusion and future research: This study proposed the concepts of time and location entropy to examine students' spatial-temporal patterns and the relationship between the identified patterns and learning performance in two different datasets. In both sets of full-time and on-the-job students, the analytic results show that a higher level of learning anytime usually represents a higher engagement level and had a better chance of higher performance. On the other hand, the level of learning anywhere has a negative relationship with learning performance. Gender is not a consistent finding across full-time and on-the-job students.\n", "role": "user" }, { "content": "Purpose: Online learning is well-known by its flexibility of learning anytime and anywhere. However, how behavioral patterns tied to learning anytime and anywhere influence learning outcomes are still unknown.\nDesign/methodology/approach: This study proposed concepts of time and location entropy to depict students' spatial-temporal patterns. A total of 5,221 students with 1,797,677 logs, including 485 on-the-job students and 4,736 full-time students, were analyzed to depict their spatial-temporal learning patterns, including the relationships between identified patterns and students' learning performance.\nFindings: Analysis results indicate on-the-job students took more advantage of anytime, anywhere than full-time students. Students with a higher tendency for learning anytime and a lower level of learning anywhere were more likely to have better outcomes. Gender did not show consistent findings on students' spatial-temporal patterns, but partial findings could be supported by evidence in neural science or by cultural and geographical differences.\nResearch limitations/implications: A more accurate approach for categorizing position and location might be considered. Some findings need more studies for further validation. Finally, future research can consider connections between other well-known performance predictors (such as financial situation, motivation, personality and major) and the type of learning patterns.\nPractical implications: The findings gained from this study can help improve the understandings of students' learning behavioral patterns and design as well as implement better online education programs.\nOriginality/value: This study proposed concepts of time and location entropy to identify successful spatial-temporal patterns of on-the-job and full-time students.\n", "role": "assistant" } ]
Examining the effects of source selection method on procurement outcomes
[ "Procurement", "Supplier performance", "Source selection", "Supplier evaluation" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: One of the most hotly contested debates in defense procurement has been the choice of contractor selection method and the procurement outcomes resulting from that choice (Kashiwagi and Byfield, 2002; Lohfeld, 2015). The low-price, technically acceptable (LPTA) method has been pitted against the full trade-off (TO) method, with the LPTA method taking the brunt of the criticism. While Department of Defense (DoD) policies encourage contracting officers to:\nTheory and hypotheses: In essence, the principal issue at hand is the value acquired by both sides of the buyer-supplier dyad. Value is the undercurrent of sourcing decisions as buyers lean on suppliers for cost savings and competitive advantage (Chin-Chun et al., 2006). Value is the net difference between benefits gained (e.g. lower-cost inputs from suppliers, superior quality goods and services purchased from suppliers) and sacrifices made (e.g. price paid) (Lam et al., 2004; Yazdanparast et al., 2010). During source selection, the value determination manifests in the evaluation of price versus non-price evaluation criteria. Surprisingly, very little research addresses how buying teams evaluate value ex ante (Koufteros et al., 2012).\nMethodology: Data collection\nResults: Descriptive statistics\nDiscussion: Two fundamental approaches to the evaluation of competing suppliers' proposals include a LPTA method and a TO method. Each method has advantages and disadvantages; but, heretofore, research has not explored the links between source selection methods and key procurement outcomes. Exploring the differences is important because buyers may not enable the co-creation of the most value if an inappropriate source selection method is used. Additionally, exigencies could nudge buyers toward emphasizing price such as the:\n", "role": "user" }, { "content": "Purpose: The purpose of this research is to explore the effects of supplier selection method on key procurement outcomes such as procurement lead time (PLT), supplier performance and buyer team size.\nDesign/methodology/approach: Data were collected from a sample of 124 archival contract records from the US Department of Defense. A multiple regression model and multivariate analysis of covariance/analysis of covariance models were used to test the effects of source selection method on pertinent procurement outcomes.\nFindings: The trade-off (TO) source selection method increases PLT, as does the number of evaluation factors and the number of proposals received. Substantially larger sourcing teams are also associated with the TO source selection method. Nonetheless, the TO method results in better supplier performance.\nPractical implications: TO source selections yield superior supplier performance than low-bidder methods. However, they are costly in terms of time and personnel. Any assessment of supplier value should consider not only the price premium for higher performance but also the transaction costs associated with the TO method.\nOriginality/value: Very little research addresses a buying team's evaluation of supplier-offered value ex ante and whether that value assessment materializes into actual value-added supplier performance. Low bidder tactics are pervasive, but price (i.e. sacrifice) is only one component of value. Benefits from superior supplier performance may yield greater overall value. If value is critical to the buyer, a TO source selection method - versus a low-bidder approach - is the appropriate tool because of higher supplier performance ex post.\n", "role": "assistant" } ]
Optimization model for transportation planning with demand uncertainties
[ "Controllable lead time", "Inventory control", "Simulated Annealing", "Stochastic demand" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Inventory systems that demonstrate demand uncertainty and variable lead time have received increasing attention in recent years (Lan et al., 1999; Ouyang et al., 2002; Song et al., 2010; Datta and Christopher, 2011). A number of studies have examined the issues of lead time reduction including its determinants and effects (Wang and Wang, 2007; Heydari et al., 2009; Glock, 2012).The literature also shows that lead time reduction is associated with additional costs and risks (Liao and Shyu, 1991; Chandra and Grabis, 2008). However, there has been limited research that explicitly investigate the issue of how change in the lead time impacts on different components of the inventory cost function. This study develops a cost function for an inventory system in an environment with controllable lead time and random demand. Additionally, a less complex approach that could be adopted by practitioners is proposed.\n2. Literature review: Lead time reduction has been one of the major benefits of successful implementation of the popular just-in-time (JIT) inventory system (Sakakibara et al., 1997; Green and Inman, 2005; Ward and Zhou, 2006; Mackelprang and Nair, 2010; Danese et al., 2012). There has been recent research on inventory models that consider lead time as a decision variable. For example, Ouyang and Wu (1997) proposed a continuous review inventory model with lead time reduction by allowing shortages with a mixture of backorders and lost sales. Liu et al. (2007) considered a supply chain involving price and lead time sensitive demand. In the classical literature reorder point/order quantity (R,Q) policy has been widely used in decision supporting system of inventory control, while lead time is considered a constant. However, in recent studies, effects of lead time have been taken into account. For example, Axsater (2006a) investigated the problem of optimal (R,Q) policy with fill rate constraint in the context of normal distributed lead time demand. Song et al. (2010) studied the optimal policy parameters and long-run average cost of (R,Q) system relating to reduction in lead times.\n3. Proposed model: Reorder point/order quantity (R,Q) policies allow customer demand to be considered as a random variable. One of the assumptions in the (R,Q) policy is that lead time is fixed and independent of cost. However, we assume that lead time and ordering cost (aka setup cost[1]) are related. The evidence for the relationship between ordering cost and lead time is commonly found in industry. Service providers of transportation usually impose a surcharge for the express delivery of goods and that determines the relationship between these two variables.\n4. Solutioning and numerical examples: The application of the proposed inventory cost function for decision making requires a pragmatic method for deriving the solutions. This study uses a standard iteration procedure which is widely used to solve the optimal solution of inventory control with uncertain demand and a SA algorithm that offers a much simpler solution based on a heuristic algorithm. SA is a meta-heuristic approach to acquire the global optimal or near-optimal solution in solving optimization problems (Hillier and Lieberman, 2005). Besides its efficiency in searching for global optimal solution, SA is widely adopted as it is simple to formulate and takes less CPU time than some complex algorithms, such as genetic algorithm (Suman and Kumar, 2006).\n5. Discussion: This study aims to investigate the total cost function of an inventory system with a controllable lead time and random demand. It was observed that the uncertainty of demand and the form of relationship between the lead time and ordering cost have significant effects on the optimum cost solution. In the numerical simulations, the basic case scenario demonstrated the application of an iterative procedure that allows one to obtain the optimal solution for (L,R,Q) policy. The results show that while change in lead time had a direct impact on ordering cost per order, it also determines the holding cost as well as penalty cost per unit time. Significant effects of change in lead time were observed on the demand during lead time. The expected demand and the associated uncertainty together affect holding cost per unit time as well as penalty cost per unit time. In overall term, change in lead time has different effects on different cost components of the total inventory cost.\n6. Conclusion: This study investigates the total cost function of an inventory system with a controllable lead time and random demand in an environment where buyers could pay extra for expedited delivery of goods. The analysis presented in this paper aims to contribute to the existing knowledge base by investigating the change in demand distribution with respect to changing lead time and the effects of change in lead time on inventory cost components of holding cost, ordering cost, and penalty cost. We find that besides the direct effects on the ordering cost, change in lead time influences the expected demand during the lead time, which in turn affects other cost components. Such effects need to be considered in parallel to the general focus on excess demand and penalty costs in random demand situations. The analysis presented in this paper shows that even when the rate of demand is a constant, demand and corresponding expected excess demand might change with the change in lead time. The proposed approach accommodates this behavior and delivers optimum total cost solutions. These solutions are demonstrated using numerical case studies. The optimal solution is derived by making use of modified iteration process, which is based on the classical iteration process that is widely used in determining the ordering policy provides an optimal solution. We also use a simpler approach based on SA algorithm to derive the optimal solution. Both methods yield largely similar results. Unlike previous models that are mathematically complex and difficult to apply in practice, SA algorithm provides a more pragmatic and less complex method that is easily adopted by ordinary logistics practitioners and requires input data that are readily accessible from their transportation service providers. The proposed approach enables logistics practitioners in less complex operations making systematic decisions using commonly available software programs. Future research could extend this study to investigate the real behavior of firms and practical applications managing controllable lead time for cost minimization.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to investigate the total cost function of an inventory system with a reorder point/order quantity policy where the lead time is controllable based on the cost paid by the buyer for the service.\nDesign/methodology/approach: Cost functions are presented to investigate how the changes in lead time affect different components of inventory cost in the present of random demand. Two methods including an iteration technique and Simulated Annealing (SA) algorithm are presented to deal with the cost optimization issue. The application of proposed model is illustrated using numerical case scenarios.\nFindings: The cost functions show that besides ordering cost, change in stochastic demand during lead time is the major factor that affects the other cost components such as holding and penalty costs. This finding is validated by numerical study. Results also show that performance of SA algorithm is highly similar to iteration methodology, while the former one is easier in application.\nPractical implications: This paper develops less complex, more pragmatic methods, easily adoptable by logistics managers for cost minimization. This paper also analyzes and highlights the unique characteristics and features of these two approaches that can help practitioners in making the right choice when faced with the identified logistics issue.\nOriginality/value: This research explicitly investigate impacts of changing lead time on inventory cost components which enables informed decision making and inventory system planning for cost optimization by logistics practitioners. Two methodologies that can be easily used by practitioners without deep mathematical analysis and is cost effective are introduced to solve the optimization problem. Detailed roadmaps of how to implement proposed approaches have been illustrated by different case scenarios.\n", "role": "assistant" } ]
Enabling and leveraging ambidexterity: influence of strategic orientations and knowledge stock
[ "Entrepreneurial orientation", "Ambidexterity", "Market orientation", "Knowledge stock" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: To compete effectively, organizations need to pursue and acquire new knowledge as well as integrate and institutionalize acquired knowledge (Cohen and Levinthal, 1990; Crossan et al., 1999; Levinthal and March, 1993; March, 1991). Prior researchers have defined ambidexterity as the ability of an organization to simultaneously engage in the paradoxical activities of searching for and discovering new knowledge (i.e. exploration) while refining and building upon existing knowledge (i.e. exploitation) (March, 1991). Ambidexterity enables firms to sense and seize new opportunities through simultaneous exploration and exploitation, allowing them to be efficient in meeting current demands as well as adaptive to changing demands (Gibson and Birkinshaw, 2004; O'Reilly and Tushman, 2004, 2013). Although ambidexterity has been theorized to lead to sustained performance and competitive advantage (Birkinshaw and Gupta, 2013; Jansen et al., 2008; Lubatkin et al., 2006; O'Reilly and Tushman, 2013; Raisch and Birkinshaw, 2008), empirical evidence on the relationship between ambidexterity and firm performance is mixed (Stettner and Lavie, 2014). Accordingly, calls have been raised for further research on its antecedents, nature and contingent effects on the relationship with firm performance (O'Reilly and Tushman, 2013; Raisch et al., 2009; Stettner and Lavie, 2014). Responding to such calls, we develop and test a theoretical framework of the interrelationships between strategic orientations, ambidexterity, knowledge stocks and firm performance.\n2. Theory and hypotheses: 2.1 Ambidexterity\n3. Method: 3.1 Sample and data collection\n4. Results: 4.1 Preliminary analysis\n5. Discussion and conclusions: Ambidexterity is seen as crucial for firm performance and survival. However, questions have been raised regarding the feasibility of attaining ambidexterity given the innate challenges of reconciling exploration and exploitation. In this paper, we expand our understanding of the antecedents of ambidexterity and study the impact of two strategic orientations, entrepreneurial orientation (EO) and market orientation (MO) on a firm's ability to pursue new opportunities while simultaneously exploiting their existing competencies. Consistent with theoretical underpinnings, our findings indicate that each of these orientations exert a positive influence on ambidexterity. In addition, we examine how accumulated knowledge contributes to or inhibits ambidexterity by examining the role of a firm's knowledge stock. Drawing from the knowledge-based view, firms are viewed as repositories of knowledge and their main competitive advantages arise from how they can use and manage what they know (Cohen and Levinthal, 1990; Grant, 1996; McIver et al., 2013). However, research has indicated that institutionalized knowledge can also be detrimental for an organization's performance because it could lead to the formation of core rigidities (Leonard-Barton, 1992). Building on these ideas, we examined the influence of two types of knowledge stock - TKS and MKS - on the linkage between ambidexterity and firm performance.\n", "role": "user" }, { "content": "Purpose: This paper aims to develop and empirically test a framework articulating the effects of strategic orientations (entrepreneurial orientation and market orientation) on leveraging ambidexterity. Further, the paper examines the moderating effects of knowledge stock (market knowledge and technological knowledge) on the relationship between ambidexterity and firm performance to gain additional insights into how ambidexterity can be leveraged in an organization.\nDesign/methodology/approach: Data were obtained from CEOs (or equivalent members of the top management team) of 234 firms. The adequacy and psychometric properties of all measures were evaluated and purified using a maximum likelihood confirmatory factor analysis (CFA), and the hypotheses were tested using ordinary least squares (OLS). A number of post hoc tests were conducted to develop a nuanced understanding of proposed effects.\nFindings: While both strategic orientations enhance an organization's ability to be ambidextrous, results show that some types of knowledge stocks facilitate, whereas other types hinder the influence of ambidexterity on firm performance.\nResearch limitations/implications: Both strategic orientations enable ambidexterity; however, technological knowledge stock impedes the effect of ambidexterity on firm performance, while market knowledge stock enhances this relation. Cross-sectional nature of the study imposes limitations on causal inferences.\nPractical implications: Different strategic orientations provide organizations with a cluster of knowledge acquisition and utilization capabilities that enable ambidexterity. However, organizations should be wary of indiscriminate accumulation of knowledge stocks - while certain types enhance the effect of ambidexterity, others may create competency traps or core rigidities and inhibit the effect of ambidexterity.\nOriginality/value: This study integrates related, yet hitherto fragmented, research streams to demonstrate the interconnectedness between strategic orientations, ambidexterity and existing knowledge stock. Several theoretical and managerial implications are identified.\n", "role": "assistant" } ]
Establishing mercury-free medical facilities: a Malaysian case study
[ "Hospital management", "Mercury", "Mercury-free medical facility" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Mercury is a toxic element that harms humans and the environment. Mercury management has intensified through the Minamata Convention on Mercury United Nations Environment Programme (UNEP) in 2009 to protect human health and the environment from adverse anthropogenic mercury release (UNEP, 2013c). This convention requires parties to meet agreed commitments leading to mercury reduction and elimination (UNEP, 2013c). Mercury has been used for various purposes, ranging from basic dye to complex measuring devices. It is estimated that almost 1m tonnes of metallic mercury had been used in the last 500 years (Hylander and Meili, 2003). Mercury is used in industrial processes such as metal production, medical industries and consumer products including batteries and electronic devices.\nMethodology: We studied a Malaysian teaching hospital, which provides healthcare services for the public and trains medical professionals. Currently, the facility is operating with 837 beds (62 percent occupancy). In 2012, staff looked after 34,473 inpatients and treated 538,987 outpatients. We adopted a framework (Figure 1) based on the life cycle approach, which aligns with the Minamata Convention on Mercury, i.e., managing mercury holistically at every life cycle phase from procurement until disposal.\nResults: Overall\nDiscussion: We found that mercury management processes are scattered and uncoordinated owing to an improper system. Mercury reduction initiatives were driven by individual awareness and technology advancement, but often limited by departmental awareness. Therefore, establishing a holistic mercury management will significantly assist in eliminating mercury use considering the recommended practices as guidance. We found that recommended practices are often interlinked, which enable steps to establish mercury-free medical facilities (Figure 2).\nConclusion: Our study showed that there are significant gaps between recommended and current mercury handling practices in every mercury life cycle phase, but at different magnitudes. Efforts toward establishing mercury-free medical facilities are on-going, but they are uncoordinated and not producing an impact without a formal structure and full commitment. We found that current bottom-up approaches must be strengthened by top down mechanisms including establishing a governance structure equipped with formal commitment and dedicated individuals to better manage mercury use. Core components needed in establishing a mercury-free medical facility include: a multidisciplinary team that identifies mercury use, purchasing policies, action plans, guidelines and monitoring systems. These components are interrelated and can be addressed using steps to strengthen individual phase-specific elements supported by funding, technology application, infrastructure, policies, training and awareness raising activities. Steps in establishing a mercury-free medical facility will be useful for hospital managers who are strategizing their mercury reduction efforts as they prepare for the Minamata Convention on Mercury, and more importantly, protecting human health and environment from mercury exposure.\n", "role": "user" }, { "content": "Purpose: Mercury is widely used in medical and healthcare facilities as dental amalgam, mercury-added medical devices, thiomersal-containing vaccines, laboratory analysis and for other general applications despite the hazards. Various agencies consistently promote mercury-free medical facilities through mercury-free alternatives and better management practices, which are in line with the Minamata Convention on Mercury that aims to protect human health and environment from anthropogenic mercury release. The paper aims to discuss these issues.\nDesign/methodology/approach: The authors conducted a gap analysis on recommended practices gathered from the literature and current practices gathered through semi-structured interviews with Malaysian medical personnel. A life cycle approach was adopted covering mercury use: input, storage, handling, accident, waste disposal and governance phases.\nFindings: The authors found that there are significant gaps between recommended and current mercury management practices. Analysis indicates improper mercury management as the main contributor to these gaps. The authors found from recommended practices that core components needing improvement include: mercury management action plan, mercury use identification team, purchasing policy, proper guidelines and monitoring systems.\nPractical implications: This study helps us to understand mercury management practices and suggests essential steps to establish a mercury-free medical facility.\nOriginality/value: This study explored the gaps between recommended and current mercury management practices in a medical facility and contributes to the Minamata Convention on Mercury aspirations.\n", "role": "assistant" } ]
Managing career lengthening in France: why such inertia?
[ "Inertia", "Career lengthening", "Career scripts", "Institutional approach" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nThe changing landscape of careers in France: The aging of the European population has consequences for the balance of social accounts and economic growth. Since the Laeken European Council in December 2001, several meetings have contributed to the definition of a policy aimed at \"raising employment levels\" and \"lengthening working life.\" In France, these objectives have given rise to numerous legislative and conventional reforms concerning retirement pension and unemployment benefit schemes. In addition, they entail the obligation for the employer to organize regularly an interview on career development - with employees aged 45 and over at the time of the survey, but since 2014 the obligation applies to all employees more than two years of seniority and without age requirement. Employees are encouraged to anticipate their careers in a relatively confused environment given the unemployment rate. As far as the average age of the sample is 51, this paper could focus on older workers (Hennekam, 2015). Nevertheless, it also seems interesting to take into account the context of careers in France. In this perspective, the purpose of this research paper is to understand the extent to which an analysis of career scripts helps to address the inertia in the implementation of a mandatory interview process, designed to anticipate career lengthening in France.\nTheoretical framework: The literature review first presents the concept of the career script and its institutional perspective. It then proposes a description of the context of careers in France.\nMethodology: In order to explain it, this section successively presents the research field, the study design and the findings.\nFindings: The ATLAS ti(c) software records 141,805 words in the 26 interviews that lasted an average of 1 h and 1 min. The word count on the corpus cleared of articles and linking words shows a consistency of vocabulary with the interviewees' perception of career scripts. The coding process isolate 3,103 quotations that represent 114,850 words or 81 percent of the corpus. This make it possible to sort by number of quotations a set of 12 metacodes that describe the most relevant elements of the context. Table II presents the classification of these 12 constructs.\nDiscussion: First of all, it is worth mentioning the limitations of this survey. The latter was conducted on a non-random sample of older workers in a sector undergoing radical change. Moreover, on the theoretical level, the institutional approach tends to overestimate resistance to change. However, these two career scripts suggest that the prospect of a longer working life is still not embedded in the career management practices neither of the agencies nor of the cadres. Pessimistic and optimistic, both scripts seem to anticipate careers according to a similar social logic. This has practical and theoretical implications.\nConclusion: The goal of this research paper was to understand to what extent an analysis of career scripts helps to address inertia in the implementation of mandatory interviews, designed to anticipate career lengthening in French media agencies.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to understand to what extent the analysis of career scripts explains the inertia in implementing mandatory interviews designed to anticipate career lengthening in France.\nDesign/methodology/approach: This exploratory approach is based on 26 interviews of French cadres (a mix of managers and professionals) aged 45 and over in the media agency sector.\nFindings: Two career scripts - one for the majority who perceive themselves in a dead end, and the other, more optimistic for those who experience job mobility - show several commonalities. Moreover, their combination suggests an entanglement with the previous business model of media agencies.\nResearch limitations/implications: The survey was conducted on a non-random sample in a radically changing sector. However, the entanglement may explain the inertia with regard to changes in career scripts and agencies' business model.\nPractical implications: As these cadres need to work longer and move from traditional media expertise to cross-functional consulting capacity, mandatory interviews should be integrated into a human resources management strategy that fosters intra-organizational dynamics.\nSocial implications: The lengthening of the professional life in France appears as a paradigm shift. The study of career sustainability in contexts becomes a critical research field for management development.\nOriginality/value: The institutional approach of managers' careers is well developed, but there is a gap in understanding the entanglement of career scripts and business models. This may help address inertia in change processes.\n", "role": "assistant" } ]
Retrospective: a cross-sectional test of the effect and conceptualization of service value revisited
[ "Customer value", "Service co-creation", "Service dominant logic (SDL)" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The 30th anniversary of the initial publication of the first marketing journal devoted to the study of service, Journal of Services Marketing, is an excellent time for celebration and reflection. Those 30 years have seen dramatic changes in technology, the economy, regulation and consumers' shopping behaviour that continue to affect how consumers evaluate service exchanges and make purchase decisions. These dramatic changes also suggest that the criteria consumers use to evaluate potential or actual service experiences deserve renewed consideration. Digital connectivity, online retailing and a growing trend towards do-it-yourself solutions to consumers' needs and wants are just three of many changes that the last thirty years have witnessed that likely affect the quality, value and satisfaction consumers attribute to a specific service experience. In turn, these assessments likely drive the management of service organizations and the marketing of service products. These facts alone justify a reconsideration of the role of traditional consumer decision-making drivers such as value.\nHistory of the 1997 Journal of service marketing article: My personal motivation in the mid-1990s was to begin a gradual move of the central focus in my own research from service quality to value. My co-authors were all doctorial students or new assistant professors, so getting published was also a major concern. The inspiration for moving to a focus on value reflected a belief that value represented a more holistic appraisal mechanism for consumers' assessments of exchanges. Specifically, my focus on quality research was achieving some success, but it was becoming apparent that quality is a unitary construct where more is generally assumed to be better. In contrast, value is a relativistic construct that seemed to better reflect market exchanges. That is, it was apparent that consumers do not always choose the highest quality product or buy from the provider with the best service. The success of Wal-Mart serves as testimony to this assumption. Rather, it seemed that consumers' decisions are restrained by budget, time, effort and possibly other considerations and generally buyers appear to be motivated to get the most they can from what they invest in an exchange. This decision criteria fit the existing definition of value (Zeithaml, 1988) near perfectly.\nDirection from service (Dominant) logic: Without going into too much detail given space limitations, it is fair to suggest that SDL and SL research is bringing renewed attention to value. Value is the core construct in the study by Vargo and Lusch's (2004, 2008a) self-described efforts to change what has been described as 200 years of thought relative to the theory that describes consumer decision-making (Ballantyne and Varey, 2008), as well as in SL research (FitzPatrick et al., 2015). Specifically, both suggest that the appropriate unit of analysis in the assessment of value is value creation rather than products as suggested in traditional economic theory, which is termed Goods Dominant Logic (Vargo and Lusch, 2008b). In essence, SDL proposes that value is created collaboratively during interactions between actors (e.g. organizations and consumers) rather than embedded in products solely by a provider. This collaboration between actors in creating value, termed co-creation, suggests value is created when organizations work with, rather than for, customers (Karpen et al., 2012). It also credits customers as being partners in the value creation process (i.e. co-creation) rather than simply users of the value inherent in products (Ramirez, 1999). Specifically, it is suggested that providers and consumers act as resource integrators sharing knowledge and skills to co-create superior value solutions to customer needs and wants (Lusch and Vargo, 2006). In contrast, SL suggests that value is created individually and collectively by providers individually, users individually and collaboratively in interactions between providers and users (FitzPatrick et al., 2015).\nImplications for value research: SDL and SL research suggest that there are a number of highly relevant value-related issues that deserve to be a primary focus of contemporary marketing research. In the sections that follow, three specific issues are identified and briefly discussed that I believe represent important research opportunities. Answers to the questions identified have the potential to be truly transformative, as each offers assistance in clarifying how consumers conceptualize and assess their product encounters and service experiences.\nSummary and conclusion: The Journal of Service Marketing has filled an important and unique niche throughout its 30-year service to the profession. Those who have served as editors and reviewers deserve the gratitude of all service scholars. In 1987, when the study of service was slowly being recognized as a research priority in marketing, a worldwide group of innovators saw a need for an outlet for quality service research and stepped forward to fill that need. My co-authors and I owe them our thanks, as do all of those who have had an opportunity to publish their research in its pages. Hopefully, the recounting of the motivation that drove the article published with my co-authors in 1997, and the areas identified where there is potentially a need for research related to the conceptualization and measurement of value will encourage others, especially developing scholars, to pursue enquires relative to this dominant consumer decision-making construct. Such efforts are needed if truly TSR is to evolve. The co-creation of value is truly the direct path to identifying ways that service can transform the lives of consumers worldwide.\n", "role": "user" }, { "content": "Purpose: This paper aims to provide an evaluation of the findings first put forward in the article \"A cross-sectional test of the effect and conceptualization of service value\" with the benefit of hindsight, and to offer directions for further research and developments in the research area.\nDesign/methodology/approach: Research directions which emanated from the publication of the paper have been examined in the light of current service(s) marketing theory and practice. As a result, promising current and future strands of research have been identified.\nFindings: The focus of both the original and this paper is on the conceptualization and measurement of the value construct. Although much theory has been advanced relative to the measurement of value, the marketing literature lacks a conceptualization and measures that reflect this theory.\nResearch limitations/implications: The implication of the paper is that marketing scholars tend to focus more on to other studies before exhausting the opportunities to identify and test appropriate conceptualizations and measures of core marketing constructs such as value, quality, sacrifice and satisfaction.\nPractical implications: Although value is universally identified as a major driver of consumer decision-making, a full appreciation of its role in the strategic marketing efforts of organizations depends on having an accurate conceptualization and measures. The original paper and this revisit motivate and assist organizations in their efforts to better understand value and its impact in the decisions made by consumers.\nSocial implications: To contribute to the needs and wants of society, organizations must identify what is valued by society. The original and this revisit identify the creation of value as a basic need in encouraging consumer behaviours when the intent is to satisfy societal needs or desires.\nOriginality/value: The original paper was highly rated, and generated discussion and important further research. It has value as part of the history of service marketing research. The retrospective analysis by the authors gives a unique insight into processes and thinking associated with understanding key aspects that contribute to the historical development of service marketing, and provides substantial food for thought for future research directions.\n", "role": "assistant" } ]
Digital competences of the workforce - a research topic?
[ "Big data", "Research agenda", "Workforce", "Curriculum design", "Digital competences" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In 2017, the main challenges for companies are neither technological trends, disruptive innovations, nor new customer behaviour. Their main duty is to adapt their culture, mindset, and competences to the new, digital way of working (Accenture, 2016; Maedche, 2016). Results matter the most, and place and time have become less relevant, especially for knowledge workers, who are sometimes even called \"digital nomads\" (Makimoto and Manners, 1997). Although there are boundaries for digital nomadism, such as submission deadlines, update calls, or the availability of colleagues, the degree of freedom for employees working in the digital age is technically unlimited. Collaboration in the digital age occurs across geographical, functional, and hierarchical boarders, while creating interactive content using social media (Lepofsky, 2016; Malone, 2004). In addition to knowledge intensive jobs, industrial jobs are affected by digitisation as well. Just think of the impact due to automatisation, such as 3D printing, drones, the Internet of Things, and robots.\n2. Research design: While arguing for a new research stream which, thus far, is not existent, we have developed and challenged our thesis based on the experiences and findings gathered from a literature analysis. Regarding the literature review, we started with checking business and IS databases such as EBSCO and JSTOR with the search term \"digital competenc*\", which resulted in only a few results. None of the hits had any relationship to the workforce. One reason for this result might be that the term \"digital competences\" has not yet been established outside of the political and pedagogy fields (see Section 3). Another issue that arose when applying the search term \"digital competenc*\" without quotes was that the single components, \"digital\" and \"competenc*\", led to unmanageably extensive result lists. Thus, a common systematic literature review on the term digital competences was not a suitable method. Therefore, our approach was to manually check the Senior Scholars' Basket of Journals (AIS, 2011) from January 2011 until July 2016 regarding the topic of the digital competences of the workforce. This timeframe was chosen with respect to the general novelty of our topic and covers the emergence of the most popular publications dealing with the effects of digitisation on the workforce, e.g. Brynjolfsson and McAfee (2014) and Frey and Osborne (2013). In addition to the IS journals and their technology foci, and as we are focussed on the workforce, we included suitable HR journals (Human Resource Management, Human Resource Management Journal, Leadership Quarterly, Industrial and Commercial Training) in our search. With respect to the aspects regarding curriculum design, we considered scientific learning journals (International Journal of Innovation and Learning, Journal of Education and Learning, International Journal of Learning Technology) as well. We carefully screened the titles, keywords, and abstracts of all volumes for any relation to our topic. Subsequently, we checked the reference lists of selected publications for further relevant sources. However, while confirming our database search mentioned above, within this set of journals and the selected timeframe, we were unable to find any literature focussing specifically on the digital competences of the workforce. We only identified a few articles that touch upon this topic, e.g. when dealing with virtual team leadership. These results indicate that there are a very limited number of scientific articles on the topic of the digital competences of the workforce. We therefore decided to also consider non-reviewed references, e.g. the publications of political institutions and consulting agencies, thereby ensuring a high level of relevance.\n3. Current definitions and meanings of digital competences: When focussing on firms and other work-related organisations, competences can be understood as a combination of abilities, (work-related) knowledge, and skills held by an individual (Nordhaug, 1993). Following Gorbacheva et al. (2016) and Muller et al. (2014), abilities innately belong to an individual (like the ability to engage in logical reasoning). Knowledge means a theoretical understanding of a concept, while skills are the practical application of that knowledge. Other definitions of competences emphasise the aspect of attitudes, instead of knowledge, as well as the target of \"problem solving\" (Holtkamp et al., 2015; Winterton, 2009). Ha (2016) distinguished between two competence dimensions: epistemic and heuristic. While epistemic competence is the formal or knowledge-related aspect of competence, heuristic competence comprises the experiences and capacity to act which can be developed once the knowledge is constantly applied. Generally, competences should be understood within a specific context, e.g. existing systems or work practices (Hoel and Holtkamp, 2012).\n4. Pros and Cons: We challenge our thesis, Researchers must focus on the topic of the digital competences of the workforce, with regard to the three arguments. First, we deal with the opinion that the relatively new phenomenon of digitisation is simply part of permanent technical progress. Related technical skills are already targeted by current research. Thus, focussing on digital competences is redundant. The second insertion is that developing digital competences is a \"lifelong\" learning process. Consequently, it is not valid to limit research to the workforce. Third, because of other established research fields regarding competence research, it is argued that digital competences research is not a task for IS.\n5. Practical and theoretical implications: In the following, we aim to elaborate both the theoretical and the practical implications related to our thesis that the digital competence of the workforce is a research topic. We emphasise the need for a comprehensive scientific definition of the term, relevant actions in the HR field, and the link between research and curriculum design.\n6. Conclusion and research agenda: When we take a critical and honest look at today's working environment, and when we ask ourselves how our current job will be designed in 20 years (if it still exists), we must recognise that we have to radically transform our thinking about the way of working and the respective working requirements. But what exactly are the consequences on an individual level? Digital competences, whether or not one likes the term, are part of the answer, as they encompass the knowledge, abilities, skills, and attitudes we need for working in the digital age. Digital competences on an individual level is an umbrella term covering both the general digital competences (e.g. virtual collaboration) which are needed for nearly every knowledge occupation, and the specific role- or task-related digital competences which are different for every occupation. Coming back to big data analytics, completely new \"digital\" jobs have recently emerged, e.g. the \"data scientist\", but it is rather unclear which competences these jobs require. However, both career opportunities and potential earnings in the digital age will depend on developing suitable digital competences. But what these digital competences are, how they can be measured, and how they can be developed are only some of the questions for the research community to answer. This paper is a call to start shedding light on these issues.\n", "role": "user" }, { "content": "Purpose: Considering working in the digital age, questions on the consequences for the individual workers are, so far, often neglected. The purpose of this paper is to deal with the question of whether the digital competences of the workforce is a research topic. The authors argue for the thesis that it is indeed a research topic.\nDesign/methodology/approach: In addition to a literature analysis of the top IS, HR, and learning publications, non-scientific sources, as well as the opinions of the authors, are included. The authors' thesis is challenged through a debate of corresponding pros and cons.\nFindings: The definition of digital competences lacks scientific depth. Focussing on the workforce is valid, as a \"lifelong\" perspective is not mandatory for research. Digital competence research is a multidisciplinary task to which the IS field can make a valuable contribution.\nResearch limitations/implications: Although relevant references are included, some aspects are mainly driven by the opinions of the authors. The theoretical implications encompass a call for a scientific definition of digital competences. Furthermore, scholars should focus on the competences of the workforce, including occupations, roles, or industries. The authors conclude by providing a first proposal of a research agenda.\nPractical implications: The practical implications include the alignment of multiple stakeholders for the design of \"digital\" curricula and the integration by HR departments of the construct of digital competences, e.g. for compensation matters and job requirements.\nOriginality/value: This paper is one of very few contributions in the area of the digital competences of the workforce, and it presents a starting point for future research activities.\n", "role": "assistant" } ]
Effects of quality antecedents on e-learning acceptance
[ "Information quality", "Service quality", "System quality", "Instructor quality", "E‐learning acceptance", "Structural equation modelling", "Taiwan", "Learning methods", "Attitudes", "Workplace learning" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Recently, the application trend of information and communication technologies (ICTs) and Web-based technologies on education provides online educational approaches to conduct teaching and learning (Liu et al., 2009; Lee, 2010). This phenomenon has further facilitated the wide adoption of electronic learning (e-learning) on practices in education and training (Lee et al., 2009), thus organizations should invest their resources on developing online courses and training programs, and further provide their learners with access to knowledge acquisition and absorption via the e-learning systems (Chatzoglou et al., 2009; Limniou et al., 2009; Ozkan and Koseler, 2009). As compared to traditional classroom settings, e-learning can overcome the limitations of traditional learning and expand the educational territories without time, distance, and space barriers (Baylari and Montazer, 2009; Lee et al., 2009). Hence, e-learning has become a potential alternative to traditional classroom settings (Liu et al., 2010), and teaching and learning are no longer restricted to traditional classrooms (Wang et al., 2007; Ozkan and Koseler, 2009).However, simply providing learners with a Web-based learning system does not guarantee a successful e-learning. Nowadays, quality assurance, improvement, and enhancement of e-learning in organizations have increasingly become a central concern for education and training practitioners, experts, researchers, and providers (Ozkan and Koseler, 2009; Jara and Mellar, 2010). Obviously, quality assessment has become an essential requirement of evaluation for e-learning acceptance (Ozkan and Koseler, 2009). However, as far as the understanding of e-learning acceptance by organizations is concerned, it appears that numerous research efforts have focused on several types of external variables (e.g. individual factors, organizational factors, and social factors, etc.) for explaining users' e-learning acceptance (Avlonitis and Panagopoulos, 2005), with limited emphasis upon the quality antecedents of users' acceptance of the e-learning system, which is an important theme for organizations if they are to evaluate the system to solve teaching and learning problems. Besides, some studies (e.g. Ahn et al., 2004; Shih, 2004; Ahn et al., 2007; Lin, 2007; Kim et al., 2008) have addressed the relationships between quality factors and user acceptance of information systems (IS)/information technology (IT) to date, but the extent to which quality factors affect users' acceptance of the e-learning system has seldom been examined. Accordingly, the quality assessment of e-learning acceptance deserves special attention because the e-learning system can be regarded as a contemporary IS whose quality assessment has not been sufficiently reported in the IS acceptance literature. Hence, the main purpose of this study is to examine whether quality factors as the antecedents to learner beliefs can affect learners' intention to use the e-learning system. Based on the foregoing, this study proposes the research model to examine the quality antecedents for employees' acceptance of the e-learning system within high-tech organizations. Sample data collected from eight international/domestic high-tech companies in Taiwan were tested against the model by using structural equation modeling (SEM). The structure of this study is organized as follows: first, this study reviews the relevant literature to develop the research model and research hypotheses; next, this study presents the methodology and the results of empirical analyses; finally, this study proposes relevant implications and suggestions for practitioners and future researchers.\nTheoretical background, hypotheses, and research model: The outline of e-learning\nMethodology: Measures\nResults: Descriptive characteristics of the usable respondents\nResearch contributions and theoretical implications: Theoretically, quality factors (i.e. information quality, service quality, system quality) in IS success proposed by DeLone and McLean (2003, 2004) have respectively been measured and most of these have been employed as predictors of IS acceptance in some previous studies, but little attention has been given to evaluate the impacts of these quality factors on e-learning acceptance. To date, some external variables related to quality assessment have been respectively mentioned as central to users' acceptance of the e-learning system, however, these variables have not been classified according to the types of quality factors proposed by DeLone and McLean (2003). Hence, based on the updated DeLone and McLean IS success model, this study groups these external variables related to quality assessment of e-learning into the following three types of quality factors: information quality (i.e. course content quality, course design quality), service quality (i.e. support service quality), and system quality (i.e. system functionality, system interactivity, system response, user-interface design). In this study, the three types of quality factors as external variables to PU, PEOU and PE are respectively proposed and examined. Additionally, this study contributes to the identification of instructor quality factor (i.e. instructor attitude towards e-learners) that may lead to e-learning acceptance. All these above-mentioned things represent important theoretical contributions in the field of quality assessment for evaluating e-learning acceptance. This study's results reveal that four groups of quality factors including information quality, service quality, system quality, and instructor quality can make the significant impacts on employees' beliefs in affecting their e-learning acceptance. The results further implicate that collectively these four groups of quality factors have greater explanatory power than any single group of factors in describing the central quality antecedents of learners' e-learning acceptance. Besides, the results also implicate that learner-system interaction and instructor-learner interaction should be simultaneously evaluated and taken into consideration in the theoretical development of e-learning acceptance to acquire a more comprehensive and robust analysis.Furthermore, based on the TAM, previous studies (e.g. Lee, 2006; Pituch and Lee, 2006; Cho et al., 2009; Lee, 2010; Liu et al., 2010) explain quality antecedents of e-learning acceptance primarily via the extrinsic motivation, focusing on perceived usefulness (PU) and perceived ease of use (PEOU). A few e-learning studies address contribution of intrinsic motivation (i.e. perceived enjoyment (PE)) to learners' acceptance of e-learning systems. However, it should be noted that while TAM is based on the assumption of individuals' rationality in evaluating and analyzing external stimuli, PE is focused more on individuals' affective experience in explaining their non-beneficial behavior of IS/IT acceptance and usage that is difficult to explain with only the perspective of TAM (Lee et al., 2005; Choi et al., 2007). In this study, the empirical evidence on capturing both extrinsic motivators (i.e. PU and PEOU) and intrinsic motivator (i.e. PE) for completely explaining quality antecedents of learners' e-learning acceptance is well documented. Hence, this empirical evidence contributes significantly to the body of research on bridging the gap of limited evaluation for quality antecedents of e-learning acceptance, which is very rare in an intrinsic motivational perspective. Specifically, the evidence implicates that learners' PE (i.e. intrinsic motivator) can generate affective arousal caused by external stimuli (i.e. quality factors) to facilitate their usage of the e-learning system. Hence, PE (i.e. intrinsic motivation) also may well be a promising theoretical foundation for the field of e-learning acceptance, because it is useful for understanding the learners' intrinsically enjoyable experience (i.e. an affective experience) in explaining human behavior of acceptance and its results within the e-learning process.\nManagerial implications and practical suggestions: This study enhances the understanding of impacts of quality antecedents to learners' beliefs on their usage intention of the e-learning system, and thus provides organizations wishing to successfully facilitate the usage of the e-learning system with implications and suggestions. Detailed managerial implications and practical suggestions are proposed next.Quality antecedents to learner beliefs\nConclusions: This study proposes a well-rounded research model that based on the extended TAM and the updated DeLone and McLean IS success model to investigate the impacts of quality antecedents on employees' e-learning acceptance for a more robust analysis. The main research results are summarized as follows. On the part of quality antecedents to learner beliefs, as for factors of information quality, course content quality has significant effects on perceived usefulness (PU), perceived ease of use (PEOU), and perceived enjoyment (PE), and course design quality has a significant effect on PEOU; with regard to the factor of service quality, support service quality has significant effects on PU and PEOU; as far as factors of system quality are concerned, system functionality, system interactivity, and user-interface design have significant effects on PU, PEOU, and PE, and system response has a significant effect on PU; as to the factor of instructor quality, instructor attitude towards e-learners has significant effects on PU and PE. On the part of relationships between learner beliefs and usage intention, PU, PEOU, and PE respectively exhibit significantly strong impacts on intention to use the e-learning system, and PEOU also indirectly affects intention to use the e-learning system via PU and PE. Synthetically speaking, four types of quality factors, as the antecedents of e-learning acceptance can provide detailed accounts of the key forces underpinning learners' perception with regard to their beliefs (i.e. PU, PEOU, and PE), and this situation can further lead to the enhancement of learners' usage intention of the e-learning system.Several limitations should be noted in this study, and the following suggestions for further research will be worth future efforts in this field. First, because this study was limited to the high-tech sector only, given this study's limited scope, caution must be exercised in generalizing from this study's sample to firms in other sectors. Second, this study's results were obtained within the context of one type of the e-learning system (i.e. the LMS). Further research should need validation across other systems. Third, this study's data were collected from eight high-tech companies in Taiwan only, given this study's limited scope, caution must be exercised in generalizing from this study's sample to the respondents of other national cultural backgrounds. Further research may include the interactions of national culture as hypotheses in the research model. Fourth, respondents might usually display different serious reactions or relaxed feelings for the same e-learning system, depending on which situation they used the system. Further research in e-learning acceptance may examine the different set of mechanisms that involve voluntary/mandatory or work/home usage settings. Fifth, in this study, a comprehensive model employing multi-indicator measures was proposed to examine the impacts of four types of quality factors on e-learning acceptance for a more elaborate analysis. However, it might not draw a clear picture of the causal relationship between one specific quality factor and e-learning acceptance while other variables changed. Hence, further research may narrow the scope of this study and focus attention on a specific scope, and use the experimental method to manipulate one specific quality factor (the independent variable) which has been selected in order to observe its effect on e-learning acceptance (the dependent variable), while other variables that may confound such a causal relationship are controlled. Finally, this study implemented a cross-sectional analysis and might not draw a complete picture of the course of e-learning acceptance over time. It may be desirable to understand the quality antecedents of e-learning acceptance and how these influences change over time with increasing experience in using the e-learning system. Further research may use a longitudinal analysis by taking into consideration the evolution of e-learning acceptance over time.\n", "role": "user" }, { "content": "Purpose: - The main purpose of this study is to examine whether quality factors as the antecedents to learner beliefs can affect learners' intention to use an e-learning system.\nDesign/methodology/approach: - This study gathered sample data from eight high-tech companies in Taiwan. A total of 680 questionnaires were randomly distributed, 522 questionnaires were returned for a response rate of 76.76 percent, and 483 usable questionnaires were analyzed, with a usable response rate of 71.03 percent. Data were analyzed by using structural equation modeling.\nFindings: - Information quality, service quality, system quality, and instructor quality, as the antecedents of e-learning acceptance can provide detailed accounts of the key forces underpinning employees' perception with regard to their beliefs (i.e. perceived usefulness, perceived ease of use, and perceived enjoyment), and this situation can further enhance employees' usage intention of the e-learning system.\nOriginality/value: - Based on the extended technology acceptance model and the updated DeLone and McLean information systems success model, this study integrates related e-learning quality factors including information quality, service quality, and system quality into the research model and further contributes additionally to the identification of instructor quality that may lead to e-learning acceptance. Also, it should be noted that the empirical evidence on capturing both extrinsic and intrinsic motivators for completely explaining quality antecedents of e-learning acceptance is well documented in this study. Hence, this study contributes significantly to the body of research on evaluating the quality antecedents of e-learning acceptance.\n", "role": "assistant" } ]
RFID and buyer-seller relationships in the retail supply chain
[ "Buyer‐seller relationships", "Identification", "Supply chain management", "Retailing", "Radiofrequencies" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Radio-frequency identification (RFID), a technology that facilitates the tracking and manipulation of physical items, is now a major technological trend. According to experienced early adopters, knowledgeable industry leaders and academic researchers, it facilitates collaboration between organizations (Cantwell, 2006; Lefebvre et al., 2006; Lekakos, 2007; Roberti, 2006). In this context, it constitutes a supply chain management (SCM) enabling technology (Attaran, 2007; Pramatari, 2007).SCM is defined as the integration of business processes among companies that collaborate in an effort to bring a product, service or unit of information from the initial supplier to the final customer (Lambert et al., 1998; Mills et al., 2004). It has increasingly become the basis for competition (Angeles and Nath, 2000) between supply chains and networks (Gomes-Casseres, 1994). SCM provides many benefits including cost reductions (Spekman et al., 1998) and a collaborative advantage, rather than a competitive advantage (Chen and Paulraj, 2004). Within the supply chain, buyer-seller relationships are the individual links, the dyads, among companies that tie it together. Without these links, there would be no supply chain. They are therefore of paramount importance for SCM. In fact, a literature review of over 400 academic articles on SCM has placed these links at the center of SCM research (Chen and Paulraj, 2004).At first glance, RFID seems to support the move from a more adversarial stance towards collaborative relationships. Paradoxically, though, this does not always seem to be the case, at least in the retail industry where RFID has attracted the most interest (LoPrinzi, 2006). RFID implementations between major retailers and their suppliers seem to be far from collaborative. Several preliminary anecdotes from the industry indicate that RFID tends to upset relationships (Fogarty, 2004; Romanow, 2004; Schwartz, 2004) by the use of power (Williams and Moore, 2007) rather than solidifying them. More recently, a study in the UK grocery retail sector indicates that suppliers will implement RFID to satisfy their client's needs at their own short-term detriment (Hingley et al., 2007). This finding is corroborated by a quantitative study in which case level tagging was found to result in substantial losses for manufacturers, but not their retailers (Bottani and Rizzi, 2008). These observations clearly seem contradictory to a collaborative approach. They therefore warrant additional empirical research and serve to remind us that little is known of RFID's impact on buyer-seller relationships and even less is known of the impact of RFID on all of the buyer-seller relationships in an entire supply chain. The goal of this paper is therefore to improve our understanding of this issue by answering the research question:RQ. How do RFID and the buyer-seller relationships of an entire supply chain affect one another?The paper is organized as follows. The next section will review the literature on how similar collaborative technologies have affected and been affected by buyer-seller relationships. Then the methodological approach will be presented. The results section will first present the model that was developed to organize the field data and communicate the findings. Then the findings will be presented with a focus on the key issues that emerged from the field data. Finally, the paper will identify how the key issues influence the dimensions of the buyer-seller relationship, how the study contributes to academia and practitioners and conclude with suggestions for future research.\nLiterature review: An RFID system can be briefly described as follows: an electronic tag containing historical, transactional or identifying data are affixed to or embedded in an object. The data are automatically downloaded wirelessly to a computer when the object nears the vicinity of an RFID reader. Once on the computer, the information can travel anywhere that is accessible by the internet or on a private network.RFID as an IOS\nMethodology: Research design\nResults and discussion: Current processes in the retail supply chain\nConclusion: The interplay between RFID and buyer-seller relationships in a supply chain is an intricate phenomenon. Based on the empirical evidence from the ten firms involved in the field study, the following summary observations can be made:* Communication/information sharing. Access to shared information is at the heart of an RFID initiative. The information related to a specific product is shared among trading partners as the product travels downstream through the supply chain. Information can also travel upstream if the EPC network is used. Currently, the information related to case level quantities and shrinkage interests the respondents the most.* Cooperation. The desire to make a shared RFID supply chain project a reality clearly necessitates a deep desire for cooperation from all members of the supply chain. Respondents were aware of this and were constantly looking for ways to demonstrate the shared benefits of the system to encourage cooperation. Once the technological infrastructure is in place, it should also facilitate supply chain initiatives, thereby enabling the supply chain to work as a more collaborative team.* Trust. An RFID system has the potential to reduce a major part of the shrinkage that has caused a lack of trust among some of the organizations in the study. By acting as a constant verification system, it provides accountability for quality and volumes of shipments. Shrinkage will tend to diminish, thus making the trading partners more trustworthy. However, as demonstrated in the field, an initial inherent level of trust that partners will not try to bypass the RFID system is necessary if it is to be effective.* Commitment. The companies included in the proof of concept were identified and chosen by the respondents as being long-term partners. Implementing the RFID system will initially necessitate a considerable investment in specific investments. A portion of this investment therefore represents sunk costs that will not be recovered should the relationship end. By investing in the system, these companies expect the relationship to last. Additionally, once RFID is implemented, the relationship should be more profitable. The natural tendency would be to build on the relationship as long as possible in order to recoup the initial investments and increase profits.* Relationship value. Additional information will be accessible to members of the supply chain who use RFID together. Shrinkage in the supply chain will be reduced. New opportunities for collaboration will increase the value provided by customers and suppliers who use RFID. The relationships with trading partners who use RFID will therefore be increased when compared to those with partners who do not use the technology.* Power imbalance and interdependence. Both suppliers and buyers feel that an RFID initiative will give the other party more power. Suppliers feel that an investment and the associated sunk costs may lock them in vis-a-vis the buyers. Buyers feel that those suppliers who use RFID technology will gain a more competitive position.* Adaptation. Some organizations that are lured by the benefits that RFID promises may use their power to influence their trading partners to adapt to the new RFID requirements. As RFID technology improves, there will be a trend to push the technology further back and further down in the supply chain. As this happens, the company in the role of manufacturer is likely to be increasingly pressured by its trading partners to adopt RFID. In our study, the retained RFID scenario was developed by minimizing the required adaptation so that the technological initiative be the least disruptive possible.* Conflict. The ten organizations in this study are aware of potential conflicts related to RFID adoption and use positive reinforcements to get their trading partners to join the RFID initiative. If the trading partner does not want to go in this direction, pressure may eventually be applied. The situation could potentially create conflict in the buyer-seller relationship.Some deep practical implications arise out of these findings. Wal-Mart's initiative acted as the flagship for RFID as an IOS project but now other supply chains are currently evaluating or going forward with similar RFID initiatives. It is important that these companies be aware of the strategic role of supplier-buyer relationships as both crucial antecedents that shape RFID infrastructure and a consequence of RFID implementation. Furthermore, the continual improvement of RFID technology will tend to strain relationships among supply chain members since the use of power to urge adaptation may lead to conflict. RFID technology brings business partners one step closer to the vision of aligning their supply chain and raising the competition to the level of \"chain versus chain,\" thus moving closer to acquiring a collaborative advantage. For researchers, the results point to the importance of considering the different layers of the supply chain when investigating collaborative technologies such as RFID, relationships between business partners, or competitive positioning. This is particularly interesting for academic fields such as SCM, industrial marketing and diffusion of innovation research.\nFurther research: The results presented in this paper should be interpreted in the light of certain limitations. First, the observer's role is intermingled in this field research with that of the participant. This has the potential risk of influencing the participants by advocating a particular position, the risk of supporting a group of individuals when neutrality is required, and the risk of the participant role's requiring too much attention for good observations to take place (Yin, 2003). The researchers were aware of those potential pitfalls and tried to remain as neutral as possible. Second, the technological scenario is based on current improvement trends in the RFID industry. More specifically, perfect read rates are an optimistic extrapolation of today's state-of-the art technology but this is seen as a temporary technical obstacle. Third, the study was conducted in a single supply chain in the retail industry, which allowed us to gain in-depth knowledge of a complex phenomenon.Future research could attempt to validate these findings in other supply chains in other industries. Moreover, future studies could delve deeper into this phenomenon. For example, a longitudinal study of an RFID initiative implemented in a supply chain could indicate whether relationship quality diminishes in the short-term only to increase later, as has been found to be the case with barcode and EDI implementations. It would also be interesting to identify the RFID adoption antecedents and the contributing factors that make trading partners want to either collaborate or resist.\n", "role": "user" }, { "content": "Purpose: - The aim of this paper is to understand why a collaborative innovation, such as radio-frequency identification (RFID) technology, engenders seemingly opposite buyer-seller relationship reactions among members of a supply chain.\nDesign/methodology/approach: - The researchers carried out a three year exploratory multiple case study using the grounded theory approach through participant-observation and collaborative action research. The research project culminated in a fully functional RFID proof of concept solution which involved multiple layers of a retail supply chain. The researchers chose the retail industry for the study because it provided the ideal conditions to answer the research question.\nFindings: - The results highlight the role of supplier-buyer relationships as both crucial antecedents that shape RFID infrastructure and the consequence of RFID implementation. Specifically, the impact on and of eight key dimensions was considered, namely communication and information sharing, cooperation, trust, commitment, relationship value, power imbalance and interdependence, adaptation, and conflict. The paper also positions open-loop RFID projects as supply chain inter-organizational systems and presents a model to analyze such initiatives.\nResearch limitations/implications: - The choice of methodology has provided the insight necessary to answer the research question. Other researchers are encouraged to validate these findings through replication with other case studies or through quantitative data to reach analytical and statistical generalizability. The authors also encourage future research on this topic in other industries.\nOriginality/value: - As more supply chains move forward with collaborative RFID initiatives, it is important that these companies be aware of the strategic role of supplier-buyer relationships as both crucial antecedents that shape RFID infrastructure and a consequence of RFID implementation.\n", "role": "assistant" } ]
Savings operations with random commencement and conclusion
[ "Savings operations", "Profitability", "Risk quota", "Probability of survival", "Saving quota", "G22", "G23", "H55" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In this paper, we propose a novel modality of random savings operations (Gil Pelaez, 1993), by considering that both the commencement and the end of the savings period are subject to some contingencies or eventualities whose probability distributions are well-known. This would be the case of an investor who wishes starting a savings operation in the present moment, in order to obtain a specific amount at a certain future instant, but whose money availability is now limited by the extra eventual expenses due to the care of a person linked to him (for instance, the spouse affected by a serious illness, or a dependent parent). In these circumstances, the investor could wish that the savings operation does not begin until the death of the linked person. On the other hand, if the investor dies, the periodic deposits will finish, in which case the final amount could not be received, or alternatively could be continued in which case the saved amount would be received by a third person, beneficiary of the operation. Thus, there is also a double possibility:\n2. Financial equivalence: Let us consider a savings operation consisting of the periodic deposit by the investor of n amounts as, where s = 1,2...,n, with maturity at instant s - 1, respectively, with the aim to reach a final amount Cm at instant m, being m <= n. In this case, the commencement of the periodic deposits is subject to a contingency and, on the other hand, the end of the periodic deposits and also the withdrawal of the final amount are subject to another contingency. In these circumstances, the equation of financial equivalence at the commencement of the operation, by using exponential discounting at a variable interest rate is for each period (De Pablo Lopez, 2000) is:\n3. The withdrawal of the saved amount depends on the investor's survival: In this case, as previously indicated, the financial equivalence at time 0 is given by the following equation (observe that we have assumed that m = n):\n4. The withdrawal of saved amount occurs at an agreed date: When the withdrawal of the saved amount is sure, either by the investor (in case of survival) or by his heirs (in the case of the investor's death), the financial equivalence at the beginning of the savings operation is:\n5. Conclusion: There is an obvious need to devise some new private savings operations which complement the public pension plans. In this paper, we have proposed a new type of saving based on the contingency derived from the life expectancy of the saver, by raising an operation in which the commencement and conclusion of the savings period are random. These savings operations represent, undoubtedly, a novelty from a financial point of view. In Section 2, the equation of financial equivalence at the commencement of the operation has been introduced by distinguishing two variants of this operation: when the withdrawal of the saved amount depends on the survival of the investor and when the withdrawal of the saved amount is presented certain. These two cases are analyzed in Sections 3 and 4, respectively, where, in addition, the expressions of the classic financial parameters are obtained: balance, partial balance, interest quota, average interest rate and expected duration. Also, some other new parameters are introduced: savings and risk quotas, and various modalities of the average interest rate.\n", "role": "user" }, { "content": "Purpose: The progressive aging of the population is suggesting that public pension plans should be increasingly supplemented by private savings schemes. Accordingly, this supposes the appearance of a wide range of innovative savings products to meet the varying needs of savers and financial institutions. In practice, most contracted savings operations are nonrandom, that is to say, all amounts involved in the transaction are sure as well as their respective maturities. Consequently, the purpose of this paper is to propose a savings operation which includes the randomness derived from the contingencies which suppose the eventual but unpredictable death of the saver and a person designated by him to receive the final agreed amount.\nDesign/methodology/approach: The methodology used in this paper is financial mathematics where the risk has been introduced as an element which defines the main characteristics of this novel saving operation.\nFindings: The proposed model extends the range of savings products by describing an actual innovation with new practical applications with respect to the traditional models of saving. In this paper, the authors have proposed a new type of saving based on the contingency derived from the life expectancy of the saver, by raising an operation in which the commencement and conclusion of the savings period are random. These savings operations represent, undoubtedly, a novelty from a financial point of view.\nOriginality/value: The main added value of this paper is that these contingencies affect the periodic deposits in each period from the first to the last maturities of installments. Moreover, the different parameters of such random transactions are defined.\n", "role": "assistant" } ]
Momentum building for more women in senior appointments
[ "Gender equality", "Career advice", "Changing culture", "Pay differences", "Quotas" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: This article is a follow-up to two previous articles, namely \"Removing the boardroom glass ceiling\" (Hurn, 2012) and \"Are cracks appearing in the boardroom glass ceiling?\" (Hurn, 2013). It has been prompted by gradually increasing progress in this area and wider discussion at both government and business level on the desirability of increasing the number of women appointed to the boards of major companies both in the UK and overseas. Much of the discussion continues to be centred on the pros and cons of the need for quotas, the stubborn and continued male domination of certain sectors of business, the existing cultural and social barriers as well as the increase in the general awareness of the current yet still unsatisfactory system. It includes the consequent development of a number of HRM and other initiatives to actively improve the situation. What is perhaps most significant in recent months is the reported growing confidence and desire of women to aspire to gaining promotion on merit to senior appointments. This paper aims to review these trends and suggests the way ahead.\nLittle support for quotas: Calls for the institution of positive discrimination with prescribed quotas for women have been generally rejected as being seen as patronising and demeaning as well as unjust. Companies should select the best for the job regardless of gender. Women want to reach the top on merit, not because of any political correctness.\nWorkplace gender inequality: The Women's Business Council, formed in 2012 to advise the government, recommended in a report in June 2013 that there is a need to face up to the problem of workplace gender inequality early in school years (Women's Business Council, 2013). It emphasises that young women should be actively encouraged to be more ambitious and believe they can succeed particularly in science, engineering and mathematics. The recent tributes to Margaret Thatcher underlined how she succeeded in overcoming the excessively macho culture she experienced in politics. The report also stresses the need to encourage those women who take time out to raise children to return to work. To achieve this, attention should be paid to providing cheaper and more readily-available child care and working practices should be more flexible. By 2015 the government is planning to bring shared parental leave to come into effect. This will permit new mothers to return to work after two weeks and the rest of their maternity leave can be shared with their partner. However, the anticipated take-up by male partners is likely to be relatively low. A major conclusion of the report is that there is an \"overwhelming business case\" for increased female participation in the labour market. It estimates there are 2.4 million women in the UK who would like to work but are unable to do so because of the restraints of the availability and affordable cost of child care. If they were able to take up employment, it is estimated that GDP could be increased by 10 per cent by 2030.\nCareer guidance: There has been growing concern to motivate young girls to apply for apprenticeships in industry and science, particularly in IT, engineering and electronics. The Women's Business Council has emphasised the importance of encouraging girls still at school to widen their horizons and aspirations. The Business, Innovation and Skills Committee considers there are two crucial times when these concerns should be addressed, namely at age 14 when girls make their GCSE course choices and at age 16 when they decide whether to follow a vocational training course or to take advanced level subjects. It is important that appropriate careers advice forms part of the curriculum initially in the primary school and is continued throughout the secondary school years. For this approach to be effective, it must be part of teacher training and the role of governors, and be backed by a more robust careers service. Business schools and local businesses should work closely with schools to develop work experience placements.\nLink between equality, diversity and success: Jo Swinson, the Minister for Employment Relations and Consumer Affairs, emphasises the Government's voluntary initiative for business, in Think, Act, Report (Government Equalities Office, 2013), which encourages employers to focus on gender diversity and covers the inter-related HR activities of recruitment, promotion and staff retention. There is an urgent need to encourage more women into middle-management roles so that they can gain experience. To achieve greater equality data needs to be gathered to compare the salaries of men and women doing comparable jobs in a company. If there are significant discrepancies, these should be investigated. Deloitte, for example, investigates any discrepancies of 5 per cent or more. The transparency of salaries is another area. This can be effectively monitored if pay statistics are published by carrying out regular audits within the company to identify gaps. Companies which follow these actions should be seen to have a corporate culture which prides itself on promoting equality and diversity among its workforce, with promotion based on merit and achievement. Ceri Goddard, the Chief Executive of the Fawcett Society which campaigns for equality, considers the pay gap between genders one of the strongest remaining gender inequalities in the UK.\nChanging the culture: The major challenge is to change the entrenched culture which is predominantly male in many professions into one more representative of society as a whole. One way to achieve this is to use existing female senior executives as mentors and sponsors of women with high potential. Another initiative was introduced by HSBC in 2010 called The Bank's Balance Programme, whereby an employee network supports the recruitment, development and advancement of a gender-balanced workforce within the organisation.\nThe way ahead: There is now a need to build on the momentum produced by these encouraging trends in the business world in order to advance the case for more women in senior appointments. The reasons for the desired culture change have now gained much wider acceptance and considerable progress has been made. The following is a summary of what more needs to become accepted as best practice:\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to review the encouraging progress and increased momentum in both business and Government for more women to be appointed to senior positions. It traces this momentum since the author's last article: \"Are cracks now appearing in the boardroom glass ceiling?\" in Industrial and Commercial Training, Vol. 45 No. 4, 2013.\nDesign/methodology/approach: Critical review of recent initiatives by both government and business.\nFindings: The article highlights the positive momentum towards greater gender equality in the workplace. It emphasises the need for more career guidance beginning early in school for girls and to change the male-dominated culture found in many areas of business.\nResearch limitations/implications: Review of literature and media articles in the past 12 months.\nPractical implications: The increased antipathy towards quotas and the need for greater career guidance in both schools and colleges, together with work experience placements for women. Recruitment and selection should be gender-free, based on ability and experience and carried out by selection panels of both sexes.\nOriginality/value: Highlights and analyses the recent encouraging trends and increasing awareness of the value women bring to a company board.\n", "role": "assistant" } ]
Research on the quality stability evaluation and monitoring based on the pre-control chart
[ "Average run length (ARL)", "Evaluation guidelines", "Monitoring guidelines", "Pre-control chart", "Quality stability" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Along with the continuous improvement of the quality control requirements and the enhancement of the quality supervision, the improvement of quality management maturity needs a set of quality stability evaluation and monitoring guidelines that should conform to the characteristics of small batch production and the process shifts e[?]0. Of all the quality stability evaluation methods at present, statistical quality control techniques are the most widely used. However, the statistical quality control methods need a large number of samples and small batch production process cannot satisfy this requirements, which restricts the application of statistical quality control techniques (Peter et al., 2011; Aqua Group, 1992). One of the main methods to reduce sample size is to use the pre-control chart for quality stability evaluation and monitoring. Therefore, in order to satisfy the requirements of 6s management, it is very necessary to study the quality stability evaluation and monitoring of small batch production process based on pre-control chart.\nQuality stability evaluation scheme of the initial production process: It is necessary to evaluate and confirme the quality stability in initial production process (International Organization for Standardization, 2008), which corresponds to the first stage of the pre-control application. It is essential to determine the optimal sample number in the pre-control stability evaluation, set stability evaluation guideline based on pre-control chart, and propose the stability evaluation flow chart of the initial production process on the basis of the definition of the pre-control limits.\nQuality stability monitoring scheme of the formal production process: It is necessary to monitor the quality stability in the initial production process (International Organization for Standardization, 2008), which corresponds to the second stage of the pre-control chart applications. It is essential to propose monitoring guidelines at different sample size, determine appropriate monitoring guidelines and optimal sample size, and propose the stability monitoring flow chart of the formal production process.\nConclusions: On the basis of the comprehensive overview of quality stability evaluation theory concerning the pre-control chart both home and abroad, the paper finds that the existing pre-control charts are based on the assumptions that the shifts e=0 and the process capability indexes C\n", "role": "user" }, { "content": "Purpose: In order to meet the requirements of 6s management and to overcome the deficiencies of the theory for using the pre-control chart to evaluate and monitor quality stability, the purpose of this paper is to probe into the quality stability evaluation and monitoring guidelines of small batch production process based on the pre-control chart under the conditions of the distribution center and specifications center non-coincidence (0<e<=1.5s), the process capability index C\nDesign/methodology/approach: First, the range of the quality stability evaluation sampling number in initial production process is determined by using probability and statistics methods, the sample size for the quality stability evaluation is adjusted and determined in initial production process according to the error judgment probability theory, and the guideline for quality stability evaluation has been proposed in initial production process based on the theory of small probability events. Second, the alternative guidelines for quality stability monitoring and control in formal production process are proposed by using combination theory, the alternative guidelines are initially selected based on the theory of small probability events, a comparative analysis of the guidelines is made according to the average run lengths values, and the monitoring and control guidelines for quality stability are determined in formal production process.\nFindings: The results obtained from research indicate that when the virtual alarm probability a=0.27 percent, the shifts e in the range 0<e<=1.5s and the process capability index C\nOriginality/value: Research results can overcome the unsatisfactory 6s management assumptions and requirements and the oversize virtual alarm probability a of the past pre-control charts, as well as the shortage only adaptable to the pre-control chart when the shifts e=0. And at the same time, the difficult problem hard to adopt the conventional control charts to carry out process control because of a fewer sample sizes is solved.\n", "role": "assistant" } ]
Do union-management learning partnerships reduce workplace conflict?
[ "Trade unions", "Learning partnerships", "Managing conflict" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Managing conflict in the workplace has long been acknowledged as a key challenge for all organisations (Costantino, 1996; ACAS, 2006; Ridley-Duff and Bennett, 2011). Similarly, the value of promoting learning in the workplace and, in the context of this paper, specifically through partnership between trade unions and their management, has increasingly gained justifiable recognition (Shelley, 2007; Stuart et al., 2010a). What is less well researched is the potential efficacy of learning in also reducing conflict in the workplace. The aim of this paper, drawing on work undertaken for TUC Unionlearn, is to critically analyse the views and experiences of a cross-section of managers and trade union representatives on this question who have worked together on a number of particularly successful learning projects in North West England.\n2. Literature review: Recent employee relations research has generated valuable insight into the strengths and limitations of management and union partnerships and their impact on conflict (Oxenbridge and Brown, 2002; Martinez Lucio and Stuart, 2005). Furthermore, Munro and Rainbird (2004) have noted more specifically the \"distinctive form\" of the learning partnership in terms of shared objectives focused primarily on employee development. The aim of this paper is to build on this knowledge by considering how learning partnerships may also reduce individual and collective industrial action in the workplace. To that end, this section first considers some key aspects of organisational conflict in terms of its causes and consequences, before turning to a critical review of the current literature on the nature and influence of union-led learning initiatives in the UK.\n3. Methodology: An initial study of existing TUC project reports, and discussions with officers at TUC Unionlearn in the North West, helped identify a number of potential case study organisations and union projects which were seen as representative of successful learning partnerships in the region. The sample was subsequently decided in part on the basis of negotiating access to the organisation. For instance, a couple of highly successful local authorities declined to be involved, the researcher suspects, on the grounds of \"research exhaustion\" because they had already featured in a number of TUC and other reports. Nonetheless, the organisations willing to participate still allowed the researcher to study a sample of organisations from the public and private sector and also representing both large and medium-sized firms. It is of note however as in any case study research that the extent of generalisation of the findings must be treated with caution. The issue of the sample choice is critically reflected on further in the concluding section.\n4. Findings: The context\n5. Discussion and conclusion: There is an extensive and critical literature on the impact of union-led initiatives with respect to the development of both individual and organisational learning in the workplace (Warhurst et al., 2007; Hollinrake et al., 2008, McIlroy, 2008, Bacon and Hoque, 2008, 2009). Furthermore, earlier discussion in this paper has reflected on the many causes and consequences of workplace conflict (Blyton and Turnbull, 2004; Ridley-Duff and Bennett, 2011; Saundry and Jones, 2012). In contrast, the focus of this study has been to assess the degree to which learning initiatives may also have an impact on the managing of conflict in organisations, be that individual or collective disputes.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to report on a qualitative study of the views of a cross-section of managers and union officials engaged in joint learning projects on the impact more generally on the employee relations climate in those organisations.\nDesign/methodology/approach: Findings are based on an analysis of interview data incorporating the views and experiences of 32 union representatives and managers engaged in promoting learning in a number of case study projects in North West England.\nFindings: Findings suggest that union-led learning and development has a clear impact on pre-empting potential performance issues for individuals in terms of better equipping them with the skills to achieve expected performance levels, assisting individuals with managing work-related stress and also to prevent discrimination against workers. Collectively, the managing of downsizing in particular through union-led learning support to retrain in order to redeploy effected staff or better prepare them for new employment opportunities was also a significant finding.\nResearch limitations/implications: The research suggests that successful union-management learning partnerships can also have a positive impact on managing conflict in those organisations. However, further research is needed to test these assertions in other union organised sectors.\nPractical implications: There are implications for management in recognising the \"added value\" that partnership working with their unions on learning can bring in terms of also both pre-empting and resolving individual and collective disputes in the workplace.\nOriginality/value: The research is one of few studies that focuses on the link between promoting learning through union-management partnerships and managing conflict in organisations and offers a clear insight into how this can be achieved in practice.\n", "role": "assistant" } ]
Organizational and managerial challenges in the path toward Industry 4.0
[ "New technology", "Innovation", "Industry 4.0" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The concept of Industry 4.0, identified as the fourth Industrial Revolution, was initially introduced in Germany in 2011 (e.g. Lu, 2017; Sanders et al., 2016) referring to the integration of physical objects, human actors, intelligent machines, production lines and processes across organizational boundaries, meant to form a system in which all processes are integrated and share information in real-time frame (Hozdic, 2015). The basic principles of Industry 4.0 are therefore the connection of machines, work pieces and systems, and businesses are creating intelligent networks along the entire value chain that can control each other autonomously (Basl, 2017). This suggests that this industrial revolution is likely to affect people, companies and entire supply chains (SCs); thus, the associated level of organizational complexity is high. For this reason, manufacturing companies are looking for efficient means to accommodate the integration of I4.0 concepts into the existing setup (Sanders et al., 2017), considered that I4.0 technologies cannot be bought on the shelf, rather they require the firm to be organized to apply them properly (Porter and Heppelmann, 2015). Moreover, the way in which I4.0 technologies are integrated into existing production systems and organizational contexts is still under investigation (Tortorella and Fettermann, 2017). The literature on this topic is scant, despite it stresses that advances related to manufacturing processes ought to be accompanied by advances also at the organizational and managerial level to avoid significant but isolated gains in process efficiency and product quality (Schumacher et al., 2016). Indeed, making Industry 4.0 applications smart requires a holistic approach and sometimes a change in the business model (i.e. new value proposition, new type of relationship with customers) (Porter and Heppelmann, 2015), which further complicates the development process (Preuveneers and Ilie-Zudor, 2017).\n2. Theoretical background and hypotheses: 2.1 The concept of Industry 4.0\n3. Methodology: Based on the purpose of our study, we decided to rely on primary data that are then analyzed by means of statistical quantitative techniques, namely, a cluster analysis, complemented by a t-test, and a regression analysis. Details of this procedure are explained in the following subsections.\n4. Results: 4.1 Results of the cluster analysis\n5. Discussion and conclusions: The aim of this paper was to analyze the I4.0 behavior of manufacturing SMEs, classifying them according to their level of implementation of I4.0 technologies, and unveiling whether organizational and managerial practices may affect these implementation levels.\n", "role": "user" }, { "content": "Purpose: Currently, the expectancy that surrounds the Fourth Industrial Revolution, commonly referred to as Industry 4.0 (I4.0), is huge. In this context, the purpose of this paper is to unveil whether and how organizational and managerial practices are associated to different levels of adoption of I4.0 technologies.\nDesign/methodology/approach: To reach this aim, the authors carried out a survey involving Italian manufacturing firms. Then, the authors used a cluster analysis and t-test to analyze data.\nFindings: Results show that two clusters of firms based on their level of adoption of I4.0 technologies (high vs low) can be identified. Then, using a t-test, the authors found statistically significant higher levels of a number of organizational and managerial practices for firms with a higher level of adoption of I4.0 technologies.\nPractical implications: This paper contributes to the debate surrounding I4.0 by stressing the organizational and managerial challenges that firms willing to undertake an I4.0 transformation have to face, which goes beyond the sole application of I4.0 technologies.\nSocial implications: Entrepreneurs and managers need to be aware that the path toward I4.0 requires not only focusing on the application of the I4.0 technologies, but also on the development of a series of organizational and managerial practices that become key to face the fourth Industrial Revolution.\nOriginality/value: The authors posit here that I4.0 requires firms to bridge the capability gap, as well as overcome cultural barriers preventing entrepreneurs and managers to change their way of doing business. To this regard, this study highlights I4.0 is an all-encompassing paradigm that involves many dimensions of the firm.\n", "role": "assistant" } ]
Transition from school-based training in VET
[ "Educational inequality", "Vocational education", "Apprenticeships", "Child welfare clients", "Dual system", "Prestigious progression routes" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The dual system of vocational education and training (VET), which combines classroom-based training and periods of apprenticeships, is often praised for effectively integrating young people into the labor market (Noelke and Horn, 2014; Beicht and Walden, 2015). In contrast, vocational secondary education has been criticized for channeling working-class children into the vocational track, and, consequently, excluding them from universities (Shavit and Muller, 2000; Protsch and Solga, 2015). Even though mechanisms of exclusion and inclusion within the VET system are often overlooked, prior research has shown that the transition from school to apprenticeship is particularly difficult, especially for disadvantaged youth (Schmidt, 2010; Brahm et al., 2014). Research on this transition is limited but previous findings show that obtaining an apprenticeship is more difficult for immigrants and students with low school performances (Helland and Storen, 2006), working-class students, and girls (Heinz et al., 1998). This paper examines the probability of dropping out after school-based training for child welfare clients - a particularly disadvantaged group of youth.\nThe Norwegian context: As other Scandinavian countries, Norway is based on a social democratic welfare model. Corresponding to this welfare model, the Norwegian educational system selects different tracks for students relatively late in their school careers (when they are about 16 years old). At this point, students have finished compulsory school education (seven years in primary school and three years in lower secondary school). No student fails compulsory school, and all students have a right to attend upper secondary school free of charge, into which almost all students enroll. In general, about 97 percent of the students in a given cohort enroll directly into upper secondary school (Statistics Norway, 2014). In the child welfare population, 89 percent of the students in a given cohort proceed directly from compulsory school to upper secondary school (Daehlen, 2015a).\nPrevious research and hypotheses: The large numbers of dropouts in upper secondary school have received considerable attention among researchers, policymakers and educators for many years. It is widely recognized that early school leavers are more likely to be unemployed, become welfare recipients and be socially excluded (Freeman and Simonsen, 2015; Backman et al., 2011; Brekke, 2014). Consequently, dropping out of school increases welfare needs and reduces tax revenue, in addition to causing educational deficiencies for those who do not complete upper secondary school, which can limit their economic and social well-being throughout their adult lives (Rumberger, 1987, 2011; De Ridder et al., 2012). Thus, researchers and policymakers want to know what causes students to drop out of school and what may be done to prevent it.\nData and method: The data come from public registers of Statistics Norway and comprise all child welfare clients who received assistance during at least one year from 1990 to 2009. Originally, the data set includes a representative sample of the majority population. However, only child welfare clients and their comparison peers who completed the school-based part of the vocational track in the period 2008 to 2011 in TIP, BC, EE, and RF processing were included. In addition, students older than 23 years were excluded in the analyses[2]. Consequently, students on the academic track, students on other vocational programs, students that have dropped out of school or students who have failed to complete the two years of school-based training, were excluded from the analyses. 4,619 child welfare clients completed the school-based part of the vocational track during this period, whereas the comparison sample comprised 5,916 peers.\nResults: Descriptive statistics\nConclusion and discussion: It is difficult to determine whether the difference in drop-out rates between disadvantaged youth and more privileged youth reflects discrimination from employers or differences in characteristics between the two student groups. It was argued that if the high drop-out rate in the child welfare client population was caused by disadvantaged characteristics, which usually characterize child welfare clients, we would expect relatively equal differences in the probability of transition to apprenticeship and to the alternative educational route between the two groups. In addition, it was argued that the drop-out rate within the child welfare population was highest in the most disadvantaged strata (among child welfare clients with low grades, low educational parental level, and non-western background).\n", "role": "user" }, { "content": "Purpose: This paper assesses the drop-out rate among disadvantaged students within vocational education and training. The purpose of this paper is to examine the probability of dropping out after school-based training for child welfare clients - a particularly disadvantaged group of youth. Child welfare clients' drop-out rate is compared with students from a representative sample of their peers.\nDesign/methodology/approach: Average marginal effects were calculated from multinomial logistic regression models. Data were from public registries (n=10,535).\nFindings: The results show that differences in observed characteristics cannot explain differences in drop-out rates between child welfare clients and the majority peers. It is argued that this drop-out rate is likely a result of employers favoring apprenticeship applicants who are similar to them or that child welfare clients lack networks, which previous research has identified as crucial in finding an apprenticeship.\nPractical implications: The results suggest a need for action targeting disadvantaged youths in the transition that follows school-based training.\nOriginality/value: The paper adds to the very scarce literature on transition from school-based learning to apprenticeships.\n", "role": "assistant" } ]
Governance for health in the Anthropocene
[ "Sustainability", "Health services", "Human development", "Anthropocene", "Ecological change" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: We are entering the Anthropocene, a new geologic epoch that - as its name implies - is being caused by humanity. We have become a force that rivals nature: our activities are causing the climate to change, the Earth to heat up and the oceans to acidify; our pollutants are spread throughout the world and contaminate food chains almost everywhere; we are degrading and depleting key resources such as forests, fisheries, corals, grasslands and topsoils that are central to the well-being and survival of humans and many other species; and we are causing a sixth great extinction.\nWelcome to the Anthropocene: Considering [the] [...] major and still growing impacts of human activities on earth and atmosphere, and at all, including global, scales, it seems to us more than appropriate to emphasize the central role of mankind in geology and ecology by proposing to use the term \"Anthropocene\" for the current geological epoch. [...] mankind will remain a major geological force for many millennia, maybe millions of years, to come(Crutzen and Stoermer, 2000, p. 17).The concept of the Anthropocene has emerged in the past two decades, having been popularized initially by the Nobel Prize-winning atmospheric chemist Paul J. Crutzen and his colleague Eugene Stoermer. In 2008, a group of Earth scientists proposed that the Anthropocene be formally designated as a geological epoch; that proposal is under active consideration by the appropriate international institutes.\nHealth implications of the Anthropocene: It should come as no surprise that these dramatic changes to the Earth's natural systems have health implications; those impacts are already being felt, albeit on a small scale thus far. After all, the most fundamental determinants of our health - air, water and food - are all the products of natural systems, as are other basic needs including materials and fuels we use to build and power our societies. In addition, the stratospheric ozone layer protects us from UV radiation, natural systems detoxify many (but not all) of our wastes, and a largely stable climate for the past 11,000 years has facilitated the emergence of modern humans and modern society.\nHealth system governance in the Anthropocene: Health professionals have an essential role in the achievement of planetary health: working across sectors to integrate policies that advance health and environmental sustainability, tackling health inequities, reducing the environmental impacts of health systems, and increasing the resilience of health systems and populations to environmental change.Report of the Rockefeller Foundation, Lancet Commission on Planetary Health[6] (Whitmee et al., p. 2007).What does all this mean for health systems? There are four areas to consider:\nGovernance for health in the Anthropocene: [...] the health community has a vital part to play in accelerating progress to tackle climate change. To quote UN Secretary-General Ban Ki-moon: \"There is no plan B; there is no planet B\". Health professionals have been at the forefront of social changes, such as those that have gradually made smoking increasingly unacceptable, driving down smoking rates--and saving many lives. I endorse the call for the health community to support the growing movement for a cleaner, more sustainable, and healthier future.Dr. Margaret Chan, Director General, WHO, in commenting on the report of the second Lancet Commission on Health and Climate Change(Chan, 2015, p. 2).The fourth area where the health care system can play a role in addressing the health aspects of the Anthropocene lies beyond the health sector, in the many areas of public policy, private-sector decisions and community and personal choices that are the major determinants of the health of the population. These are not areas where the health sector has jurisdiction, although it can have influence.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to explore the pressing issues facing health and health systems governance in the Anthropocene - a new geological time period that marks the age of colossal and rapid human impacts on Earth's systems.\nDesign/methodology/approach: The viewpoint illustrates the extent of various human induced global ecological changes such as climate change and biodiversity loss and explores the social forces behind the new epoch. It draws together current scientific evidence and expert opinion on the Anthropocene's health and health system impacts and warns that many these are yet unknown and likely to interact and compound each other.\nFindings: Despite this uncertainty, health systems have four essential roles in the Anthropocene from adapting operations and preparing for future challenges to reducing their own contribution to global ecological changes and an advocacy role for social and economic changes for a healthier and more sustainable future.\nPractical implications: To live up to this challenge, health services will need to expand from a focus on health governance to one on governance for health with a purpose of achieving equitable and sustainable human development.\nOriginality/value: As cities and local governments work to create more healthy, just and sustainable communities in the years ahead, health systems need to join with them as partners in that process, both as advocates and supporters and - through their own action within the health sector - as leading proponents and models of good practice.\n", "role": "assistant" } ]
Service innovations enabled by the "internet of things"
[ "Internet of things", "Business modelling", "Intermediating", "Network processes", "Service innovations" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: It is often said, in contemporary economic and political debates, that we live in a \"service society\". It is argued that growth, employment, efficiency, international competitiveness as well as living standards in society depend on service innovations.\nResearch method: We have for some years studied service innovations based on use of mobile phones, especially mobile payments (Andersson et al., 2009, 2011a, b, 2012). For the current IoT project we have consulted survey articles on service innovation (Carlborg et al., 2014; Henneberg et al., 2013) and searched for contemporary articles with titles and keywords including \"service innovation\". The review indicated that a more systemic view on service innovation has emerged (e.g. Maglio and Spohrer, 2008; Ford and Mouzas, 2013). We found IMP related texts on how innovation processes in general should be approached in a network perspective. Hakansson and Waluszewski (2007) and Waluszewski (2011) argued that an innovation must be understood as embedded in several network contexts for the innovation process to succeed.\nService innovation: Already Penrose (1959) argued that the value of physical products depended on their use in the buyer's operations. The service dominant logic (SDL) developed in Vargo and Lusch (2004) takes this idea much further as a perspective on creation of value by exchange of services. Researchers like Maglio and Spohrer (2013) continue on this line of research and argue for a new service science and dominant logic perspective on business models and service innovation. As indicated in a service innovation review (Carlborg et al., 2014), a growing stream of marketing and management researchers have begun to devote interest in the field. Service innovation has begun to be a research topic in its own right, not only a side activity to product innovation. The evolving view during the last decades is to focus on interrelation of service innovation and product innovation. There is a need then also to consider how service innovations include changes of organizational design and the nature of producer-user interactions (Araujo and Spring, 2006) as well new service-based business models in manufacturing industries (Kindstrom, 2010). With reference to the general characteristics of innovations in Waluszewski (2011) service innovations will have to compete with existing value creating services and cope with the different economic logics of \"use\", \"supply\" and \"development\" in the networks.\nIoT: It is often argued that with the IoT we are witnessing the next major step in the evolution of ICT industries and infrastructures. IoT is said to represent a major departure in the history of the internet, \"as connections move beyond computing devices, and begin to power billions of everyday devices, from parking meters to home thermostats\"[1]. EPoSS (2008) provides a short definition stating that IoT is a world-wide network of uniquely addressable interconnected objects[2]. An impressive number of devices will, or potentially may, be connected in the new IoT infrastructure. Objects can communicate regardless of their position in space. Their spatial position, and changes in position, may be observed.\nA conceptual framework: In the research method section we have motivated in detail our selection of variables to be overlapping, intermediating, objectification of actors and business modelling. We will in this section further discuss these variables.\nThe connected vehicle - an evolving service innovation process: The Connected Vehicle case, in which we are involved as academic observers, concerns a fairly new process leading to the emergence of systems in which cars and trucks increasingly become connected to the internet and to wireless networks. In the Methodology section we presented the data collection for the case.\nInterpretation of the case in the conceptual framework: Do we find evidence of relations between the four variables in the model? In IMP research, innovation is to a large extent about trying out new combinations and this often means bridging different types of \"gaps\" (Hoholm and Hakansson, 2012, p. 254). In a similar vein, Bernardi et al. (2012) states that although existing relationships in an innovation process represent an important element of stability that counterbalances the uncertainties in the innovation process, there are situations when innovation emerges in a network by actors with weak or without previous relationships. Furthermore, the innovation needs to be embedded in different network contexts. Hakansson and Waluszewski (2007) distinguish between the developing, the producing and the using network contexts. It is suggested that bridging mechanisms/strategies - bridging established networks in a more limited sense or developing new networks in between - should be in focus for innovation research. In our conceptual framework both intermediating and overlapping are involved in bridging. In our analysis of interaction between the four variables we will use intermediating as the common variable. Below we present five attributes of service innovation processes and formulate propositions based on these attributes.\nSome ideas for future research: We suggest that further research into IoT-based service innovation processes should go deeper into different aspects of the role of intermediating. In this paper we have, first and foremost suggested that service innovation processes as manifested through changes in intermediating, will be strongly connected to processes of overlapping, business modelling and the objectification of actors. Below we propose three research subjects building on the propositions discussed in the previous section.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to develop a new conceptual framework that reflects network dynamics in Internet of Things (IoT)-enabled service innovation processes.\nDesign/methodology/approach: Based on literature on service innovations, business networks and IoT, dynamic concepts are selected. Aided by information about an evolving case \"The connected vehicle\", propositions about interaction between the variables in the framework are formulated.\nFindings: A conceptual framework consisting of four interacting variables: overlapping, intermediating, objectification of actors and business modelling is developed, linking several streams of research. Propositions are motivated and issues for further research questions formulated.\nResearch limitations/implications: The framework may stimulate further research on IoT-enabled service innovations.\nPractical implications: Understanding network dynamics for developing and implementing business models for service innovations.\nOriginality/value: The conceptual framework provides an original contribution to understanding IoT-enabled service innovations.\n", "role": "assistant" } ]
How the exchange relationship affects employee commitment: The mediating role of organizational trust
[ "Organizational commitment", "Organizational trust", "Social exchange", "Economic exchange" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Trust is a crucial element of social relations within an organization. In the interpersonal level, while employees' trust in managers affects their attitudinal, behavioral and performance outcomes (e.g. job attitudes, intentions (Dirks and Ferrin, 2002), satisfaction (Tan and Tan, 2000; Matzler and Renzl, 2006), loyalty (Matzler and Renzl, 2006), enterprising behavior (Costigan et al., 2006), innovative behavior (Tan and Tan, 2000) and organizational citizenship behavior (Wong et al., 2003; Mayer and Gavin, 2005)), trust of managers in their subordinates (Spreitzer and Mishra, 1999; Lamsa and Pucetaite, 2006) and the reciprocity of trust in manager-subordinate dyad (Butler, 1991) positively affect organizational outcomes. In addition to this interpersonal level, employees' trust in their employing organization (hereafter organizational trust) also interacts with various employee-related outcomes in the organizational setting. Building a trustworthy relationship can affect employee satisfaction (Deery et al., 2006), organizational identification (Edwards and Cable, 2009), intention to stay (Tan and Tan, 2000; Montes and Irving, 2008), cooperative relationship (Montes and Irving 2008), organizational citizenship behavior (Restubog et al., 2008) and organizational commitment (Tan and Tan, 2000).\n2. Trust in the organizational setting: Trust is defined in many different ways in the literature. These definitions can be categorized based on various personal traits (attitudes, beliefs, behaviors and dispositions) or in response to diverse referents (trust in someone, something or specific characteristics) (McKnight and Chervany, 2001) and both of these variables affect the antecedents and outcomes of trust (Dirks and Ferrin, 2002). Since defining trust is important to figure out the nature of research, most scholars \"tend to develop narrow conceptualizations of trust that fit the type of research they do\" (McKnight and Chervany, 2001, p. 37). To avoid this common pitfall, the current study analyzes trust in the organizational setting by capturing its overall meaning and distinguishing it from the similar forms.\n3. Making employees feel trusted: social and economic exchange: According to Whitener et al. (1998), since trustor's perception arises based on his or her observation on trustee's behavior, it is the management's responsibility to initiate a trusting relationship with employees through providing the necessary actions and behaviors. Social exchange theory (Blau, 1964) helps to understand this mechanism with providing the role of trust in an organization setting (Aryee et al., 2002; Konovsky and Pugh, 1994). The theory is built on the idea that a person who helps another person expects a return and gratitude in return for this favor; if his or her help is properly reciprocated by the other person, this social reward will result in further assistance (Blau, 1964). This \"mutual exchange of services\" will create a social bond between two parties: \"giving a present or doing a favor demonstrates trust in another; the other's reciprocation validates this trust as justified\" also (Blau, 1964, pp. 4, 107). Therefore, people tend to trust in each other depending on what they give and what they obtain from this relationship (Fulmer and Gelfand, 2012).\n4. Research methodology: 4.1 Measurement\n5. Discussion of findings and limitations: The findings of study showed that organizational commitment positively relates with the social and economic exchange relationship in between employees and employer organization. These results are in line with the findings of previous studies in the literature. As a contribution of this study, these findings also support the conceptualization of social and economic exchange in a wider sense and the study provides reliable scales for each dimension. In the current study, the social exchange is not reduced into the socio-emotional relations between employee-organization dyad; it was conceptualized as an aspect of how an organization responds to major social challenges such as diversity, justice, social responsibility and transparency. In a similar vein, the economic exchange is also defined in a wider sense by recognizing how organizations manage the issues about structure, policies, culture and leadership. In addition to these findings, as the focal question that compels this study, organizational trust was found as the mediator in both exchange models. This result showed that trust at the organizational level should be also taken into account when designing the social and economic exchange relationship with employees.\n6. Conclusion: According to Jeffrey Pfeffer, a Professor of Organizational Behavior at Stanford University, our modern business organizations have become shockingly inhumane due to the problems such as increasing stress, layoffs, longer hours, work-family conflict or economic insecurity; Pfeffer states that \"companies should care about what they are doing to the social environment, not just the physical environment\" (Walsh, 2018). Although building a trustworthy relationship between employees and their organizations is seen as one of the most traditional way of addressing some of these social problems, we still have a limited understanding on its underlying mechanism at organizations. The current study attempts to address this void by taking an exchange perspective at organizations. In doing so, it paves the way of taking a holistic view when defining the social and economic exchanges at organizations by involving wide-ranging issues from social responsibility to leadership.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to analyze the impact of social and economic exchange relationships on organizational commitment in line with the mediation effect of organizational trust.\nDesign/methodology/approach: In order to test the hypotheses of the study, a survey was conducted on a sample of 213 employees who were working at manufacturing companies in Turkey.\nFindings: The findings of the study reveal that both dimensions of the exchange relationship positively affect organizational commitment and these links are mediated by organizational trust.\nResearch limitations/implications: The study provides reliable scales to measure the social and economic exchange relationship between employees and employing organizations. Although the sample of the study was relatively small and drawn from a single country, the Cronbach's a values of scales were obtained above the recommended threshold value.\nPractical implications: Organizational leaders might adopt an exchange perspective to build a trustworthy relationship with their employees. Developing such a mindset is very important at an employment structure, which has become highly flexible and contingent during the last decades.\nOriginality/value: The study attempts to distinguish the twofold nature of the exchange relationship in organizations based on a theoretical model to reveal the impact of each dimension on organizational level outcomes in conjunction with the mediating role of trust. In doing so, the study contributes to the literature by incorporating social and economic exchange in a holistic view as well as defining each dimension in a broader sense by including some employee-related challenges of business organizations such as diversity, social responsibility, leadership, ethical culture and so on.\n", "role": "assistant" } ]
Objectives priority in university strategy map for resource allocation
[ "University", "DEMATEL", "Strategic planning", "Balanced score card", "Strategy map" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Diverse changes in universities' environment such as globalization, increasing information technology, internationalization of education and research collaborations emerge the essence of competitive advantages for universities. In addition since country vision emphasizes on superiority in education and science regionally and globally, each university needs to enhance the quality of education and overall academic standard. To promote university educational quality they require performance measurement which encourages universities to improve its advantages and become a reference for student (Chen et al., 2006, 2009). Universities as non-profit organization do not face pressure of survival but they should focus mostly on their mission, vision and performance management much more than financial ratios. The primary objective of non-profit performance management system is to determine how well an organization fulfills its mission (Grigoroudis et al., 2012). Furthermore, government stress on thorough performance evaluation and ranking among domestic universities.\n2. Concept of BSC and strategy map: In economic age companies succeed by benefiting from economies of scale and scope. During those years financial control systems are developed in companies. In information era in last decades of twentieth century excellence merely in financial indicators is not useful anymore. In this era companies require a control system which converts intangible assets to tangible outcome (Kaplan and Norton, 1996). Traditional performance measurement systems try to control behavior, on the other hand BSC put strategy and vision, not control, on the center. It just puts goals and leaves adoption of the required behavior and actions to reach these goals to employees (Kaplan and Norton, 1992). Frigio and Krumwiede (1999) found that performance management systems using BSC are significantly more effective comparing to others.\n3. Methodology: In this research, Ferdowsi University of Mashhad (FUM) is selected as the example. The proposed framework to construct strategy map from BSC is illustrated in Figure 2. First we define four perspective based on literature review which includes financial, stakeholder, process and learning and growth perspectives. Then 11 objectives defined in FUM strategic plan divided into 20 to fit into four perspective of BSC. The appropriate place of each objective among BSC perspectives is obtained from literature review and expert opinion while mission and vision are on the top as in Figure 3. After that, a casual relationship is applied to priorities objectives using DEMATEL method. Finally the strategy map is developed based on DEMATEL analysis result. DEMATEL method is expressed in Section 3.\n4. Results: This section includes measurement of relationship between FUM objectives. Most important objectives are emerged as the result of this method.\n5. Conclusion: In this research we utilize DEMATEL method to determine causal relationship among objectives of FUM. Using DEMATEL as a structural decision-making model to develop the strategy map of university and prioritizing its objectives is the distinction of this research from previous studies. Those studies develop their strategy maps only by expert's rule of thumb. Through DEMATEL method the main cause and effect objectives of university is determined.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to provide a model for Ferdowsi University of Mashhad (FUM) to prioritize its objectives. This will be achieved by providing strategy map using balanced score card (BSC) method.\nDesign/methodology/approach: This research result is based on 21 managers' opinion about relation among university objectives, gathered by questionnaire. DEMATHEL method which is a structural decision-making model is used to process data in questionnaires and discover casual relationship between university objectives.\nFindings: According to this research result \"Having productive competent human resource that are committed to Islamic values and professional ethic\" has the highest priority among FUM objectives while \"Achieving educating excellence especially in graduate programs and research\" has the lowest.\nPractical implications: FUM managers need to focus on their human resources and enhance their competency in order to achieve academic excellence.\nOriginality/value: Achieving superiority in university competitive position based on its education and research activities is FUM vision. As a non-profit organization due to resource restriction and environmental circumstances it has to fulfill this goal with higher productivity. BSC as a performance management system will help it to achieve this goal by translating vision into objectives and defining casual relationship between them. This method is rarely studied in the context of universities especially in Iranian universities. This research applied strategy map, one of BSC implementation stages, in a public university in Iran to illustrate series of objectives that leads to mission accomplishment.\n", "role": "assistant" } ]
Validity and reliability: the Iranian version of Lawton IADL in elderly community dwellers
[ "Mental health", "Validation", "Elderly community", "Lawton IADL" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nBackground: The normal aging process leads to decreased functional abilities, this process is accelerated by acute illness, advanced chronic disease and long stays in hospital, resulting in less independence and decline in physical and mental abilities, with daily activities becoming much harder for the elderly (Graf, 2009; Graham et al., 2008). The scale of ability that includes advanced activities, Instrumental Activities of Daily Living (IADL) and the basic Activities of Daily Living (ADL) can show the highest and lowest levels of daily activities in older people (Vergara et al., 2012). IADL tasks are not physically more complicated than ADLs (Lawton et al., 2003). IADLs are more neuropsychologically driven and are necessary to maintain individual independence in the community (Halter et al., 2017).\nMethods: Participants and at a collection\nResults: Of 422 participants referred to outpatient clinics, the majority were male (62.8%), with average 66.9 years old (SD = 5.95). A total of 29.2% of participants had a higher educational level. Ambulation and locomotion scores showed 85.8 and 86.7% of participants were independent, respectively (Table 1). The total average IADL scores found in both males and females, according to comorbidities; stroke 4.97, Parkinson 5.36 and dementia 6.29.\nDiscussion: In this validation study, significant correlations in the performance of IADL items were observed, with other concurrent measurements, such as the level of ambulation, locomotion, FAAM-ADL and SF36-PCS; although there was a weak correlation observed with the mental and cognitive status; due to having cognitive impairment as an exclusion criterion. Age and gender were not observed to affect in the validity of the IADL Iranian version.\nConclusion: The Lawton IADL scale has great construct validity and has high indicator reliability value. This tool culturally conformed to the Persian language, which was an important factor.\n", "role": "user" }, { "content": "Purpose: Normal aging changes, acute and chronic illness, and the long stay in the hospital made the decline in elderly physical and mental abilities in non-instrumental and Instrumental Activities Daily Livings (IADL). This paper aims to determine the validity and reliability of Lawton IADL in the Iranian elderly community relevancy to physical, mental and cognitive measurements.\nDesign/methodology/approach: A cross-sectional study with 422 participants, age 60 years old and over selected by face-to-face interviewing from the geriatric clinic, Tehran medical university from January to June 2019.\nFindings: Of 422 participants referring to outpatient clinics the majority were male (62.8%); with an average of 66.9 +- 5.95 years old. Reliability was significantly high (Cronbach's alpha = 0.96, p < 0.001). There was a significant relationship between IADL with Short-Form Health Survey36-Mental Component Score (SF36-MCS), with p < 0.01. Reliability was significantly high (Cronbach's alpha = 0.96, p < 0.001). In the factor analysis with eigenvalues more than one, two components were found; that the first factor named as observational daily activities, whereas the second factor named \"advanced daily activities\".\nResearch limitations/implications: The limitation of this study was the low number of similar studies to compare the results and the stronger discussion.\nOriginality/value: Determining IADL's dependency in the community elderly is important to maintain their self-care manages. More studies are needed to manifest the relationship between mental health and IADL's independence. Therefore, validation in different settings is important in planning for the geriatrics team.\n", "role": "assistant" } ]
Robert Owen: a historiographic study of a pioneer of human resource development
[ "Robert Owen", "Human resource development", "Historiography", "Work reform", "Social reform", "Communitarianism", "Utopianism", "British industrial revolution", "British factory system", "Management" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: As post-Colonial America continued the practice of indentured servitude as its \"original labour force\" and managed labor through ownership of \"bound white servants\" (Galenson, 1981, p. 32), back in Scotland the nineteenth century industrialist Robert Owen (1771-1858) was creating worker-oriented factory villages with humanitarian oversight.\nPurpose: To bring history to the forefront in HRD research and address how historical figures inform current practices and concepts, I began to study Robert Owen in 1999. I set out to examine his and others' historical writings and artifacts to see if there was a relationship between his ideas and endeavors and the human resource concepts upon which contemporary workplaces are developed. This historiographic study is an effort to provide evidence of Robert Owen as a pioneer of HRD.\nRationale and significance: The rationale for this research is situated within a historical framework of workplace reform and the development of human resources. Knowledge of workplace reform and how it is influenced by contexts is critical to develop ideas and interventions that make positive changes on employees, leadership, organizations and society. Workplace reform can also inform better relations between workers and management. Its basic tenets match those of modern human resource development, namely cooperation and communication between management and workers, skills development, diversity and quality of work life.In studying Robert Owen, the relationships between the nature of work and workers and between workers, managers and business owners and the role that society plays in workforce development and workplace reforms help us make comparisons across time that can \"illuminate the present by highlighting both what is recurrent and what is new, what is durable, transient and contingent on our present conditions\" (Brown, 1991, p. 1). As Bedeian (1998) said: \"A discipline [HRD] that downplays or ignores the cumulative impact of the past on its present fails to fully exploit the explanatory and interpretive potential of understanding how and why present theories and methods have their particular nature by virtue of the past\" (p. 4).This lack of a historical perspective contributes to the inability of HRD to define itself (Lee, 2001) and to establish itself as a discipline (Hatcher, 1998). \"Collective historical memory serves as the basis for establishing the identity of a discipline\" (Carson and Carson, 1998, p. 38).The present study is significant in a number of ways. First, it is one of only a few empirical studies of Robert Owen as a pioneer in any work-related position (Butt, 1971). Next, only a handful of historical studies of individuals related to HRM or personnel management were located. The study is significant because only one publication was found on the origins of HRD, namely Swanson's (2001) work on Channing Dooley's contributions to training during the Second World War, and finally an extensive search revealed no human resource development studies using a historiographic research method.This study of Owen is also significant in helping HRD and related professionals value what they are trying to create in modern workplaces through insights into what has come before. Historical figures and their contexts can and should play a role in how people are treated and developed within contemporary workplaces.\nConceptual foundation and method: The conceptual foundation for the present study suggests that historiographies can inform the present; \"individuals do significantly affect the course of events\", both contemporary and in the future (Lockyear, 1984, p. 46) through reconstruction. Reconstructionism was the epistemological foundation of the study. Characteristics of reconstructionism include the \"building blocks of validated sources, the origin and context of each source, how the source is relevant to the research, and that interrogation [albeit post hoc] reveals truth\" (Booth, 2005, p. 464).Theoretically, I counter the postmodern criticism of work-related history as unsupportive of \"emancipatory discourse\" (Jenkins, 1999) through knowledge that (Popper, 1972) allows me to better understand the \"real world/workplace\" by providing historical evidence that informs the emancipatory characteristics of the modern workplace such as cooperation and employee development, attributes of Owen's approaches to the workplace, workers and society.Historiography was chosen for this study because it provides a way for the historian/researcher to understand the actor as well as the context of the action. Carson and Carson (1998) added:Historiographers strive to disentangle [...] motives through the difficult and somewhat subjective technique of empathetic reconstruction [...] which requires [them] to enter the minds and understand the feelings of the individual (p. 31).Like many research methods, historiography begins with a general question, such as \"Is there historical evidence to establish Robert Owen as a pioneer of HRD?\". To answer this question I examined primary sources including eye-witness accounts by contemporaries or seminal publications by Owen and others. Secondary sources provided views of journalists or scholars who later reconstruct the story (Handlin et al., 1970). Each source was examined, authenticated, analyzed and its suitability as a reliable source was evaluated (Handlin et al., 1970).In the present study, in conjunction with an extensive literature, I examined numerous primary and secondary sources during my visits from 1998 to 2005 to New Lanark and Edinburgh, Scotland and New Harmony, Indiana and visits to Ireland and Wales. I visited organizations known to hold Owen's works, including the The Workingman's Institute and Library at New Harmony, IN, and the Owen Memorial Museum in Newtown, Wales[1].As I discovered ideas/concepts in one source I then sought confirming and/or disconfirming evidences from other sources. As an example, an article in the Glasgow Herald described worker's responses to Owen's return to his New Lanark mill after a long absence traveling abroad: \"Mr Owen is so justly beloved by all the inhabitants employed at New Lanark\" (Glasgow Herald, January 10, 1814). To confirm, I located discussions of this event in several secondary sources such as Donnachie (2000) and Duncan (2004): \"Men, women and children were chanting his name [...] shouting their gratitude\" (p. 210). Many of my early interpretations were critiqued by the Honorable Hatton Davidson, past Curator of the Robert Owen Memorial Museum located in Newtown, Wales.The next section discusses the many contextual influences on Owen. It provides an understanding of how he developed his thinking and drove his actions.\nContextual influences: With few exceptions, HRD-related research discounts the historical contexts in which people live and work. This is unfortunate, because contextual factors such as economics, politics and societal demands greatly influence workers and workplaces. Wren (1987) suggested that an individual's \"contributions can only be evaluated in light of the context of the times\" (p. 341). HRD professionals can and should learn by understanding the connections of contexts over time. A good example is a comparison between the impact that the growth of technology during the British Industrial Revolution had on the workplace with the technological explosion within US and global organizations of the late twentieth and early twenty-first centuries.To better understand Robert Owen within the context in which he lived and worked, it was necessary to examine biographical, historical and socio-economic contexts of the British Industrial Revolution for the period 1790-1850 as the backdrop against which his philosophy and principles were cultivated.Early life\nLinking Owen to contemporary HRD: The following is a discussion of the themes that emerged from the present research that specifically connects Owen's ideals and actions with modern concepts of human resource development. The themes are:* managing people and profit;* education and training;* pioneering workplace innovations; and* from factory owner to \"social father\": new harmony and beyond (see Table I for a summary).Managing people and profit\nDiscussion and conclusions: Historiographers draw conclusions not from empirical data and statistical processes but from \"impressionistic analysis where conclusions are more on narratives that capture mood, feelings, underlying thoughts and motivations\" (Topolski, 1991; Carson and Carson, 1998, p. 31). This research highlighted the ideas, thoughts and work of Robert Owen, whose \"application of certain principles of business conduct, of education and human organization, were startlingly modern\" (Gorb, 1951, p. 128). The innovations he cultivated and implemented provide evidence that a focus on workers' moral character and community environment coupled with humane management leads to more productive and \"happier\" workers and families. He demonstrated that social ills can be addressed through \"benevolent leadership\", work-based social innovations and community-based education. Owen's writings and actions suggested that he understood education and training were keys to pull workers from destitution and forced slavery, to create a new community-based work environment and thus to contribute to economic and social conditions (Gorb, 1951).Owen, like few of his contemporaries operated on the principle that the working conditions/environment of employees should not be compromised for the sake of profit. \"Few masters of business affairs concerned themselves with the welfare of the people or with the solution of social problems as did the cotton spinner of New Lanark\" (Branigin, 1972, p. 18).Although a workplace and social reformer, Owen was a contemporary of the paternalism practiced by industrialists of the time portraying him as unable to have a personal and reciprocal relationship with his workers \"whose lives he wanted passionately to improve\" (Anderson, 2002, p. 427). As an aristocratic factory owner he was expected to be detached from the lower classes.While there is some evidence that Owen did not associate himself much with his workers and saw them in a detached, fatherly way, he admonished his fellow industrialists for referring to workers as \"bad, worthless and wicked\" (\"Address to the inhabitants of New Lanark\"; Owen, 1816a, b).Another way to see Owen's relationship with his workers is as a precedent of organizational culture and socialization. In his Third Essay (Owen, 1814) he observes that people are \"trained and formed by its leading existing, and the character of the lower orders in Britain is now formed chiefly by trade, manufacturers, and commerce\" (Anderson, 2002, p. 426). In \"A New View of Society\" (Owen, 1816a, b), he stated:To remedy this evil, not one legal punishment was inflicted, not one individual imprisoned; but checks and other regulations of prevention were introduced; a short plain explanation of the immediate benefits they would derive from a different conduct was inculcated by those instructed for the purpose [...] Those employed became industrious, temperate, healthy; faithful to their employers, and kind to each other, while the owners received benefits far beyond those which could be obtained by any other means than those of mutual confidence and kindness.And even if he was less than \"democratic\" as a leader, he was within the political debates of the time among those who pressed the hardest for civic and human rights for the lower working class (Tsuzuki, 1992).The present study identified Owen's visions and actions to achieve greater autonomy and justice in the workplace, and the optimistic belief that this goal and more can be achieved through the creation of a positive work community. Owenism, the notion that humans can find ideal solutions to the complex problems of society and the workplace, has been supplanted by labels such as \"visionary\" and \"socially responsible\", accolades and monikers for leaders who work for the betterment of organizations and society (O'Toole, 1996).\nImplications for human resource development: This original research contributes to HRD research and practice in four ways:1. It provides historical support for the ideals of Owenism as part of the history of modern workplace concepts such as diversity, equality and learning.2. Robert Owen (1771-1858) pioneered workplace reforms and workplace justice and social responsibility, common ideas in modern workplaces and management concepts. Owen believed that all people are educable and have potential \"Not only business but society at large was to regard the working man as a person who could be developed to a point where he could contribute to the improvement of economic and social conditions. In his experiment with education associated with work, Owen anticipated a concept which only in recent times has been accepted by managers as part of his approach to administration\" (Gorb, 1951, p. 144). Owen believed in and experimented with the idea that the workplace should be part of, not separated from the community. The applicability of Owenism, communitarianism and to a lesser extent utopianism, has inspired leaders to visualize and promote a humane and equitable workplace and focus on human potential versus human chattel. The common thread in these perspectives is the desire to achieve greater autonomy and human rights in the workplace, and the optimistic belief that this goal can be achieved through development of human resources and related workplace reforms towards a new kind of organization where financial success is a result of developing human potential.3. Historiography as a research method has potential for HRD research. It should be considered a supplement to more traditional research techniques, especially in areas where methods are required to assess multiple causes from multiple time-bound contexts. Indeed, researchers may benefit for the creativity, skepticism, and reticence demanded by historiography (Goodman and Kruger, 1988).4. Finally, on tracing the origins of our current thinking on matters of humane workplaces and developing human potential on the job, we can see that Owen's work and ideas as a businessman and workplace reformer an early attempt to integrate a set of social and ethical ideas into a socio-workplace framework that linked the development of human resources with social reforms.The evidence cited herein suggests that Robert Owen implemented and enhanced many of today's HRD-related activities and concepts. He was one of the first recognized business leaders to focus on employees and their well-being and the relationship between work and society. Thus, Robert Owen should be considered and accepted as a true pioneer of human resource development.\n", "role": "user" }, { "content": "Purpose: - The purpose of this paper is to investigate the ideals and activities of the nineteenth century Welsh industrialist and reformer Robert Owen (1771-1858), and how they informed modern human resource development (HRD) concepts and practices and provided evidence of Owen as a HRD pioneer.\nDesign/methodology/approach: - Historiography provided a method to understand how historical figures, and the context in which they lived and worked, inform contemporary research and practice.\nFindings: - Contextual factors of economics, politics and societal demands and the influences of Owen's early life, his immersion within the British factory system and the creation of the New Lanark mill village, Owen's great work experiment, revealed a strong impact on his thinking and actions. Thematic findings included: managing people and profit, education and training, pioneering workplace innovations, and the failure of the New Harmony, Indiana community. Themes provided unique historical evidence that education and development of workers, and the creation of humane work and community environments are linked across time and contexts to modern concepts of human resource development and thus supported Owen as a HRD pioneer.\nPractical implications: - Understanding the ideals and workplace experiments and contextual influences on a historical figure such as Robert Owen illustrate how modern concepts of workforce training and education, diversity, equality and justice and social responsibility originated and the importance of contexts on their development and success.\nSocial implications: - Contexts of economics, politics and societal demands greatly influence organizations and the creation of humane workplaces that nurture human potential.\nOriginality/value: - The study brings history and historiography as a research method to the forefront of HRD research and practice. The study provides the beginnings of a collective historical memory that can contribute to HRD defining itself and establishing its identity as a discipline.\n", "role": "assistant" } ]
A historical approach to realistic job previews: An exploration into their origins, evolution, and recommendations for the future
[ "Recruitment", "Human resources management", "Realistic job previews" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Throughout the history of organizations there has been a need to exchange information about the content of the work that needed to be performed. This information has been essential in helping workers to be able to successfully prepare for both the organization and the specific duties of the employment position. In order to help ensure a better fit between the organization and the employees, this exchange of information should be conducted during the recruitment process and prior to a mutual selection by the organization and employee into an employment contract.\nRealistic job previews: Research in realistic job previews has been conducted for over 50 years (Breaugh, 1992). Phillips (1998) noted that realistic job previews have garnered the greatest level of attention of all recruitment issues. A realistic job preview (RJP) is defined as a technique that presents job applicants with a \"realistic\" view of what they should expect from the organization. To do so, the organization provides job applicants with \"all pertinent information without distortion\" (Wanous, 1980, p. 37) which is both positive and negative (Ganzach et al., 2002; Wanous, 1973, 1977, 1980).\nPsychological processes underlying RJPs: The question of why RJPs work was first investigated by Wanous (1973). His \"met expectation\" hypothesis, also known as the \"reality shock\" model, states that RJPs lower the initial job expectations of new employees to a level that is more likely to be congruent with what is actually encountered on the job (Miceli, 1985; Wanous, 1977). In doing so, he found support for the claim that individuals whose expectations are met are more likely to be satisfied with the job, and less likely to leave voluntarily. Zaharia and Baumeister (1978) interviewed departing employees and found that the misalignment of initial employee expectations with the actual job situation was a key source of dissonance.\nHistorical evolution of the realistic job preview research: Weitz and Nuckols (1955) provided the first evidence that employees can possess unrealistic expectations about the nature of work they perform. Although primarily interested in studying the relationship between employee satisfaction with specific aspects of their jobs and its impact on performance (operationalized as job survival), Weitz and Nuckols found an interesting pattern of results. Specifically, the authors generated a list of 58 satisfaction and 82 dissatisfaction items for inclusion in a questionnaire and then distributed the instrument to 2,710 insurance agents. A large percentage of the sample marked two items that were not part of the initial hypotheses. Agents appeared to be highly dissatisfied when there was a misrepresentation of their employment contracts or of their job requirements and advancement possibilities (Weitz and Nuckols, 1955). The employees' disillusionment with their work, then, may have resulted from misrepresentation of the job during the recruitment phase. Further, the authors discovered that those agents who claimed that their managers misrepresented the job or job possibilities during the interview were more likely to voluntarily quit than were those who did not agree with these statements. This finding inspired the authors to speculate about the content of recruitment communications, most of which are likely positively skewed by the recruiters.\nMixed results of realistic job previews: Despite the logical relationships between realistic job previews and outcomes such as job satisfaction and reduced intentions to quit, results have often been mixed, with many studies finding either weak or no relationships (Breaugh, 2008). For example, two early meta-analyses (Premack and Wanous, 1985; McEvoy and Cascio, 1985) found differing results. Premack and Wanous (1985) found that realistic job previews had an average effect size of 0.06 on job survival which included very little between-studies effect size variance. McEvoy and Cascio (1985), however, found the average effect size for the same relationship of realistic job previews and job survival to be 0.09 with significant between-studies effect size variance. These conflicting results were addressed by Wanous and colleagues (Wanous et al., 1989) in their important work in outlining the numerous judgment calls that must be made when conducting a meta-analysis. In doing so, the authors found that, in addition to only using 11 of the 20 studies from the Premack and Wanous (1985) meta-analysis, McEvoy and Cascio (1985) also used two additional studies that were highly significant outliers which greatly impacted their results. Another study found that realistic job previews negatively impacted the decision by European graduate students who were being mentored to accept a job offer at their mentor's organization (Spitzmuller et al., 2008). Specifically, the study found that realistic job previews may have a suppressor effect (Conger, 1974) by decreasing the intentions of the protege to pursue employment opportunities with the organization.\nThe future of realistic job previews in research: Realistic job previews (RJPs) have been studied for over half a century (Breaugh, 1992). With several organizational benefits such as decreased turnover, increased job satisfaction, and stronger organizational commitment, RJPs offer many advantages to organizations and warrant its place as a central component in the research of organizational scholars. To date, scholars have found RJPs to be effective in different environments, contexts, formats, and organizations. Undoubtedly, RJPs will continue to be studied and implemented for years to come both in testing the theoretical conceptualizations posited in prior work and in advancing the empirical studies into dynamic new areas. Based upon our examination of the historical lineage of the RJP research presented in this paper, we suggest the following important areas of research still waiting to be discovered.\nThe future of realistic job previews in practice: There are numerous ways in which the impacts of realistic job previews are yet to be explored that may serve as direct benefits to organizations. Internships are one such possibility for reconceptualizing the traditional RJP. Internships provide students with work experience that is both structured and career relevant (Taylor, 1988). Due to the interns' education and organization-specific knowledge, they are ideal job candidates (Zhao and Liden, 2011). Additionally internships, like temporary workers or the use of probationary hire periods, allow for observation and assessment of the performance an individual can actually do, not just what they say they can do in an interview (Klehe and Anderson, 2007). The use of internships and similar forms of pre-hire assessments offer two main benefits. First, the organization is able to better assess the individual's actual performance with employees interacting with and evaluating the individual. Second, the individual is better able to determine his or her fit and ability to accomplish the required tasks. As such, unrealistic expectations will be reduced and the intern will be in a better place to assess fit within the A-S-A framework (Schneider, 1987) and self-select out if appropriate. By utilizing multiple interns, the organization can implement a real-options approach (McGrath, 1997, 1999) by reducing the costs of recruiting, selecting, and training an unqualified new employee. In doing so, the organization is able to hedge against the downside risk while maintaining the upside of increased performance. Should the organizational leaders decide to offer employment to a fraction of the interns, they will be able to make a more informed selection decision by comparing the interns' performance and eliminating the underperformers.\nConcluding thoughts: While the refinement of realistic job previews has evolved a long way, there is still a need to provide a socialization experience to workers that adequately prepares them for what they might experience on the job. History has shown that providing newcomers with realistic information can have numerous beneficial effects for an organization in terms of performance, turnover, job satisfaction and a number of other important organizational outcomes. As such, organizations should actively promote the use of both RJPs and ELPs within the hiring and orientation processes for new employees. Further, it is our hope that recruitment researchers will benefit from our review of the extensive realistic job preview literature, the discussion of the psychological processes underlying RJPs and the mixed results, as well as our recommendations for future research.\n", "role": "user" }, { "content": "Purpose: The aim of this paper is to provide some historical understanding of a popular recruitment procedure called a Realistic Job Preview (RJP). As long as individuals have worked for others there has been a need to exchange information about a focal job. Information can be exchanged through myriad channels. The aim here is to trace the origins of RJPs and discuss the initial studies that generated attention and interest in what has become known as \"realistic recruitment\".\nDesign/methodology/approach: Along with a historical account, this paper provides a summary of the limitations associated with this method, proposed psychological processes mediating effectiveness of RJPs, and issues with development, mode of presentation, implementation of RJPs, and an important alternative/accompanying technique (ELP).\nFindings: While this technique has been used for many years, it will continue to be a quality addition to any worker socialization program.\nOriginality/value: The value of this paper is that it places this technique in an historical context.\n", "role": "assistant" } ]
Self-gifting guilt: an examination of self-gifting motivations and post-purchase regret
[ "Consumer behaviour", "Retailing", "Post-purchase regret", "Self-gifting" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\nIntroduction: Gift giving motivations around ritual occasions, particularly Christmas giving, are a well-researched area (Clarke, 2006, 2008). However, self-gifting is a performative process because these actions relate to an individual's inner thoughts and comments which act as both predictors and outcomes for particular or potential purchase decisions. Hence, self-gifting differs from traditional gift purchases because the requirement of asking others about gift preferences is eliminated. Previous research has not examined the antecedents of self-gifting behavior and the effect on post purchase regret. Shoppers often face many purchase decisions during half yearly, Black Friday or post-Christmas clearance sales events. One such decision is to purchase items for themselves (Mukhopadhyay and Johar, 2009). These purchases are considered self-gifts and constitute a form of hedonic shopping and personal indulgence (Mick and DeMoss, 1990). In the context of shopping during sales periods, self-gift purchase decisions may emerge from hedonic shopping and indulgent behavior that contributes to self-gift behaviors and possibly regret. In the context of mid-year clearance sales in Australia, the purchasing motivations of consumers often move between buying for others, household requirements and purchasing gifts for themselves. Yet, little has been done to understand or draw linkages between the established antecedents of self-gifting shopping behaviors and the specific contexts of self-gifting motivations. The multi-dimensional conceptualization of self-gifting behavior (Mick and DeMoss, 1990) helps to examine the relationships between hedonic and indulgent trait shoppers, the types of self-gifting during these clearance sales and the extent of post-purchase regret. Thus, this research contributes to gifting behavior theory and literature.\nSales events and gift shopping: Shopping is perceived as a recreational pursuit born of a consumer culture that maintains conspicuous consumption, frivolous spending and demands for instant gratification (Betts and McGoldrick, 1996). Our senses have a powerful effect on the physical, cognitive, social and emotional aspects of the shopping experience (Gentile et al., 2007). The appropriate sensory stimuli can calm, relax, de-stress, energize, improve mood, influence decision making and hence control the propensity to spend (Soars, 2009). However, the atmosphere and general ambience generated by other shoppers creates an increased energy, enthusiasm and excitement about the sales event (Martin, 2012). There is a thrill, an excitement and the madness of a dream; a dog-eat-dog fear of missing out (Betts and McGoldrick, 1996). Therefore, the more sensory and exciting the experience, the more engaging it will be (Soars, 2009).\nHedonistic shopping and indulgence: The hedonic shopping experience is personal, subjective and reflects an emotional or psychological worth of the purchase through values such as the joy from excitement of shopping, or the escape from everyday activities that the experience offers (Carpenter and Moore, 2009). Satisfaction-seeking fulfills a biological need whereas pleasure-seeking and hedonism co-exist in the mind of the consumer, and aims to increase the \"quality\" of the experience (Campbell, 1987). Shoppers anticipate the pleasure that sales generate; the novelty of a new product; the thrill of the chase and gaining a bargain. Thus, hedonic consumers shop because it \"feels good\" and make no attempt to consider the full consequences of action through regret or post purchase dissonance.\nSelf-gifting motivations: Heath et al. (2011) indicates that the frequency and increased consumption during clearance sales, together with the shopping appeals and communication media, creates a shopping environment that encourages self-gifting behavior. Accordingly, most clearance sales present a valid context in which shoppers would expect to engage in extensive self-gifting. Self-gifts can be spontaneous or planned, cognitive or affective. However, self-gifts are controlled, connected mentally to an occasion or context and are particular or singularly relevant to the purchaser (Mick and DeMoss, 1990). Furthermore, the type of self-gift depends on an inspiring reward or a motivation because buying a self-gift is \"a kind of self-love\" (Heath et al., 2011). Other categories include relaxing, therapeutic self-gifts and thoughtful birthday self-gifts. Such purchases reflect a variety of motivations that cover rewards, celebrations or simply being nice to oneself and fulfilling certain \"felt\" needs. Initially, Mick (1996) designated four categories of self-gift occasions. There was PuritanicSG that equates to SGReward; RomanticSG elements reflect the SGHedonic; HolidaySG now incorporates the SGCelebration category while therapeuticSG emerges unchanged. It is therefore evident that the motivations and the decision-making process to self-gift varies and consequently leads to cognitive fatigue and possible post-purchase regret.\nPost purchase regret: There are two aspects of post purchase dissatisfaction. The first concerns post purchase dissonance and the second is post purchase regret. The seminal work of Festinger (1957) argues that cognitive dissonance refers the pressures of non-fitting consumption relationships and ultimately requires changes in behavior, cognition and exposure to new information and new opinions. Reduction in post purchase dissonance occurs through actions of the provider or others where a more favorable attitude or evaluation from the customer is a goal (Hunt, 1970). Initial regret theory focused on irrational decision making (Loomes and Sugden, 1982), but developed as negative emotion theories where the present situation would be acceptable if different choice decisions prevailed (Zeelenberg and Pieters, 2004). Also, Sweeney et al. (2000) refer to regret as an individual's psychological discomfort subsequent to the purchase decision. Post purchase regret involves coping and risk minimization strategies that are more personal and generally involves self-judgment and feeling sorry (Watson and Spence, 2007). Consumer post purchase regret is a frequently felt emotion when a decision outcome compares unfavorably with other possibilities, but not the final choice (Inman and Dyer, 1997). Similarly, there is a state of feeling sorry due to the experience of misfortunes, limitations, losses, short comings, transgressions or mistakes (Landman, 1993). Previous research of antecedents of post-purchase regret have considered the effect of product type or demographic characteristics (Ali and Ramay, 2011), prior knowledge of consideration sets (Lin and Huang, 2006), influence on customer satisfaction (Inman and Dyer, 1997; Taylor and Schneider, 1998), repeat purchase intentions (Inman and Zeelenberg, 2002) and brand switching (Zeelenberg and Pieters, 2004). According to Sugden (1985) regret not only involves the wish that one had chosen differently, it also involves the belief that the original decision was wrong at the time and includes the elements of self-recrimination or self-blame. The more negative the belief about the purchase, the greater the experience of regret because regret is the result of the comparison between \"what is\" and \"what might have been\" (Creyer and Ross, 1999).\nResearch hypotheses: Self-gifting purchase decisions stem from various sales shopping experiences and indulgent acquisitions that vary from \"cheer up\" oneself, to highly controlled, carefully considered self-gifts, intended to celebrate a singularly relevant event (Campbell, 1987; Mick and DeMoss, 1990; Rohatyn, 1990). Post purchase regret involves self-judgments of decision outcomes that result in feelings of sorrow due to the perceived misfortunes, short comings or mistakes, associated with the self-gifting decisions (Landman, 1993; Inman and Dyer, 1997; Watson and Spence, 2007). It is therefore evident the motivations and the decision-making process to self-gift varies, lead to cognitive fatigue and varying levels of post-purchase regret. Sales shopping experiences relate to practices and beliefs about shopping and associate with indulgent practices which appear closely linked to drive the purchase of merchandise in general, and self-gift shopping in particular (Figure 1). Consequently, the first hypotheses is:\nMethod: An initial pilot survey of 50 random participants clarified the content and structure of the indicative factors to formulate a series of ranking questions that offered a succinct survey. The final instrument was an online format that sought Likert-style responses within a 1 (disagree) though to a 5 (agree) ranking system. The timing of the main study was outside the typically heavy gift-giving periods observed in western societies. Irrespective of the timing, sales serve as an economic motivation by creating higher value for customers when shopping for family, friends and self. An Australian list supplier provided a rented list of potential respondents in order to generalize the study across Australia. The specified age range for participants was between 18 and 45. The list supplier made e-mail contact with the potential respondents and supplied a link to the university website and the questionnaire. Participation in the online survey was voluntary, anonymous and allowed four weeks for participation. After cleaning the data, 307 responses were acceptable. These participants were predominantly female (75 percent) and an overall average age of 38 years. Whilst the data represented a small response rate, the sample size is adequate for the selected analysis and outcomes determined through AMOS version 21.0.\nResults: Byrne (2010) describes exploratory factor analysis (EFA) as being \"designed for the situation where links between the observed and latent variables are unknown or uncertain\". The approach in this study covered a factor analysis, hypothesis test and describes a path analysis. The data were subject to a principal component analysis factor exploration using SPSS version 21. Table I presents the factor structures and associated Cronbach Alpha. Individual factor loadings are consistent and Alphas range from 0.872 to 0.965. These structures are acceptable for further development within an AMOS (v21) hypothesis analysis (Table II) and path model (Table III).\nDiscussion: To date there is little understanding of linkages between the antecedents that lead to self-gifting (Davies et al., 2010; Heath et al., 2011). Certainly, an area of interest is the interactions between shopping practices, self-gift motivations and regret. One important aspect of this study is the design and testing of the relevant concepts. All factors offer succinct content and exhibit strong Cronbach Alpha measures that define the concepts of hedonic shopping, indulgent views, self-gift concepts and regret. Consequently, the connections between these concepts allow insights into the relationships between the variables. The study qualifies the idea that the shopping motivations of hedonic shopping and indulgent actions make moderate contributions to the purchase of self-gifts across the four motives. The hypothesis testing showed shopping and indulgent views did not encourage regret nor did reward and hedonistic self-gifts. The only evidence of post purchase regret emerged from the concepts of SGTherapeutic and SGCelebratory.\nMarketing implications: Generally, consumers appear more self-oriented in their purchases and consumption behaviors. Understanding self-gifting behaviors holds important retail management implications and opportunities for self-gift promotional strategies through advertising or service encounters. This section illustrates the possible promotions representative of each shopping motivation, self-gift concepts and regret.\nConclusion, limitations and future research: Self-gift giving research continues to mature and develop both theoretically and empirically. While the established theory related to gift and self-gift giving covers a broad spectrum, the connections between purchase influencers and self-gift giving motivations holds minimal coverage. This is a special, self-directed, internalized purchase situation which differs from the buying of traditional gifts that fulfill various obligations, rituals and protocols. Contrary to earlier research, the results show various positive relationships because different shoppers engage in self-gifting for different reasons and not all self-gift motivations attract regret. The research attempts to remedy those shortcomings concerning the complex interrelated phenomenon of hedonistic shopping and indulgence that influence self-gifting behaviors and to gauge the incidence of post purchase regret associated with self-gifting during clearance sales.\n", "role": "user" }, { "content": "Purpose: Self-gifting is a performative process in which consumers purchase products for themselves. The literature to date remains silent on a determination and connection between the extents of post-purchase regret resulting from self-gifting behavior. The purpose of this paper is to examine identification and connection of self-gifting antecedents, self-gifting and the effect on post purchase regret.\nDesign/methodology/approach: This study claims the two antecedents of hedonistic shopping and indulgence drive self-gifting behaviors and the attendant regret. A total of 307 shoppers responded to a series of statements concerning the relationships between antecedents of self-gifting behavior and the effect on post-purchase regret. Self-gifting is a multi-dimensional construct, consisting of therapeutic, celebratory, reward and hedonistic imports. Confirmatory factor analysis and AMOS path modeling enabled examination of relationships between the consumer traits of hedonistic shopping and indulgence and the four self-gifting concepts.\nFindings: Hedonic and indulgent shoppers engage in self-gifting for different reasons. A strong and positive relationship was identified between hedonic shoppers and reward, hedonic, therapeutic and celebratory self-gift motivations. hedonic shoppers aligned with indulgent shoppers who also engaged the four self-gifting concepts. The only regret concerning purchase of self-gifts was evident in the therapeutic and celebratory self-gift motivations.\nResearch limitations/implications: A major limitation was the age range specification of 18 to 45 years which meant the omission of older generations of regular and experienced shoppers. This study emphasizes the importance of variations in self-gift behaviors and of post-purchase consumer regret.\nOriginality/value: This research is the first examination of an hedonic attitude to shopping and indulgent antecedents to self-gift purchasing, the concepts of self-gift motivations and their effect on post-purchase regret.\n", "role": "assistant" } ]
Assessing sustainability performance of high-tech firms through a hybrid approach
[ "Assessment", "High-tech firms", "Sustainability performance" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The Millennium Ecosystem Assessment Synthesis Report shows that economic development brings unprecedented energy consumption and environmental pollution. In total, 60 per cent of the world's ecosystems are deteriorating due to a lack of sustainability (Szilagyi et al., 2018). Countries around the world not only encourage firms to voluntarily fulfil their environmental responsibilities but also issue laws to regulate their practices and to promote sustainable development (Tien-Shang Lee, 2012). Currently, with the increasing environmental awareness of consumers and relevant government regulation restrictions, firms are facing great pressure regarding their environmental policies and procedures (Mathiyazhagan et al., 2014; Mumtaz et al., 2018). With the advent of the knowledge economy and the big data era, the sustainable development of high-tech firms has become a hot research topic (Silva, 2008; Xu and Lin, 2018). High-tech firms are economic entities that increase production capacity, create economic benefits and win market competition through innovative technologies, research and development of new products and the use of advanced equipment (Bojnec and Ferto, 2014; Yu et al., 2018). As key players in the upgrading of industrial structure, high-tech firms are better than resource-based firms at using knowledge and technology to improve product and service quality and thereby increase economic returns (Lin et al., 2012; Kuo et al., 2014; Xie et al., 2018). As international competition becomes increasingly fierce, high-tech firms, as representative of national science and technology innovation strength, are crucial for increasing national competitiveness (Ester et al., 2010; Wonglimpiyarat, 2016). Under the pressures of rapid market changes and technology upgrades, the assessment of sustainability performance can help high-tech firms identify problems and determine the path towards sustainable development. Many studies have been conducted on the sustainability assessment of resource-based firms. Such firms usually focus on clean energy, waste recycling, and other factors. In contrast, high-tech firms pay more attention to innovation-driven practices (Chen et al., 2018). Therefore, a new assessment framework is needed to assess the sustainability performance of high-tech firms.\n2. Literature review: 2.1 Sustainability framework and performance\n3. Methods: A hybrid approach is used to assess corporate sustainability performance in this study. First, a qualitative literature research method is used to select a number of sustainability criteria. Based on these results, the quantitative statistics of word frequency analysis are used to assist in screening the criteria to enhance their objectivity. Second, according to the characteristics of the criteria and expert opinions, cluster analysis is used to divide the criteria into several sustainability aspects, and then a sustainability performance assessment framework is built. Finally, in order to overcome the problem of expert semantic ambiguity, grey theory and DEMATEL are combined to assess the sustainability performance of high-tech firms. The specific procedure is described as follows.\n4. Results and analysis: 4.1 Case background\n5. Implications: 5.1 Theoretical implications\n6. Conclusion: Sustainability performance assessments can provide high-tech firms with clear directions for their sustainable development. Since the third scientific and technological revolution, high-tech firms have attained key roles in adjusting the industrial structure and promoting sustainable economic development. Building a scientific framework of sustainability assessment is an important part of studying the sustainability performance of high-tech firms. Although studies have assessed corporate sustainability from many perspectives, knowledge gaps remain. With the rapid development of society, the understanding of sustainability is changing. In addition, the criteria for assessing corporate sustainability performance are becoming more complicated. In this situation, the existing assessment systems or methods may no longer be appropriate. Therefore, to address this research problem, we propose a hybrid approach to assessing the sustainability performance of high-tech firms. First, the assessment criteria of sustainability are obtained from the literature as well as a high-frequency word analysis of firm features. Thus, the constructed assessment framework is more scientific and comprehensive. Second, we propose obtaining the assessment aspects via a cluster analysis and verifying the current theory. Finally, we use the GDEMATEL to evaluate firms in the Dalian High-tech Zone to explore the aspects and criteria that represent driving factors and core problems in the sustainability performance of high-tech firms.\n", "role": "user" }, { "content": "Purpose: In light of the lack of subjective criteria and scientific rationality in current sustainability performance assessment, the purpose of this paper is conducted to improve the sustainability performance assessment of high-tech firms by developing a hybrid approach that integrates quantitative and qualitative research methods.\nDesign/methodology/approach: This study proposed a hybrid approach that integrates word frequency analysis, cluster analysis, grey theory and the decision-making and trial evaluation laboratory (DEMATEL) method. Specifically, this study identifies useful criteria using quantitative word frequency analysis as well as qualitative literature research. Then, cluster analysis is used to divide these criteria into different categories. Subsequently, this study applies the grey theory associated with the DEMATEL method to assess the sustainability performance of high-tech firms.\nFindings: The results reveal that the socio-environment is an important aspect underlying the corporate sustainability performance of high-tech firms. Therefore, high-tech firms should enhance their pollution emission control capabilities and increase investment in energy-conservation and emission-reduction technologies to drive sustainable development. In addition, increasing green product sales revenue and improving the guiding capability of green consumption are core issues that firms must address.\nOriginality/value: This study assesses the sustainability performance of high-tech firms by applying a hybrid method. This method can be used to construct a framework for scientific sustainability performance assessment and to provide a clear direction for the sustainable development of firms.\n", "role": "assistant" } ]
Predicting corporate failure: some empirical evidence from the UK
[ "Business failures", "Bankruptcy", "United Kingdom" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1 Introduction: In the post Enron-Andersen debacle era, there has been widespread debate among various stakeholders in the quest to identify firms likely to go bankrupt and/or become financially distressed (Gerald, 2002). Notwithstanding the gigantic number of corporate failure models, which try to give early warning indicators of financial distress, developed by various renowned researchers ranging from academicians to practitioners, Morris (1997) suggests that the best-known models are Altman's ZETA in the USA, Taffler's Z-score models in the UK and these have now been in commercial use for about 20 years. Discriminant analysis has been employed to identify bankruptcies in sectors such as manufacturing industries in the USA (Altman, 1968), small-scale business (Edmister, 1972), railroads industry in the USA (Altman, 1973), banks (Sinkey, 1975) and insurance companies (Trieschmann and Pinches, 1973; Pinches and Trieschmann, 1977). Despite the vast literature, Shukla (1994) raises the question that, in the present-day of environmental hurly-burly, why do some organizations prosper and grow, while others collapse and disintegrate? Ironically, research interest on corporate failure dates back to 1932.Interestingly, as at the time of writing, an all-embracing approach to corporate failure is non-existent, despite its relevance to stakeholders (Balcaen and Ooghe, 2006; Thornhill and Amit, 2003). Therein lays the justification of this paper, which attempts to answer the preceding question by developing a suitable Z-score model using multiple discriminant analysis (MDA) and then applying it in order to measure the financial health and the risk of failure of UK manufacturing.The rest of the paper is orgainsed as followed: Section 2 provides a brief review of the relevant literature. Section 3 discusses the methodology used for the study. Section 4 discusses the empirical results and finally, Section 5 concludes the study.\n2 Review of relevant literature: Corporate failure includes creditors' or voluntary liquidation, and appointment of receiver (Taffler, 1982, 1983, 1984; Morris, 1997; Bankruptcy Yearbook & Almanac, 2001). Beaver (1966) adopted univariate discriminant analysis (UDA) and concluded that financial ratios of failing firms differ from those of non-failed firms. The study extended Patrick's (1932) ratio analysis approach and concluded that ratio analysis could prove a useful tool for company successful firms with failed ones. Charitou et al. (2000) concluded that the ratio of retained earnings to total assets maintains its unique ability to predict failure in the other two years prior to failure.The statistical technique (UDA) adopted by Beaver (1966) and other researchers begs the question of dependence on a single ratio rather than taking a holistic view of possible complex factors that may indicate future bankruptcy. Zavgren (1983) argues that UDA creates inconsistent signals since different variables could give conflicting forecast. Therefore, alternatives that guarantee consistency are imperative (Altman, 1968). This sets the scene for the studies using MDA. Altman (1968) style is related to that of Beaver's (1966), but for the concurrent use of multiple financial ratios in a given year, to predict an imminent collapse. In Altman's models, the highest contributor was the profitability ratio, earnings before interest and taxes/total assets, whilst the least was working capital/total assets. Altman (1968) argues that the profitability ratios contribution is not surprising, considering that the incidence of profitable firms' failure is almost nil.It is important to note that, events of this day and age make Altman's argument not necessarily valid given the collapse of profit superstars like Enron and WorldCom. Further, creditors demand cash not profit for settlement of debt. In a follow-up study, Altman et al. (1977) were motivated to improve Altman's earlier study by using a more complex MDA to predict failure. This surveillance highlights a distress that the variables are irregular across numerous studies. Furthermore, these two studies were exceedingly precise in the short-run, but the precision shrunk vividly when the facts were for periods that were more than two years prior to ruin.Various researchers have criticised Altman's (1968) work based on lack of evidence of ex ante predictive ability of ratios (Joy and Tollefson, 1975; Moyer, 1977). Moyer (1977) for instance argues that better explanatory power could be obtained if market values of equity/book value of debt and sales/total assets variables were eliminated. Although various models now exist, they are all different (Argenti, 1983; Kip, 2002); nevertheless, they all follow the concepts of the original one devised by Altman (1968). The most important of the previous studies, which ignored the various limitations of Altman model, modified it in the following. These include:* assignment of prior probability membership classes (Deakin, 1972);* employment of a more appropriate \"quadratic classifier\" (Altman et al., 1977);* use of cash flow-based models (Gentry et al., 1987);* use of quarterly financial statement information (Baldwin and Glezen, 1992); and* investigation of the use of current cost information (Keasey and Watson, 1986; Aly et al., 1992).Other researchers noted that the Altman model is not just providing a basis for predicting corporate failure, but also a tool to assist in credit evaluation, internal control guideline and a guide to portfolio selection (van Horne, 1974; Bolten, 1976; Reed, 1976). Finally, Moyer (1977) stated that \"the result achieved from other dynamic approaches have not been sufficiently better than static naive model to justify their serious attention at this time\".Taffler (1977) used the statistical technique MDA, to develop a linear model to determine whether failed firms could be distinguished from non-failed firms in the UK. The model provided an overall classification accuracy for the year prior to failure as compared to two and three years before failure. The major contribution of his study was the development of a Z-score model for predicting company failures in the UK. Interestingly, Taffler (1977) claimed his model had 100 per cent predictive accuracy. In a follow up study, Taffler (1982, 1983) argues the pairing technique used in corporate failure studies was no more successful than selection by any other method. MDA models are reliable to a certain extent in predicting corporate failure.However, it is difficult to assess this reliability since the assumptions of discriminant analysis may be valid or invalid in reality (Zavgren, 1985). As a result, a more reliable approach based on less or no assumption should be considered. Notwithstanding, Aziz and Humayon (2006) suggest that the use of MDA and logit models dominate most previous studies. Therein lays the justification for the adoption of MDA for the present study.\n3 The research question and methodology: Economic cost of insolvency is immeasurable. Therefore, the need for reliable empirical model that predicts corporate failure promptly and accurately is imperative. Notwithstanding the above significance of predicting corporate failure, it is alarming to note that not a single well-specified theory of how and why businesses fail has yet been developed (Kip, 2002). The study adds to existing literature on predicting corporate failure by answering the following research questions:RQ1. To what extent, is it possible to predict corporate failure in the UK. Manufacturing sector using the Z-score model?RQ2. To what extent, is it possible to discriminate between failed and non-failed companies based on Z-score model?3.1 Data collection of failed companies\n4 Empirical analysis and results: The empirical finding was structured to address the following in the relevant sections: the initial model development and testing, assessment of type I and II errors and variables modification and modified model development and testing.4.1 The initial model development\n5 Summary and conclusion: This paper sought to assess the usefulness of financial ratios together with MDA to distinguish between failed and non-failed companies. Specifically, a set of financial ratios adopted from Argenti (1983) were used to develop a Z-score model to approach corporate failure prediction in general and discriminate between failed and non-failed companies in particular. The justification for the use of financial ratio together with MDA was motivated by the criticism of the univariate theory and technique as discussed in literature review section of this paper. The Z-score model proved to be superior in predicting corporate failure as compared to the individual ratios used in isolation. Furthermore, the modified model proved to be extremely accurate in predicting bankruptcy correctly in 97.3 per cent as compared to the initial model of 79 per cent. Interestingly, the classification accuracy of both models faded with the introduction of a new sample to test the predictive accuracy of both models.The results of this study contrast sharply with Altman (1968) and Taffler (1977) model's predictive accuracy, which recorded a higher validation sample of 96 and 98.5 per cent as compared to 94 and 99 per cent, respectively, of the initial sample. This paper has enough grounds to support Moyer's (1977) argument that using genuinely post-dated sample of firms, the predictive ability of Z-score would fade. The results also indicate that, though this paper adopted the style of Deakin (1972), its conclusion differs from the aforementioned work with respect to reducing the type II error to nil in terms of the modified model. The justification for this finding might partially be due to the unbiased nature in sample selection, sample sizes and the use of relatively recent data for the research.Moreover, the various hypotheses tested had inconclusive results due partially to the numerous assumptions and limitations inherent in this paper. Specifically, the ratios and its dimensions used to distinguish between failed and non-failed and to predict corporate health or risk were helpful but could also be misleading in the midst of creative accounting. Perhaps, the greatest limitation of this study was the decisive factor used for including variables in the model, unlike other researchers, who applied the stepwise factor analysis. An area for further study is the development of a new set of ratios since the variables used were adopted from a 1983 paper. It may well be that these ratios are no more relevant in the 1990s and beyond and perhaps not reliable in the manufacturing sector, considering the extent of misleading picture they portray (see assessment of type I and II errors above). A further study could therefore start from scratch and develop variables using the factor analysis of the SPSS. We conclude that predicting corporate failure is a grey area, and though much work has been carried out, the final word on corporate failure is not yet said. Notwithstanding, the major contribution of this paper is the discovery of the superiority of the net profit margin in predicting failure vis-a-vis the gross profit margin.\n", "role": "user" }, { "content": "Purpose: - The purpose of this paper is to use relevant financial information of private medium-sized failed and non-failed manufacturing firms in the UK, during the period 1994-2004 to determine whether corporate failure can be predicted by developing a Z-score model.\nDesign/methodology/approach: - Multiple discriminant analysis is used to develop the Z-score to support the notion that Z-score is an innovation to overcome the numerous difficulties associated with using single ratios to measure companies' health or risk of failure.\nFindings: - This paper advances the notion that the net profit margin is superior to the gross profit margin in discriminating between failed and non-failed UK manufacturing companies in terms of its significant contribution to the Z-score, though the latter exceeds the former slightly using the univariate analysis.\nOriginality/value: - This research contributes to the area of benchmarking by providing a method to more accurately predict corporate failure.\n", "role": "assistant" } ]
Ethical attribute and brand concept congruity enhances brand evaluations
[ "Social responsibility", "Sustainable brands", "Conspicuous consumption", "Brand concept", "Ethical attributes", "Perceived congruity" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Under the guidance of its former CEO Indra Nooyi, Pepsi Co. introduced Pepsi Next - an aspartame-free cola beverage brand. This introduction coincided with a strategic modification of Pepsi Co.'s brand portfolio through the acquisition of brands associated with health benefits, such as Tropicana and Quaker Oats. In introducing its new brand and adding brands associated with health benefits, Pepsi Co. sought to contribute to consumers' long-term well-being (Chatterji, 2013). Although Pepsi Co.'s strategy was initially praised in the media (Seabrook, 2011), Pepsi Co.'s stock prices stagnated, while Coca Cola's doubled (Chatterji, 2013).\nConceptual background: Ethical attributes\nStudy 1: interactive effect of ethical attributes and brand concepts: Study 1 examined whether congruity between ethical attribute type and brand concept influences brand evaluations (H1, H2).\nStudy 2: mediating role of perceived congruity: Study 2 seeks to replicate the congruity effect of ethical attributes and brand concept, and explores the mediating effect of perceived congruity (H3).\nStudy 3: replication with different product category and brands: To increase the generalizability of previous findings, Study 3 tested the ethical attribute-brand concept interaction, as well as the mediating role of perceived congruity in a new product category. While Studies 1 and 2 included private-label brands, Study 3 involved a set of national brands to rule out the potential confounding of retailer and brand perceptions on consumers' brand evaluations.\nStudy 4: the moderating role of brand usage conspicuousness: The results of Studies 1, 2 and 3 suggest that ethical attribute-brand concept congruity positively influences consumers' brand evaluations. Brand consumption in the product categories tested in Studies 1, 2 and 3 (hoodies, backpack and carry-on luggage) are relatively conspicuous. Study 4 examines whether the observed effects of ethical attribute-brand concept congruity emerge at high (but not low) levels of conspicuousness of brand consumption (H4).\nConclusion and implications: In light of the increasing adoption of ethical attributes by brands, this research explored what ethical attribute types are most likely to increase consumers' evaluations of symbolic or utilitarian brands. It examined whether brands competing within the same product category and associated with differentiated brand concepts might benefit from offering different types of ethical attributes.\n", "role": "user" }, { "content": "Purpose: This paper aims to examine to what extent congruity between ethical attributes (i.e. product attributes with positive implications for the environment, human rights, social issues and animal welfare) and brand concept (i.e. the unique meaning associated with a brand in consumers' minds) influences consumers' evaluations of brands offering ethical attributes.\nDesign/methodology/approach: Four studies involving North American consumers empirically tested the moderation effect of brand concept on consumer evaluations of ethical attributes and the mediating role of perceived congruity.\nFindings: This research finds an interactive effect of ethical attribute type and brand concept on brand evaluations, such that congruent ethical attribute-brand concept pairings (i.e. a utilitarian [symbolic] ethical attribute offered by a brand with a utilitarian [symbolic] brand concept) result in more favorable brand evaluations (Studies 1, 2, 3 and 4). Consumers' perceptions of congruity between ethical attributes and brand concepts mediate this interactive effect (Studies 2 and 3). Moreover, a positive congruity effect of ethical attributes and brand concepts emerges at higher levels of conspicuous brand consumption (Study 4).\nResearch limitations/implications: It is important to acknowledge that the current research did not specifically consider the case of utilitarian and symbolic ethical attribute offerings by luxury brands. This is a question that is left to future investigations.\nPractical implications: For marketing managers, findings indicate that brands gain from ethical attribute introductions only when these attributes are congruent with the brand concept. In addition, brands benefit to a greater extent from offering congruent ethical attributes when brand consumption is conspicuous.\nOriginality/value: The findings of this research contribute to the literature on the effect of ethical attributes on consumers' responses to brands and highlight the importance of brands' choice of ethical attributes.\n", "role": "assistant" } ]
Impact of brand personality and consumer ethnocentrism in China's sportswear market
[ "Brand personality", "Consumer ethnocentrism", "China", "Sportswear", "Country-of-origin", "Purchase intention", "Brands", "Ethnocentrism" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The success of the Beijing 2008 Olympic Games, along with China's rising economy, has instilled a new national confidence and pride in many Chinese citizens. According to surveys conducted by the Boston Consulting Group in 2008, many Chinese consumers claim that purchasing domestic brands is the right thing to do and that they will increasingly favor local brands in the near future. For example, 42 percent of the Chinese consumers in their study preferred local apparel and footwear brands, whereas only 19 percent preferred foreign brands (Anestis et al., 2008). However, other studies have suggested that while Chinese youths value patriotism, they show little partisan behavior when it comes to international marketers (Madden, 2004). Other factors - such as brand image, price, and product quality - are the most important variables influencing their purchase decisions (Ogilvy Group China and Millward Brown ACSR China, 2008), meaning that although the Chinese have a strong sense of patriotism, those feelings do not significantly influence how they purchase domestic and foreign brands. The brand choices that the Chinese make seem to be more about rationalism than about nationalism (Anestis et al., 2008; Hille, 2003; Lau, 2003).\nLiterature review and hypotheses: Brand image and brand personality\nConceptual framework: The purpose of this exploratory study is to examine the effect of brand personality and consumer ethnocentrism on the evaluation of product quality and purchase intention for domestic and foreign brands in China's sportswear market. For the purpose of this study, a conceptual framework was developed and is shown in Figure 1: brand personality and consumer ethnocentrism are considered latent-independent (exogenous) variables, while product quality perceptions and purchase intentions are considered latent-dependent (endogenous) variables. We also examined the relationship between quality perceptions and purchase intentions toward domestic and foreign brands in China's sportswear market.\nMethodology: Choice of product category\nData analysis and results: The framework in Figure 1 and the corresponding relationship were investigated by studying Chinese consumers. Structural equation modeling (SEM) using AMOS 19 was employed for confirmatory factor analysis (CFA) and path analyses. We followed the two-step approach recommended by Anderson and Gerbing (1988). In the first stage, the measurement model was analyzed to ensure sufficient reliability and validity of the constructs. In the second stage, the proposed causal model of the relationships between constructs was tested. Model fit criteria suggested by Hu and Bentler (1999) and Tabachnick and Fidell (2007) were used for both the measurement and the structural model: kh2/df, goodness of fit (GFI), adjusted goodness of fit (AGFI), comparative fit index (CFI), root mean square residual (RMR), and root mean square error of approximation (RMSEA). Acceptable models should have kh2/df<=2, AGFI>=0.80, RMR<=0.1, RMSEA<=0.06, and GFI and CFI greater than 0.90.\nDiscussion and implications: This research aims to empirically investigate the effect of brand personality and consumer ethnocentrism on consumers' quality perceptions and purchase intentions toward domestic and foreign brands, specifically in China's sportswear market. The study contributes theoretically and empirically to research on the effects of brand image and consumer ethnocentrism in emerging markets. We believe that the findings from this study will help domestic and international marketers have a better understanding of the role of brand image (specifically brand personality) and consumer ethnocentric tendencies in influencing young Chinese consumers' product evaluation and purchase intentions. In addition, few studies have simultaneously examined the effects of brand image and consumer ethnocentrism (moderated by COO) on consumers' purchase intentions and product evaluations, although both factors may have a direct effect, which provided new research and managerial implications. Although these findings are based on the evidence from China, China is an emerging economy; therefore, with successful replications, our findings may be generalizable to other emerging countries or regions.\nLimitations: It is important to point out that our study is limited to the sportswear market in China and only focuses on the young Chinese generation. Previous studies have suggested that COO and consumer ethnocentrism effects tend to vary by product category (Balabanis and Diamantopoulos, 2004; Han and Terpstra, 1988). Hence, future studies should replicate this study with other product categories with the specific COO of foreign alternatives. In addition, future research needs to be done if the results are to be generalized to the entire Chinese population in light of significant age-related differences in consumer attitudes and behaviors.\n", "role": "user" }, { "content": "Purpose: This study aims to investigate the effect of brand personality and consumer ethnocentrism on perceived quality and purchase intentions in China's sportswear market.\nDesign/methodology/approach: In order to examine and compare the effect of brand personality and consumer ethnocentrism on the evaluation of product quality and purchase intention toward domestic and foreign brands, COB (the country with which the brand or firm is associated) and COM (the country in which final production takes place) cues were used in this study. Chinese college students represented the sample of 385 respondents. Structural equation modeling was used to investigate the causal relationships among brand personality, consumer ethnocentrism, perceived quality, and purchase intention.\nFindings: Results revealed that brand personality has a significant effect on Chinese consumers' product quality perceptions and purchase intentions toward both domestic and foreign brands. Furthermore, it is found that Chinese consumers' ethnocentric tendencies have no significant impact on their intentions to buy either domestic or foreign sportswear brands. However, ethnocentric Chinese consumers do positively evaluate the quality of domestic brands when domestic brands are also made domestically and negatively evaluate the quality of foreign brands when foreign brands are also manufactured non-domestically.\nOriginality/value: The study contributes theoretically and empirically to research on the effects of brand image and consumer ethnocentrism in emerging markets. The findings from this study will help domestic and international marketers have a better understanding of the role of brand personality and consumer ethnocentric tendencies in influencing young Chinese consumers' product evaluation and purchase intentions. In addition, few studies have simultaneously examined the effects of brand image and consumer ethnocentrism (moderated by COO) on consumers' product evaluations and purchase intentions, which provided new research and managerial implications.\n", "role": "assistant" } ]
An integrated framework for ISO 9000 motivation, depth of ISO implementation and firm performance: The case of Taiwan
[ "ISO 9000 series", "Motivation (psychology)", "Business performance", "Taiwan" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The application of standardized quality system models (e.g. ISO 9000) in business is considered to be a most important phenomenon in recent times in quality management development and globalization (Dick, 2000). According to, the latest statistics released by the (ISO, 2006), as of the end of December 2005, 161 countries accounted for 776,608 certifications. Such an impressive number does indeed make ISO 9000 a universal and significant phenomenon.While the nature and scope of ISO 9000 is well understood, several issues remain unaddressed. First, some studies suggest, that the most prominent reason for implementing ISO 9000 is that customers prefer to buy from suppliers that are ISO certified (Carlsson and Carlsson, 1996; Rao et al., 1997; Jones et al., 1997; Casadesus et al., 1998, 1999; Huarng et al., 1999; Acharya and Ray, 2000; Yahya and Goh, 2001; Najmi and Kehoe, 2001; Terziovski et al., 2003; Costa and Lorente, 2004; Bhuiyan and Alam, 2005). Other studies, however, show that internal motives are the most important consideration in certification (Idris et al., 1996; Najmi and Kehoe, 2001; Gotzamani and Tsiotras, 2002). Obtaining ISO certification does not automatically lead to improvements in the processes of organizations. The literature reveals that organizations that pursue ISO 9000 certification willingly are more likely to report improved organizational performance than those that only obtain ISO 9000 certification under customer pressure (Lee, 1995; Jones et al., 1997; Douglas et al., 1999; van der Wiele et al., 2000; Singels et al., 2001; Yahya and Goh, 2001; Llopis and Tari 2003; Terziovski et al., 2003; Park et al., 2007). Therefore, it is necessary to examine the question: \"What is the relationship between internal and external motivations, and what is the influence of motivation on the depth of ISO 9000?\" Studying the relationships among the internal and external motivations and the depth of ISO 9000 implementation could help clarify the reasons for the failure of some externally-driven companies.Second, there has been considerable debate in the literature as to whether ISO 9000 has a positive impact. Many studies (Elmuti, 1996; McAdam and McKeown, 1999; Huarng et al., 1999; Lipovatz et al., 1999; Yahya and Goh, 2001; Arauz and Suzuki, 2004; Casadesus and Karapetrovic, 2005a; Briscoe et al., 2005) have demonstrated the benefits of implementing ISO 9000. These studies have claimed that the implementation of ISO 9000 standards has improved operational and business performance. However, Rao et al. (1997), Terziovski et al. (2003) and Naveh and Marcus (2005) have indicated that while implementing the ISO 9000 standards led to improved operational performance, it did not give rise to better business performance. Corrigan (1994), Stephens (1994) and Lima et al. (2000) found that the implementation of ISO 9000 standards did not result in improved productivity, quality, or profitability. In light of these inconsistent findings, Terziovski et al. (2003) stated that whether or not ISO 9000 is beneficial is likely to be the major determinant of the degree to which managers will embrace or reject ISO 9000 certification in the twenty-first century. Therefore, it is important to also consider the question: \"Does ISO certification guarantee a better performance for organizations, and, if so, how does the depth of ISO 9000 implementation impact a firm's performance?\" Investigating these issues could be of benefit to practitioners in deciding whether to pursue ISO 9000 and the depth to which the standards are implemented.Judging from the above, it is important to integrate the motivations behind implementing ISO 9000 certification, the depth of implementation, and the performance of ISO 9000. Our contention is that these constructs have some close relationships, for the motivations create an environment where the ISO 9000 implementation can take place, and the implementation affects ISO 9000 performance. Other researchers who support our contention are Gotzamani and Tsiotras (2002), Brecka (1994) and Kochen (1993). Gotzamani and Tsiotras (2002) reviewed the literature (Brecka, 1994; Kochen, 1993) and summarized the potential of the ISO 9000 standards and their contributions, by stating that:... the long-term effectiveness and real value of the quality assurance standards is not based on their content and requirements but on the way that companies adopt and implement these requirements. The key for their success lies in the companies' real commitment to quality improvement and their true motives for certification, which finally dictate the way and depth to which the standards are implemented.The literature on ISO 9000 contains considerable research that focuses on the respective impacts of motivations and implementation processes on ISO 9000 performance. However, there is a lack of systematic empirical evidence integrating the motivations behind the implementation of ISO 9000 certification, the depth of implementation, and ISO 9000 performance. Thus, the framework needs to be empirically tested. Therefore, by using the structural equation model (SEM), this study empirically examines the relationship between the ISO 9000 motivation and ISO 9000 implementation depth and looks at how implementation depth influences firm performance.Several important reasons exist for focusing research efforts on Taiwan. The Taiwanese economy has passed through a transition, having been transformed from an agricultural economy to become, as a result of an export processing phase, a worldwide OEM/ODM powerhouse. According to the Global Competitiveness Report 2004-2005 completed by the World Economic Forum (WEF), Taiwan was ranked fourth globally and first in Asia in terms of the growth competitiveness index (GCI) rankings. Moreover, Taiwan was ranked 17th in the business competitiveness index (BCI) rankings and 12th in the sophistication of company operations and strategy rankings.Taiwan is well known to be heavily dependent on exports. The above rankings suggest that the strength of Taiwan as an exporter lies in its business adaptability and efficiency. According to the ISO (2006) report 76 percent of ISO member countries are developing economies or economies in transition. The experience of a developing country with a strong international orientation, business adaptability and efficiency, should be valuable to ISO 9000 practices as it may provide directions for firms in developing countries intending to globalize. It is for this reason that this study focuses on Taiwan, and based on the experiences of Taiwan in terms of ISO 9000 implementation offers important insights into ISO 9000 practices.The remainder of this paper is organized as follows. The next section introduces a research model and related hypotheses based on a literature review. Followed by the section, that describes the research methodology, including the sample and data collection procedures, the operationalized measures of variables, the statistical approach adopted, and the tests for reliability and validity. Followed by the section, that presents the results of testing the structural model. The penultimate section discusses the implications of the results for researchers and practitioners and reevaluates the validity of the findings. The paper concludes with further research implications for this study.\nConceptual framework and research propositions: The conceptual framework on which this study is based is aimed at integrating the motivations behind implementing ISO 9000 certification, the depth of implementation, and ISO 9000 performance, and is shown in Figure 1. The model proposed in this study includes three components. The first component describes the effect of motivations on the depth of implementation of ISO 9000. The concept occurs alongside the concept whereby \"motivation increases the likelihood of employees to work hard\" (motivations - the depth of implementation) proposed by Deci (1971), and the concept whereby \"intrinsic motivation mediates the effect of external factors\" (external motivations- internal motivations- the depth of implementation) proposed by Ambrose and Kulik (1999). The second component describes the effect of the depth of ISO 9000 implementation. It was derived from a number of studies on ISO 9000 standards in the quality management area. For example, Joubert (1998), McAdam and McKeown (1999) and Terziovski et al. (2003) have shown the existence of a positive relationship between the implementation of ISO 9000 and firm performance. The third component describes the relationships among different dimensions of firm performance, and is derived from a number of studies on ISO 9000 standards, quality management and the marketing literature. For example, Elmuti (1996) believed that ISO 9000 enables firms to improve their organizational and market performance, and hence results in increased market share and even improved business performance. Besides, Naveh and Marcus (2005) have claimed that \"more attention is needed to this important issue of the relationships between operating and business performance.\"A literature review in relation to each of these three components is more extensively discussed below and then the hypotheses of this study are presented.Motivations for certification and the depth of ISO 9000 implementation\nResearch methodology: This section discusses the sample and data collection procedure, the operationalized measures of variables as well as the statistical approach used in this study. Moreover, the results of the tests performed for reliability and validity are also presented.Sample and data collection procedure\nSubstantive test results: By using the validated scales, the research propositions were tested using a SEM to represent the causal relationships specified by our hypotheses. This stage of the analysis involved the evaluation of the relationships between the latent constructs. The estimated standardized parameters for the SEM are shown in Figure 2, along with the levels of significance. Overall, the results presented here indicate that the structural model has a good fit with the data (kh2=447.20, df = 212, p<0.05; RMSEA = 0.06, Comparative fit index - CFI = 0.98, and GFI = 0.90). The goodness of fit statistics are listed in Table IV and the covariances among the latent constructs are shown in Table V. With regard to the tests of the individual research propositions, these parameters are reported as shown in Table VI. All of the paths in the model are supported, except for paths H1b and H3b. The test of the proposed hypotheses is based on the direct and indirect effects in the structural model. The path coefficients between constructs provide an indication of the relative strength of each relationship (Joreskog and Sorbom, 1993). Both hypotheses were tested at p<=0.05, the two-tailed significance level.\nDiscussion: The primary purpose of this study was to investigate the relationships among ISO 9000 certification motivations, as well as the implementation depth and performance. The results and their implications are discussed throughout the remainder of this section.Discussion of results\nLimitations and avenues for further research: Several limitations of this study must be mentioned. These limitations also indicate worthwhile avenues for further research.First, the sample was restricted to Taiwan. While this sample is valuable for some developing countries, the ability to generalize these results regarding ISO 9000 to other countries remains limited.Second, our study uses perceptual measures of the magnitude of operational, market and business performance. To make more precise predictions of how ISO 9000 affects business performance, additional research could develop objective measures of these constructs.Third, this study focuses on the motivation and implementation of ISO 9000 as drivers of ISO 9000 effects. Future conceptual development and empirical research can consider a broader array of antecedent factors and examine their effects, such as company size, industry, information technology or the existing quality system.Fourth, this work investigates overall ISO 9000 implementation performance. However, ISO 9000 implementation comprises numerous practices. Researchers should be encouraged to further identify the relationships among ISO 9000 practices, as well as which ISO 9000 practices are directly or indirectly related to business.To summarize, this study contributes to of the body of literature on ISO 9000 by examining the motivation behind implementing ISO 9000 certification, the implementation depth and ISO 9000 performance. We hope that this study can foster further debate and research in this area.\n", "role": "user" }, { "content": "Purpose: - The purpose of this study is to empirically examine whether business can benefit from ISO 9000, and examines how motivation impacts the depth of ISO 9000 implementation and how the depth of ISO 9000 implementation impacts a firm's performance in Taiwan.\nDesign/methodology/approach: - A survey questionnaire was mailed to the 1,668 companies with ISO 9000 certification, and a total of 441 usable responses were returned. Using a structural equation model, this study empirically examines the relationship between ISO 9000 motivation and ISO 9000 implementation depth and how implementation depth influences firm performance.\nFindings: - The results demonstrate that a positive relationship exists between the extent to which companies implement ISO 9000 and firm performance. Additionally, internal motivation fully mediates the relationship between external motivation and ISO implementation depth. Furthermore, the implementation of ISO 9000 directly and positively influences operational performance and indirectly affects market performance, in turn positively impacting business performance.\nOriginality/value: - This study represents a first attempt to construct a conceptual framework that integrates the motivations behind implementing ISO 9000 certification, the implementation depth underlying processes, and ISO 9000 performance.\n", "role": "assistant" } ]
After the disaster: Using the Big-Five to predict changes in mental health among survivors of the 2011 Christchurch earthquake
[ "Mental health", "Disasters", "Earthquake", "Personality", "Big-Five", "Emotional stability" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: On 22 February 2011, a moment magnitude (M\nMethod: Participants\nResults: The purpose of this study is to examine the relationship between personality and changes in psychological distress following the 22 February 2011 Christchurch earthquake. Because Emotional Stability captures individual differences in people's responsiveness to threat (DeYoung, 2010), we predicted that Emotional Stability would protect people from the increases in psychological distress expected to occur in the months after the earthquake. That is, people who are psychologically hardy to challenges in the environment (i.e. those who are high on Emotional Stability) should be more resistant to increases in psychological distress than those who are susceptible to such challenging circumstances. On the other hand, because personality is - by definition - (relatively) stable across time (Donnellan et al., 2006), we expected that psychological distress assessed at Time 1 would be unassociated with changes in Emotional Stability in the months after the 2011 earthquake.\nDiscussion: The earthquake that shook Christchurch on 22 February 2011 turned a once-vibrant city into a disaster area. In addition to claiming the lives of 185 people, the magnitude of the tremor left much of Christchurch's infrastructure in shambles (Kaiser et al., 2012). The extent of the devastation also crippled the economy of the Canterbury region (Parker and Steenkamp, 2012) and led many within the community to seek comfort in religion (Sibley and Bulbulia, 2012). Indeed, the destruction caused by the earthquake exacted a tremendous toll on the Christchurch community - the effects of which will likely be felt for years to come.\nConclusion: While the papers included in this special issue of Disaster Prevention and Management demonstrate that the 22 February 2011 Christchurch earthquake exacted a tremendous toll on the affected community, people's reactions after the devastation highlight an admirable degree of resilience that must not be overlooked. Indeed, our data show that the residents of the Canterbury region were remarkably resilient in the face of an overwhelming amount of destruction. At the same time, our findings identify a risk factor likely to increase people's levels of psychological distress after a natural disaster. Therefore, efforts aimed at rebuilding communities affected by disasters should target this at-risk population to ensure that they are provided with the resources needed to assist them through these difficult times. Only then can the vibrancy and illustriousness of a pre-disaster community be restored to - and perhaps even surpass - its original state.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to examine individual differences in people's resilience to changes in psychological distress following the 22 February 2011 Christchurch earthquake.\nDesign/methodology/approach: Data were based on a subsample of the New Zealand Attitudes and Values Study (NZAVS) - an annual nation-wide longitudinal study of New Zealand adults that began in 2009. In both waves of the NZAVS examined here, participants completed measures of the Big-Five, psychological distress, and demographic covariates. As such, the analyses, which focus on participants who were living in the Canterbury region before the 2011 earthquake (n=325), use measures of personality collected in late October of 2010 (Time 1) to predict changes in psychological distress after the devastation that unfolded on 22 February 2011.\nFindings: Time 1 levels of Emotional Stability were inversely associated with increases in psychological distress following the 2011 earthquake. Psychological distress assessed at Time 1, however, was uncorrelated with changes in Emotional Stability.\nPractical implications: These results show that Emotional Stability protects people against decrements to mental health following a disaster. Thus, efforts to rebuild disaster-stricken communities should ensure that those who are particularly likely to experience increases in psychological distress (i.e. those who, before a disaster, are low on Emotional Stability) receive the help they need.\nOriginality/value: This study assesses a subsample of respondents from a longitudinally based national probability study to show that Emotional Stability exerts a cross-lagged effect on changes in psychological distress following a natural disaster. The access to such measures affords us the rare opportunity to explain how people cope in the wake of a catastrophic disaster.\n", "role": "assistant" } ]
Job satisfaction in developing countries: An evidence from a matched employer-employee survey in Vietnam
[ "Job satisfaction", "Generalized order logit model", "Vietnam labour economics", "J50", "O15" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The economic reform that Vietnam embarked upon in the late 1980s to change from a centrally planned economy into a market economy has helped the country to reach the status of a lower \"middle-income\" country in 2010. High and continuous GDP growth rates as well as successful implementation of economic development programs have resulted in overall improvement of people's welfare and led to significant poverty reduction[1]. As the income level increases, the questions that policy makers face also change. Although job creation remains important, attention is now being shifted to job quality and decent employment (i.e. characteristics of jobs that are associated with higher level of job satisfaction)[2]. In a developing country like Vietnam, job satisfaction could affect the potential growth of firms and the nation. It has both direct and indirect impacts on the productivity, competitiveness and technology upgrading capacity of firms.\nMethodology: Following the previous literature, we postulate that job satisfaction can be described by the following latent variable model:(1)\nData description: As labor force surveys and household surveys in Vietnam typically do not collect satisfaction measurement, in our paper, we rely on a unique matched employers-employee survey conducted in 2008 by the North-South Institute (Canada) in collaboration with the VASS. The survey was designed investigate the behavior of FDI firms and their linkages with and spill-over effect to domestic firms. The firms were drawn randomly and separately for FDI firms and domestic firms from the Government Statistical Office's Vietnam Enterprise Census with a view to be representative of both FDI firms and Vietnamese domestic firms (Nguyen, 2010).\nEstimation result and discussion: Previous studies have often used the standard ordered logit model without paying due attention to the validity of the parallel lines assumption. In our empirical analysis, we tested the validity of this assumption using the Brant test (William, 2006). Table II presents the kh2 statistics of the Brant test for three sets of estimation results, separately for the whole sample, for workers in the FDI firms and for workers in the domestic firms. The Brant test allows us to test the validity for the overall model and for each of the parameters in the model. The Brant test statistics for overall model was significant, indicating that the parallel lines assumption did not hold for the whole model. The Brant tests for individual parameters suggest that the effects of a number of covariates changes across the cut-off points of the ordered logit model, suggesting that the parallel lines assumption did not hold for these variables. They include gender (male), ethnicity (ethnic), union membership (union), job status (manager) and the training expense (cost of training as percent of revenue). These statistical tests suggest that the generalized ordered logit model that relaxes the parallel lines assumption should be used in our empirical analysis.\nConclusion: In this paper, we investigated the determinants of job satisfaction for Vietnamese workers. Although job quality and decent work have drawn attentions of researchers and policies makers in Vietnam, the concept of job satisfaction is still relatively new in Vietnam. It is not normally captured in labor force surveys, household surveys or enterprise surveys. We rely on unique employer-employee matched survey data with relatively rich information regarding satisfaction levels that was made available by a joint effort by the Vietnamese Academy of Social Sciences and Canada North-South Institute. We have been able to investigate efficiency wage policy exercised by firms, the relative income hypothesis using different income comparison and compare worker's job satisfaction level between the FDI sector and domestic sectors.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to contribute to this literature on developing countries by investigating the determinants of job satisfaction in Vietnam where the economics literature on this issue is virtually non-existent. The authors also contribute to the literature on income comparison by extending beyond the within-firm co-worker income comparison.\nDesign/methodology/approach: The authors estimate a generalized order logit model for job satisfaction as statistical tests suggest that the parallel-lines assumption, which is often invoked in previous studies using the standard logit model, does not hold.\nFindings: For Vietnam, the authors find that absolute and relative incomes as well as human resource practices such as efficiency wage and training policy have an impact on workers' satisfaction. Workers in the foreign direct investment (FDI) sectors behave a bit differently from their peers in the domestic sector.\nOriginality/value: Taking advantage of a unique matched employer-employee data set collected in 2008 by the North-South Institute (Canada) and the Vietnam Academy of Social Sciences, the authors are able to investigate the impact of a number of important job characteristics on job satisfaction such as absolute and reference incomes, wage policy, training plan for workers, union membership and job position, and, at the same time, to disentangle the possible differences in job satisfaction of workers in domestic vs FDI firms.\n", "role": "assistant" } ]
Cost of quality consideration following product launch in a present worth assessment
[ "Cost of quality", "Present worth" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: It has long been held that field failures, though regrettable, might need to be accepted at some level. The economic feasibility of better products - gold plated and bullet proof designs - has limits. Juran (1962) suggested that the cost of field failures can be reduced by increased expenditures in prevention of the failures through robust designs and improved appraisal. (The Hubble telescope is an example of enormous field failure costs being amassed in spite of a successful appraisal effort that found the problem before product launch. The results of the appraisal test by a null corrector were discounted in favor of contradictory information from a more sophisticated instrument that happened to be improperly assembled (Dunar and Waring, 1999)). But it was felt that there was an optimal point beyond which the money spent in the failure reduction effort exceeded the money saved by the reduced cost of the failures. The relationship between the cost of prevention plus appraisal efforts and the cost of failures was shown graphically by Gryna (1988). The cost per good unit of production was laid out along the ordinate scale and the level of defectives, decreasing outwardly, on the abscissa. This graph showed that increasing the expenditure on appraisal plus prevention corresponded to a decrease in the cost of failures until the defective level reached around 25%. This was the point at which Total CoQ was optimal (i.e., at its lowest point). Beyond this point, an asymptotic increase in the cost of prevention plus appraisal was required in order to drive field failures any lower.\nThe concept of CoQ: CoQ is the term that is given to any cost that would not have been incurred if the quality of the product was perfect. Schiffauerova and Thomson (2006) put it this way:\nTangible costs: Several costs are associated with the failure of a product in the field. Principle among these is the cost of fixing or replacing the product. The cost of either of these responses includes one or more of the following:\nIntangibles of CoQ: Pricing constraints\nCoQ and the business case: The costs of poor quality are real and often substantial. Omar and Murgan (2014) cite work that suggests that costs of quality may represent as much as 30 percent of all manufacturing costs and account for between 15 and 25 percent of total sales. Why then is CoQ not made a part of the evaluation of the project - the business case - at the project planning phases? Why are quality costs not a cash flow consideration? Sower and Quarles (2003) submit that ROI determinations of the cost spent on prevention and appraisal actions against the cost savings in internal and external failure costs would be possible with CoQ data. They point out that a new financial instrument, Return on quality (ROQ), is coming into use. ROQ \"measures increases in revenues associated with increases in customer satisfaction against costs associated with quality efforts to obtain those improvements\" (Sower and Quarles, 2003, p. 3). Sower et al. (2002), state that a study by Viger and Anandarajan (1999) found that \"managers who have access to quality cost data make different decisions than managers who do not have quality cost data\" (p. 2). While they do not explain what they mean by \"different decisions,\" they note that in many cases the information on CoQ is limited to \"vague estimates of the amount of scrap produced in the short term\" (p. 2). Given an instrument such as ROQ, management decision making would be better informed and perhaps significantly different.\nCase study: In an endeavor to understand the magnitude of the impact CoQ data would have on a business case, a case study is developed in which a project is evaluated: first, using the standard assumptions of cash flow, and second, adding post-launch costs of imperfect quality to the cash flow consideration. No difference is anticipated in the value of the project - as reflected by a positive PW - when standard warranty cost data is added to the project assessment. However, the amount of change brought to the value of the investment will be assessed for a range of cash flow conditions from warranty costs associated with standard field performance to costs associated with a special cause affecting a critical component or product function. The upper extreme implies a level of quality concern that could lead to product recall and possible liability litigation. Sensitivity analysis will seek the point at which accepting the null hypothesis - no significant material difference in the ROI of the standard business case vs the case including CoQ - alters to a rejection of the null.\nPW determination: Standard cash flow assumptions\nSensitivity analysis: Adding CoQ cash flows appear to have little impact on the evaluation of a project based on PW when the performance of the product does not deviate from historically low levels of quality concerns. But what if an extraordinary problem is considered? While a business case is not intended to factor in failures in a design or problems with manufacturing, it may be worth exploring these rocky margins in an effort to assess risk and plan mitigation strategies.\nConclusion: Imperfect quality occurring in product after the launch of a program is universally acknowledged as undesirable. What is not as widely accepted is that the costs of quality can and should be considered in the evaluation of the worth of the program prior to program development and product launch. The impact these costs can have on a PW calculation is on the order of 30 percent, even if modest assumptions regarding quality performance are employed. The sensitivity analysis done in this project suggests that, even limiting cost estimates to replacement cost and analysis of failed material, a negative PW can be arrived at which would either discourage investment altogether or prompt risk mitigation actions well up front of an actual quality issue occurrence.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to investigate the impact of costs of quality on the present worth (PW) of a new product launch at a North American automobile manufacturer.\nDesign/methodology/approach: The paper is based on the examination of various cash flows associated with a new product launch within an automobile manufacturer. Standard cash flows and a PW analysis were examined and compared to non-standard cash flows which take into consideration post-launch cost of quality (CoQ). A sensitivity analysis was used to determine if any CoQ factors affected the integrity of the product launch.\nFindings: The paper concludes that there is an impact on the PW of a program when CoQ is considered as a cash flow element. CoQ should be considered in a product launch PW analysis preceding any commitment to invest.\nResearch limitations/implications: This study suggests that, given data on the costs that will accrue assuming standard quality concerns following product launch, and the occurrence of special cause issues, the business case can establish a better estimate of the costs a program will face under varying levels of post-launch quality. An understanding of the potential cost consequences of quality issues can shape the understanding of the risks in a planned project.\nPractical implications: The paper shows that CoQ can have a significant impact on a PW analysis. CoQ concerns should be considered during pre-launch planning of a new product.\nOriginality/value: The paper satisfies the need to study when a manufacturer considers investment in the launch of a new product with CoQ concerns.\n", "role": "assistant" } ]
Career control, career dialogue and managerial position: How do these matter for perceived employability?
[ "Employability", "Career control", "Career dialogue", "Managerial role" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Employability is an important topic for current work and careers. It refers to chances and competences to obtain work and to continue to work in the future (see e.g. Forrier and Sels, 2003; Van der Heijde and Van der Heijden, 2006). The importance of one's own responsibility for employability has been largely emphasized in the literature (see e.g. Clarke, 2008). Nowadays careers require a large amount of self-responsibility and self-control, as changes and developments in work are going fast, and urge the individual to act on it, preferably to pro-act on it (Parker et al., 2010; Van Dam, 2004). People in general are expected to keep up with relevant developments in their work, to make their own choices, and to maintain their own employability. Indeed, research has confirmed that concepts such as proactivity, career self-management and career control are positively associated with career outcomes such as objective and subjective career success and job, career and life satisfaction (Guest and Rodrigues, 2015; Seibert et al., 2001). Career control as a concept refers to a behavioral competency that focuses on being able to actually shape one's career by pro-actively taking action (Akkermans et al., 2013). This concept has parallels with theoretical notions on protean and boundaryless career attitudes (Briscoe and Finkelstein, 2009). In line with other studies on career control (e.g. Akkermans et al., 2013), we make use of a self-assessment of career control in this study. De Vos and Soens (2008) have shown that people who take ownership of their careers are more confident in their ability to find another job and report higher levels of career satisfaction. Hence, career control seems to be a highly relevant concept in the context of employability. However, the relationship between career control and employability has so far scarcely been addressed. This is an important gap in the literature. Moreover, examining the relationship between career control and employability also contributes to the ongoing debate to what extent employees can act as real free agents, mastering career control and thereby influencing their employability levels (Forrier et al., 2009). In this way, the research has also practical value for employees who are interested in whether steering their careers can enhance their employability.\nTheoretical framework: Employability\nMethods: Data were collected in 2014 conducting a cross-sectional survey among Dutch managers and employees of a large science-based multinational active in health, nutrition and materials located in the Netherlands. In the Netherlands, the company employs about 5,000 FTE's working at different locations. The company implemented a program for its employees in the Dutch locations aimed at enhancing their employability, which consisted, among other things, of a campaign in which using the career dialogue was promoted. All managers in the Netherlands (n=590) received an invitation to participate in the survey, as well as a random selection of Dutch employees in a non-managerial position (790 invited). In total 460 surveys were returned (33.3 percent overall response rate; 206 employees in a managerial position (34.9 percent) and 254 employees in a non-managerial position (32.2 percent) responded). In all, 81 percent of the managerial employees hold a higher vocational education degree or university degree, compared to 59.8 percent of the non-managerial employees. This final sample was representative for the population that was invited to participate in the study in terms of gender and organizational discipline. It is important to note that employees under the age of 35 were slightly under-represented in the final sample.\nResults: Measurement models\nDiscussion: The current study aimed to examine three-way interactions among career control, career dialogue and managerial position in predicting perceived employability. We first found that career control is positively related to perceived employability, which is in line with earlier empirical work (Akkermans et al., 2013). Important to note is that the relationship between career control and perceived employability is positive and significant for all groups of employees in our study, except for the managers who did not participate in a career dialogue about their own career. Notwithstanding the fact that this positive relationship could not be confirmed for this group of employees, in general this study confirms the idea that enhancing perceived employability, requires the individual to take responsibility for and control over their career (see e.g. Clarke, 2008). Despite this promising result, more research is needed to understand which career competencies (see e.g. Akkermans et al., 2013) and other personal strengths, next to career control, are most relevant for perceived employability. After all, personal strengths are likely to boost perceived employability through motivating employees to exploit and broaden their employability radius (Thijssen et al., 2008) as well as their perceptions of being employable (Wittekind et al., 2010).\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to examine three-way interactions among career control, career dialogue and managerial position in predicting perceived employability. The authors expected that participation in career dialogue strengthens the positive relationship between career control and employability. Furthermore, the authors expected that managers benefit more from career dialogue than employees. Hence, the relationship between career control and employability was expected to be strongest when employees engage in career dialogue and hold a managerial position.\nDesign/methodology/approach: Data were collected in 2014 conducting a cross-sectional survey among managers (n=206) and employees (n=254) at a Dutch location of a large science-based multinational. Moderated regression analyses were used to test the hypotheses.\nFindings: Career control was positively related to perceived employability. This relationship was significantly stronger for the managerial group that did participate in a career dialogue than for the managerial group that did not engage in a career dialogue. For the non-managerial group of employees participation in a career dialogue did not strengthen the relationship between career control and perceived employability.\nPractical implications: Career control is beneficial for enhancing perceived employability among employees regardless of their position in the organization. Hence, training employees to master this competency may be a fruitful starting point for enhancing employability.\nOriginality/value: This is the first study to investigate whether the relation between career control, career dialogue and employability differs for employees with a managerial and a non-managerial role.\n", "role": "assistant" } ]
Transparency: a two-way mirror?
[ "Open management", "Corporate governance", "Utility programs", "Ethics", "Trust" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: For a few decades, the word transparency has been used very extensively, notably in the field of corporate governance. It has been endowed in most discourses with every virtue. This omnipresence and the lack of critical analysis has first got on our nerves, then raised our suspicion. Our paper therefore, aims to answer the following question: \"what do the prevailing discourses on transparency hide?\". This exploration is achieved through three steps. The first step aims to identify the origins of the concept, to evaluate the extent of its use and to discover the theoretical connections with managerial, economic and even philosophical literature. Second, section concentrates on the adverse effects of discourses on transparency: the dissimulation of a struggle for power and the expression of a disciplinary power. Third, points to the necessity of breaking away from the anti-humanist assumptions on which economic science has developed and which underpin most discourses on transparency. It suggests rebuilding management science on radically different bases, notably by a stronger emphasis on the concept of trust.\nTransparency: a pervasive concept: For about two decades, the word transparency has invaded a large range of discourses. It is especially pregnant in the debate on corporate governance (LARGO, 2002). But surprisingly, this extensive use does not go along with an accurate definition. Therefore, it is necessary to understand what is at stake, to explore the theoretical - economic and philosophical - roots of the concept.The expression of an emerging aesthetic\nThe exclusion of ethics: The usual reasoning is the following one. Individuals are opportunistic; they care only for their self-interest and if they think that profits are higher than costs, they will behave in a reprehensible way. The only means to prevent them from such actions is to exert on them a permanent surveillance; transparency appears as the best device to achieve this control. Reality is, however, more complex: the concept of transparency may be manipulated to serve the interests of a few; more generally, it leads, at least in the dominant discourses, to an exclusion of ethics and to anti-humanism.A door open to manipulation\nTowards an alternative theory: To break away from the dominant theories is not an easy task. As Charreaux (2002, p. 58), acknowledges \"the introduction of an alternative paradigm generates costs which are related to a more precise study of the underpinning theoretical frameworks and to the learning of new investigation methods, which differ from the usual economic modelisation and the traditional statistical tests\". More generally, this inversion in logic is confronted by all the obstacles, which have been noted by Kuhn (1970) in the course of scientific revolutions.However, the field has not remained entirely uninvestigated. In the philosophical perspective, we can refer to Habermas's (1987), Levinas' (1998) and Ricoeur's (1988) works. In the organisational field, we can go on with the researches conducted by the French Economist F. Perroux and his disciples (Bloch-Laine and Perroux, 1967 and 1968; Bloch-Laine, 1967; Perroux, 1973; Woot, 1968) who themselves were deeply influenced by Schumpeter and by the Roman Church's social doctrine (see Thomas Aquinas). These theoreticians adopt radically different assumptions regarding the ultimate aim of the enterprise and the nature of man. In his innovative but today unfortunately forgotten book, Pour une doctrine de l'entreprise (1968, for a doctrine of the enterprise) Ph. de Woot provides us with interesting insights. We will refer to his argumentation in the following sub-section.Alternative purposes for the enterprise[9]\nConclusion: In a period, which has been rich in huge scandals and notably financial scandals, the claim for transparency is swelling. It goes along with discourses marked by positive connotations and therefore not easily refutable: who could be against transparency? The concept then becomes an ideal without limits and its attractiveness \"discourages the auditor's resistance\" (Bourguignon, 2004). However, we have pointed to the adverse effects of extended transparency, and notably generalised amorality. We have demonstrated that it takes its roots in a very pessimistic conception of the nature of man, which reduced him to a calculating individual, alienated to his self-interest, a perspective, which indeed excludes ethics. To break away from such a pernicious conception, which has developed taking as departing point the concept of utility promoted by Bentham, we have to rebuild management science on alternative assumptions. Political and strategic participation (and not only objective participation) constitute two major conditions, which in their turn require trust. Trust in the full sense of the word and not the residue of erudite calculations.If you want to get rid of weeds, it is of little use cutting the flowers, the leaves, or even the stalk; you have to tear out the roots. At the beginning of this paper, we had a seductive flower called \"transparency\". We have identified its scientific name: \"information asymmetry\" and we have dug down to its ideological roots: \"utility\". But this kind of plant grows only in a specific kind of soil; we call it anti-humanism. Our paper aims to set down a few milestones towards radical humanism. Therefore, it takes for its own the French philosopher E. Morin's call \"to enter a politics of civilisation, a politics of humanism\" (Gibier, 2002).Figures 1 and 2 show the comparison of the two approaches which have been explored: the first one points to the ideological context in which dominant discourses on transparency are embedded and to its ultimate consequences; and the second one suggests an alternative model.\n", "role": "user" }, { "content": "Purpose: - Transparency is assumed to improve markets' efficiency, to enhance better corporate governance and finally, to ensure moralisation of business life. The paper aims to the realities that prevailing discourses on transparency dissimulate.\nDesign/methodology/approach: - First, a genealogical method is used to explore the theoretical roots, which underpin claims for a greater transparency (reduction of information asymmetry in academic discourses). Second, the analogies between mainstream theories on corporate governance (which legitimate the demand for transparency) and the Panopticon, the famous architecture conceived by Bentham are pinpointed. Finally, the reflections developed by philosophers such as Ricoeur and Henriot on ethics are exploited to explain why transparency is unable to produce the expected effects.\nFindings: - First, discourses on transparency often conceal reinsurance manoeuvres and power struggles. Second, transparency appears as a modernised manifestation of panopticism: a common base (the concept of utility), the same fundamental assumption (the opportunistic nature of man) and a similar answer (discipline). More generally, panopticism and discourses on transparency are embedded in the same utilitarian philosophy which does not grant any space either for responsibility or for ethics and therefore, favours generalised amorality.\nResearch limitations/implications: - As a consequence it is suggested that one should explore the possibility of elaborating an economics theory, which would get rid of the concept of utility and a theory of the enterprise, which would take as key concepts creativity and trust, instead of opportunism.\nOriginality/value: - Through the deconstruction of prevailing discourses on transparency, the paper points to the intrinsic anti-humanistic character of mainstream economics and organisation theories and suggests alternative issues. It therefore may contribute to both enlightenment and emancipatory aims (in the Habermasian sense).\n", "role": "assistant" } ]
The need to apply new theories to "Sport CSR"
[ "Culture (sociology)", "Corporate social responsibility" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Sport provides an interesting paradox; on the one hand, it generates much respect and even adulation through the endeavours and achievements of athletes and is widely recognised for providing many benefits to society. On the other, professional sportsmen (and sometimes women) are derided for taking performance-enhancing drugs, collusion to fix sports contests, involvement in violence and an increasing disconnection with \"normal\" society; administrators and those involved in the business of sport are even more discredited for being inept, corrupt, greedy and out of touch. Within this, where does sport corporate social responsibility (CSR) sit? As a neutral vehicle that mainly \"does good\" or one that attempts to pull the wool over critics eyes in attempts to distract attention away from the negative publicity surrounding sports (or is it both?) Although we feel that considerable progress has been made in mapping the growth of CSR in the sports industry, we feel that most analyses fall into portraying sport CSR in the first category.\nPolitical CSR: Political CSR actually emerged from instrumental viewpoints of CSR. Yet, importantly, it also recognizes that business is implicitly and explicitly engaged in political activity. This includes its interaction with a wide range of stakeholders and political institutions and all interactions should potentially be considered in any assessment of work that purport to relate to institution - public relations). Political CSR tends to be critical of many CSR analyses in that they display a normative vacuity, whereas political CSR also has a normative element to it. This means that analysis of CSR (including sport CSR) should actively attempt to examine potential negative social consequences of business actions on stakeholders at macro- and micro-levels. In other words, the narrow actions of CSR should not be analysed in isolation.\nApplying political CSR to sport CSR: For us, research into sport CSR should learn from at least two separate themes here. First, that rigorous and independent measurement of \"sport CSR\" happens as part of the research process and that such a measurement system needs to take into account the wider political ramifications of the businesses that support sport CSR. For example, CSR that is associated with professional sport clubs or mega events should not be viewed in isolation to the core business activities of that club or tournament. So, the impact of Liverpool Football Club's core activities on a wide range of stakeholders (including the local community) could be taken into consideration when a review of all their CSR projects is undertaken. Similarly, an event like the 2010 World Cup should review the activities of FIFA and its core partners along with the CSR that was associated with it (see for example, Levermore, 2011). Second, a stronger regulatory framework guided by a normative undertaking to improve the general governance of sport should be explicit in analyses of sport CSR. In some ways, the debates on ethics from within some parts of FIFA are an example of the recognition of such a debate.\nCritical CSR: Critical CSR extends from political CSR's focus on normative motives and concerns over the \"big picture\" and the regulatory environment. It draws on critical theory and its concern for emancipation, empowerment and an analysis of unequal power relations to critique CSR to express concern over:\nGround for applying critical CSR to sport CSR (deconstructing the nature of CSR): Being heavily influenced in our earlier academic careers by critical theory, we firmly believe that all CSR research should include some of the elements noted above. Unsurprisingly, the first three of the critical CSR issues raised above have been applied to sport CSR in our work (such as Levermore, 2013). We focus attention on the fourth aspect noted above with the hope that more detailed work will follow in all five areas, especially that of emancipation and empowerment.\n", "role": "user" }, { "content": "Purpose: This paper aims to highlight how critical theory and political CSR might be applied to deepen our examination of the complexities associated with 'sport CSR'. The debate on the use of corporate social responsibility (CSR) in the sports industry is starting to move beyond \"mapping the territory\", which characterized the initial examination of this new direction in CSR. This viewpoint suggests that it is time for \"sport CSR\" to turn to a range of CSR perspectives found in mainstream management debates as they are under-applied at the moment.\nDesign/methodology/approach: This paper examines the current state of research in sport CSR and offers a discussion on the possible ways to apply two under-utilised mainstream perspectives - political CSR and critical CSR - to sport CSR.\nFindings: A review of literature highlights how sport CSR has tended to pay insufficient attention to the maladies, dilemmas and broader structural concerns and political ramifications associated with sport CSR. This means that other viewpoints noted and applied in this journal, such as \"critical CSR\" and political CSR are largely neglected.\nOriginality/value: The value of this article lies in highlighting how critical theory and political CSR might be applied to deepen our examination of the complexities associated with \"sport CSR\".\n", "role": "assistant" } ]
The impact of R&D sourcing strategies on basic and developmental R&D in emerging economies
[ "Basic and developmental R&D", "Emerging economy firms", "R&D sourcing strategies", "Governance modes" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Different contexts have unique environmental features (Chen and Miller, 2007). For instance, developed economies are characterized by a strong institutional background and a higher level of technological development. By contrast, emerging economies typically have a weak institutional background and a lower level of technological development (Hoskisson et al., 2000). These differences across developed economy and emerging economy contexts provide different opportunities and problems, thus affecting firms differently depending on in which context they operate. For instance, emerging economies can provide lower cost supply and manufacturing opportunities whereas they can also cause problems, such as internal and external resource scarcity and limitations in absorptive capacity. Therefore, the characteristics of the emerging economy context might influence firms' R&D sourcing strategies and the nature of R&D. Extant research investigating the link between R&D sourcing strategies and types of R&D have not taken into account the external environment context in which firms are residing. Therefore, this study asks the following research question:\n2. Theory and hypotheses: 2.1 Emerging economy context\n3. Method: 3.1 Data\n4. Results: Table I reports descriptive statistics, such as means, standard deviations, minimum and maximum values of variables used in this study. It is important to show the distribution of offshore and domestic R&D variables in the sample. In this regard, Table I shows that 2 percent of firms in the sample undertake offshore R&D outsourcing activities and almost 1 percent undertakes offshore R&D insourcing, with standard deviations of 15 and 10, respectively. Regarding domestic R&D, Table I indicates that 8 percent of firms in the sample undertake domestic R&D outsourcing and only 1 percent undertakes domestic R&D insourcing, with standard deviations of 28 and 13, respectively. The distribution of these variables suggest that emerging economy firms leverage more from domestic sources compared to offshore ones due to limitations they have in accessing and utilizing external resources. The distribution of the variables is in line with previous studies that have investigated R&D sourcing (see Rodriguez and Nieto, 2016).\n5. Discussion and conclusion: Firms follow different R&D types, such as basic and developmental R&D. Different types of R&D require different level of heterogeneity and cognition in knowledge. Even though previous studies have investigated what determines firms' different R&D types (Shimizutani and Todo, 2008; Mudambi and Swift, 2014), the literature still needs some consideration. In addition, existing studies have focused on mostly developed economies. However, the context of emerging economies provides a new setting that enables us to investigate what determines emerging economy firms' R&D activities. Emerging economies lag behind in accessing important resources compared to developed economies. Therefore, the context of emerging economies creates difficulties in accessing knowledge sources for firms' residing in such contexts. This study thus extends existing research by looking at the effects of R&D sourcing strategies on two types of R&D: basic and developmental R&D from the perspective of emerging economies. The effects of R&D sourcing strategies on R&D types become important because these strategies produce different level of cognition and heterogeneity in knowledge inputs. In particular, R&D sourcing is important for emerging economy firms because it enables these firms to upgrade their knowledge bases. Moreover, firms need to be aware of the implications of these sourcing strategies whether it is abroad or home (Cassiman and Veugelers, 2002, 2006; Nagarajan and Mitchell, 1998). Previous research in this field, however, has been limited to the investigation of the impact of R&D sourcing on performance such as innovation, sales and exports (Bertrand, 2011; Mihalache et al., 2012; Ceci and Masciarelli, 2010; Steinberg et al., 2017). As Rodriguez and Nieto (2016) suggested, it is important to look at the effects of R&D sourcing strategies on different activities of a firm. Thus, this study analyzes the effects of R&D sourcing on basic and developmental R&D from the perspective of emerging economy firms. In addition, this study further downs this research area particularly by exploring the effects of governance modes of each R&D sourcing strategy - outsourcing vs insourcing - on basic and developmental R&D.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to investigate the impact of R&D sourcing strategies and their governance modes on basic and developmental R&D. Following the concept of cognitive distance, this research proposes that there are trade-offs between basic and developmental R&D when emerging economy firms engage in different R&D sourcing strategies. R&D sourcing can enable emerging economy firms to access different level of heterogeneity of knowledge inputs depending on the cognitive distance between the firm and its suppliers. Distance in cognition increases when firms obtain knowledge from abroad and independent suppliers in comparison to the acquisition of knowledge from home boundaries and affiliates.\nDesign/methodology/approach: Tobit maximum likelihood estimation approach is used.\nFindings: Using data from Turkish firms, this study finds out that offshore R&D with an outsourcing governance mode affects basic R&D. In contrast, domestic R&D with an insourcing mode influences developmental R&D.\nOriginality/value: This research extends recent efforts to better understand the determinants of different R&D types by examining offshore and domestic R&D together and by taking into account different governance modes of each R&D sourcing strategy. This study becomes important because it investigates this issue from the perspective of emerging economy firms.\n", "role": "assistant" } ]
Retail evolution model in emerging markets: apparel store formats in Brazil
[ "Brazil", "Emerging markets", "Apparel store formats", "Big middle theory", "Low income consumers", "Retail evolution model" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Diversification in retail formats has arisen as a result of the dynamic character of business models developed by retailers, who create, alter, and discard formats in response to changes in the competitive environment (Reynolds et al., 2007). In Brazil, the substantial increase in the purchasing power of low-income consumers has accelerated the pace of modernization within retailers located in the large urban unplanned shopping districts (USD) serving these segments. Apparel stores, a major component of the store mix of USDs, represent a wide diversity of formats, sizes, strategies, and levels of retail modernization. One can find traditional, outmoded, local, single-unit formats next to and intertwined with very modern retail formats of large national and international chains. They provide therefore a fertile ground for investigation of the diversification of store formats and the evolution cycle of modernization which is under way in the Brazilian retailing industry serving low-income consumers.\nConceptual background: Traditional and modern retailing in emerging markets\n\"Theories\" of retail evolution: Many different hypotheses have been proposed to offer useful frameworks to help our understanding of the evolution phenomenon of retail institutions and formats.\nStore formats and retail marketing mix: A retail format is a type of retail composition adopted by certain companies to satisfy consumers' needs (Levy and Weitz, 2011). Retail format classification can help to identify the competitors and the main changes that occur in the sector (Dunne et al., 2010). It has been recognized that definitions of store formats, or types, tend to be imprecise, overlap, and are often confusing (Gonzalez-Benito et al., 2005; Dawson, 2000). Retail stores which serve overlapping needs and target segments tend to be classified according to a certain store format. From the consumers' point of view, store formats within one industry (like apparel, food, or electronics) can be considered as competing categories that satisfy specific consumption needs or different shopping situations. As such, different retail formats characterize the retail competitive structure, and indicate how they attract distinct market segments (Gonzalez-Benito et al., 2005) due to their different value propositions. In response to changes in the consumer behaviour and in the competitive environment, retailers tend to create, alter and discard formats (Reynolds et al., 2007).\nThe context of this research: Over the last ten years, low-income Brazilians have experienced a substantial increase in their purchasing power. The minimum wage increased faster than inflation, and this has helped to promote a less unequal distribution of income in the country. To illustrate the striking rise in income, the minimum wage in 2000 of only about US$80 has almost quadrupled in the last decade. The rapid expansion of this new kind of \"consuming low-middle class\", not only in Brazil, but also in countries such as China and India, has transformed these large emerging countries into the leading consumer markets for many consumer goods industries.\nMethodology: The apparel stores located in Unplanned Shopping Centres (USDs) are the units of analysis in this work. These serve mainly low-income consumers. In this descriptive exploratory study, we adopted a judgment sampling approach to choose three representative USDs, reflecting centres of different sizes and locations in the city of Sao Paulo. Surveys were conducted in 108 apparel retailers within these three areas. In order to identify the different store formats, the instrument for observation was designed to capture the stores' marketing mix characteristics; a short questionnaire was developed to obtain complementary information from the stores' managers (or owners). The observations and interviews were conducted by a single researcher in order to avoid differences in perception that could create some personal bias in the analysis. One of the authors directly conducted the field work, strictly following a methodological criteria to objectively evaluate the stores characteristics. Cluster analysis was the statistical data treatment method used to derive taxonomy of store formats. Methodological details are presented below.\nResults - analysis and discussion: From the total marketing mix variables presented in Table I, the following variables were used in the cluster analysis, as they represent the major elements of the definition of store formats: assortment width, price, type of service, promotional signage, layout and presentation, and store size.\nRetail formats in emerging markets - analogies with the Big Middle model formats: Although the classification schemes are mainly descriptive, rather than explanatory tools, they can be useful in a relationship study or for hypothesis specification regarding the relationship between the objects of the classification (Bailey, 1994). Based on the taxonomy developed for the Brazilian apparel retailing stores and drawing from the existing literature about the evolution of retail formats, especially from the \"Big Middle\" concepts, analogies were derived comparing store formats, and tentative propositions or hypotheses were developed about the evolutionary life cycle of retailing formats, in transitioning emerging markets.\nFinal considerations: Considering the fast-growing economic importance of the new \"low-middle class\" of emerging markets, and also the relevance of apparel retailing, this paper has investigated the retail diversification phenomenon taking place in USDs in the city of Sao Paulo, where there is a wide diversity of formats, sizes, strategies, and levels of retail modernization in apparel stores. Based on the identification of the marketing mix variable characteristics, the study was successful in empirically deriving a taxonomy of apparel formats which operate in USDs. We identified four distinct clusters which portray the major types of store formats in the very diverse universe of Brazilian apparel retailing: \"Old Fashioned\", \"Low Price\" \"Specialized\", and \"Big Middle\". These represent not only a wide variety of retailing strategies, but also seem to be at different stages of retail evolution and modernization.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to develop a taxonomy to describe and synthesize the retail strategies and store formats in the fast-changing modernization context of emerging markets. It identifies how these different formats are related to the existing literature about retail evolution. The paper proposes analogies between the empirical findings and the Big Middle Theory model developed by Levy et al. (2005).\nDesign/methodology/approach: Structured observations and personal interviews were conducted with managers/owners of 108 apparel stores located in three unplanned shopping districts in Sao Paulo, Brazil. Cluster analysis was used to derive the store formats.\nFindings: The results identify four store formats - old fashioned, price focused, specialized, and consolidated - reflecting diverse marketing strategies and different stages in the retailing modernization cycle. There is a striking resemblance between these empirically derived formats and the four types of retail segments (Big Middle, Low price, Innovative, and in Trouble) proposed by the Big Middle model.\nResearch limitations/implications: There are limitations due to the exploratory nature of the research: the specific context (Sao Paulo, apparel retail sector) where the empirical data was collected restricts the generalization of the results to other situations; lack of precision in the appraisal of the variables might limit the research replicability. Nevertheless, this research contributes to expanding the scarce knowledge about the retailing phenomenon in emerging markets. An extension of the Big Middle model is suggested to depict the existing retail formats in emerging markets, offering therefore a theoretical contribution to the retail literature.\nPractical implications: The study provides retailers with a framework against which to map their format, competitive strategy, and stages of the modernization cycle, enabling them to better adjust the configuration of their marketing mix variables. It also offers a classification scheme of stores formats which will help to describe the apparel retailing industry in emerging markets.\nSocial implications: Public policies are suggested to technically assist the modernization process and survival of more vulnerable retail formats.\nOriginality/value: Considering the fast-growing economic importance of the new \"low-middle class\" of emerging markets, and also the relevance of apparel retailing, this research is relevant and unique because it helps to bridge a gap in the limited literature and knowledge in this area.\n", "role": "assistant" } ]
A destination's gastronomy as a means for holiday well-being
[ "Tourist behaviour", "Well-being", "Information sources", "Local food", "Destination gastronomy", "Meal" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The topic of well-being has entered a diverse set of fields with different interpretations, such as economics and national well-being (e.g. Everett, 2015), management and organisational well-being (e.g. Koopman et al., 2016), education and child well-being (e.g. Bernal, 2015) and quality of life and subjective well-being (e.g. Magyar-Moe et al., 2015).\nA destination's gastronomy: A destination's gastronomy is a socio-cultural element of a destination's cultural heritage that defines and mirrors the local culture, traditions and natural environment (Long, 2004). It is also an inherent element of each traveller's holiday (e.g. Quan and Wang, 2004). A destination's gastronomy thus has the potential to affect travellers' holiday well-being, which is considered to have multidimensional elements in this study, including physical, functional, emotional and social well-being aspects. Physical well-being is here perceived to derive from aspects such as convenience, nourishment and healthiness, but also food safety (Cohen and Avieli, 2004), which refers to food-related practices and food hygiene for preventing foodborne illnesses and diseases. Functional well-being is about what allows humans to function well (Magyar-Moe et al., 2015) and refers here to the daily meals that structure the holidaymakers' days for regular food intake and for gathering with family members and friends (e.g. Rozin et al., 2002). Emotional well-being is about what makes the human feel well and includes positive feelings, such as happiness, issues related to how the individual experiences various aspects of one's life and cognitive interpretations of how satisfied one is with the various aspects of life in general (e.g. Avni-Babad, 2010; Diener et al., 2006; Magyar-Moe et al., 2015). Emotional well-being relates to the emotional aspects of happiness and satisfaction gained by a destination's gastronomy. Social well-being comes from dining with family and friends, which is a major social activity, and from other social activities,such as social food preparation and co-cooking (e.g. Magyar-Moe et al., 2015).\nMethod: This study benefits from a quantitative research approach (Veal, 2011) examining travellers' pre-trip information sourcing about a destination's gastronomy and on-site meals for vacation well-being.\nResults: Demographics\nDiscussion and conclusions: This study contributes to well-being research by providing insights to two inevitable traveller activities related to destinations' gastronomy: pre-trip food information sourcing and the daily meals consumed.\n", "role": "user" }, { "content": "Purpose: To provide insights into holiday well-being, the purpose of this paper is to examine two inevitable traveller activities related to destinations' gastronomy: pre-trip food information sourcing and the daily meals consumed.\nDesign/methodology/approach: A survey was carried out among 243 Finnish travellers. The findings are based on univariate analysis (t-test, ANOVA and regression analysis).\nFindings: Pre-trip behaviour to ensure holiday well-being is based on travellers' interests in food, an emotional desire for a sense of safety and a functional desire for convenience, while they collect information from the internet and guidebooks about recommended food places and local food as well as food safety and price level. Travellers' place the highest importance on dinner for their holiday well-being, especially foodies - those travellers with a keen interest in food. Breakfast is the second most important meal contributing to holiday well-being.\nPractical implications: These findings inform destination marketing organisations about what food dimensions they should emphasise in destination gastronomy-related marketing communication for tour operators and hotel and local restaurants about the essence of dinner and breakfast for holiday well-being.\nOriginality/value: The study provides insights into the role of destinations' gastronomy in holiday well-being, which deserves to be studied in the current era of experiences and food interest.\n", "role": "assistant" } ]
Stakeholder engagement in sustainability accounting and reporting: A study of Australian local councils
[ "Stakeholder engagement", "Stakeholder theory", "Social", "Environmental", "Local councils", "Sustainability accounting and reporting" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The involvement of stakeholders in the accounting and reporting process enables organisations to identify and incorporate their material concerns, issues, perceptions, needs and expectations (ACCA, 2005; Bebbington et al., 2007; Fries et al., 2010; GRI, 2013; AccountAbility, 2015). Stakeholder engagement as defined by AccountAbility (2015, p. 6) is:The process used by an organisation to engage relevant stakeholders for a clear purpose to achieve accepted outcomes. It is now also recognised as a fundamental accountability mechanism, since it obliges an organisation to involve stakeholders in identifying, understanding and responding to sustainability issues and concerns, and to report, explain and be answerable to stakeholders for decisions, actions and performance.\n2. Stakeholders, engagement and stakeholder theory: Freeman (1984, p. 46) defines stakeholders as \"any group or individual who can affect or are affected by the achievement of the organisation's objectives\". This definition indicates the existence of a two-way impact relationship between an organisation and its stakeholders. First, stakeholders can affect organisational goals which implies that organisational performance can benefit from the activities and participation of its stakeholders (Gao and Zhang, 2001). This impact legitimises stakeholder engagement in decision making to achieve the best possible organisational performance. Second, stakeholders are affected by the achievement of organisational goals and this impact legitimises a stakeholder's right for stakeholder engagement (Gao and Zhang, 2001). The inclusion of a broad range of stakeholders enables organisations to recognise and address their social and environment impacts (Horisch et al., 2014). The pressure to communicate more proactively and more frequently with stakeholders demands an engagement with a diverse range of stakeholders through enhanced inclusiveness, partnership and dialogue (Andriof et al., 2003; Crane and Livesey, 2003). Organisations often do not deal with stakeholders on an individual basis, they address the demands of multiple stakeholder groups (Andriof et al., 2002, 2003).\n3. Sustainability accounting and reporting and the role of stakeholder engagement: 3.1 Sustainability accounting and reporting\n4. Research methods: The case study research strategy (Gomm et al., 2000; Stake, 2003; Simons, 2009; Swanborn, 2010; Yin, 2013) was used for this study. A case study is an empirical investigation undertaken to gain an in-depth understanding of a particular phenomenon (Yin, 2013). It is considered as both a process of inquiry of a case and the outcome of that inquiry (Stake, 2003). Case study was considered an appropriate research strategy for this study because:\n5. Findings: 5.1 The sustainability accounting and reporting process\n6. Discussion: The majority of the research on sustainability accounting and reporting has examined planning, accounting and reporting as three independent organisational practices (Schaltegger, 2012). The current study's findings suggest that a link between strategic planning, accounting and reporting, and that the involvement of stakeholders is essential to establish an efficient stakeholder-centric accountability system.\n7. Conclusion: This study has highlighted that stakeholder engagement plays a critical role in the sustainability accounting and reporting process. The qualitative research generalisations (Parker and Northcott, 2016) that arose from a study of specific local councils in Australia indicate that a range of stakeholders influence decisions concerning the development of strategic plans and sustainability indicators, and the measurement and communication of sustainability performance.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to examine how stakeholders are engaged in the sustainability accounting and reporting processes of Australian local councils.\nDesign/methodology/approach: Managerial stakeholder theory through the use of the notion of stakeholder salience provides a theoretical basis for exploring stakeholder engagement in the sustainability accounting and reporting process. Case study research was used to explore the stakeholder engagement practices of three Australian local councils. Data collection methods included interviews and document analysis.\nFindings: The findings of this research identified the importance of stakeholder engagement in the entire sustainability accounting and reporting process, the development of strategic plans and sustainability indicators, the measurement of sustainability performance and the preparation of sustainability reports.\nResearch limitations/implications: This study, by integrating the sustainability accounting and reporting literature with the stakeholder salience concepts of power, legitimacy, urgency and proximity, illustrates the critical role of stakeholder engagement in the sustainability accounting and reporting process of three local councils.\nPractical implications: This study has implications for public sector organisations (PSOs) and their stakeholders in relation to stakeholder engagement in sustainability accounting and reporting. The findings of this study will also be useful to corporations in understanding the importance of stakeholder engagement in sustainability accounting and reporting.\nSocial implications: The public sector is expected to be a leader in sustainability and this paper provides evidence of three councils who through their stakeholder engagement provide exemplars of useful practices that could be adopted by other entities.\nOriginality/value: Prior research in PSOs has primarily focused on the sustainability accounting and reporting process but has given limited consideration to the involvement of stakeholders. The focus on stakeholder engagement through the use of managerial stakeholder theory extends the role of stakeholders from merely being an audience for sustainability reports to an influential contributor in the sustainability accounting and reporting process.\n", "role": "assistant" } ]
Knowledge transfer from business schools to business organizations: the roles absorptive capacity, learning motivation, acquired knowledge and job autonomy
[ "Vietnam", "Absorptive capacity", "Knowledge transfer", "Job autonomy", "Learning motivation", "In-service training business students" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: In the age of knowledge economy, knowledge is a crucial asset that creates a firm's sustainable competitive advantage (Nonaka, 1991). Nonaka observes the following:\nTheory and hypotheses: Theoretical background\nResearch methods: Research context and sample\nData analysis and results: As previously mentioned, this study used CFA to validate the measures of the constructs and then employed SEM to test the proposed theoretical model and hypotheses. The maximum likelihood estimation method was used to estimate the parameters in the CFA and SEM models. This estimation method was employed because the screening process revealed that although the data exhibited slight deviations from normality, all of the univariate kurtosis and skewnesses were within the range of [-1, 1] (Muthen and Kaplan, 1985).\nDiscussion and implications: Recognizing that in-service training business students can serve as a channel as well as a source of knowledge transfer from business schools to business organizations, this study investigated some key determinants of such a type of knowledge transfer. Based on the AMO framework, this study proposed that both students' absorptive capacity and learning motivation will enhance the knowledge acquired by students from their business schools. Subsequently, students' acquired knowledge together with their absorptive capacity and learning motivation will affect their knowledge transfer from business schools to business organizations. In addition, job autonomy plays the role of a mixed moderator in the relationship between acquired knowledge and knowledge transfer. The study results, based on a data set collected from a sample of 344 in-service training business students in Ho Chi Minh City, Vietnam, revealed that both students' absorptive capacity and learning motivation underlie students' acquired knowledge, and acquired knowledge is a determinant of knowledge transfer. In addition, students' learning motivation enhances knowledge transfer but students' absorptive capacity does not. Further, job autonomy does play the role of a mixed moderator, that is, it has a positive effect on knowledge transfer and also strengthens the impact of acquired knowledge on knowledge transfer. These findings offer a number of implications for theory, research and practice.\nConclusions, limitations and directions for future research: In conclusion, in-service training business students play a dual role in the transfer of knowledge from business schools to business organizations: a channel as well as a source of knowledge transfer. This dual role has largely been ignored in the literature on university-to-industry knowledge transfer. In transitioning markets like Vietnam, where business management practice of business organizations still lags behind that of organizations in advanced economies, in-service training business students, together with other channels such as consulting contracts, are critical sources for enhancing business management practice of Vietnamese business organizations. The findings of this study suggest that students' absorptive capacity, students' learning motivation and organizations' job autonomy are critical for effective knowledge transfer from business schools to business organizations. Hence, participating parties - business schools, business organizations and in-service training business students - should pay attention to these key factors and design and implement appropriate strategies and action programs to enhance the effectiveness of this type of knowledge transfer.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to investigate the impacts of absorptive capacity, learning motivation and acquired knowledge on knowledge transfer from business schools to business organizations, as it has been realized that in-service training business students can serve as a channel as well as a source in this type of knowledge transfer. The study also examines the mixed moderating role of job autonomy in the relationship between acquired knowledge and knowledge transfer.\nDesign/methodology/approach: A sample of 344 in-service training business students in Vietnam was surveyed to collect data. Confirmatory factor analysis was used to validate the measures, and structural equation modeling was used to test the hypotheses.\nFindings: The findings reveal that absorptive capacity positively affects acquired knowledge but it does not have an effect on knowledge transfer. Learning motivation, however, has positive effects on both acquired knowledge and knowledge transfer. The findings also indicate that acquired knowledge is a determinant of knowledge transfer. Finally, job autonomy plays the role of a mixed moderator in the relationship between acquired knowledge and knowledge transfer.\nPractical implications: This study signals the participating parties -business schools, business organizations and in-service training business students -that absorptive capacity, learning motivation, acquired knowledge and job autonomy are critical to the transfer of knowledge from business schools to business organizations.\nOriginality/value: This study contributes to the literature on university-to-industry knowledge transfer by providing empirical evidence for key determinants of knowledge transfer from business schools to business organizations through a new channel of knowledge transfer - in-service training business students.\n", "role": "assistant" } ]
Correlates of career priority and family priority among hospital-based nursing staff
[ "Women", "Career priority", "Family priority", "Nursing staff" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: There has been increasing interest in the career prospects of professional and managerial women as organizations strive to increase their human capital (Wittenberg-Cox and Maitland, 2008). In an increasingly competitive business environment, organizations cannot afford to ignore half the population's talents (Burke and Mattis, 2005; Michaels et al., 2001). Yet, women continue to have difficulty advancing their careers (Barreto et al., 2009; Davidson and Burke, 2011) in both developed and developing countries.\nAn overview of the literature: In efforts to both understand and support the career advancement of women, Schwartz (1992) suggested that organizations should offer women with young children, who prefer to do so, a less demanding role for a period of time so they can fill family obligations then returning to more demanding roles when their children get older. Women without children would continue to hold demanding jobs and careers similar to those held by ambitious men. She referred to the former as career-family women and the latter as career-primary women. Schwartz, perhaps as an oversight, did not make this same distinction among men.\nMethod: Procedure\nResults: Descriptive statistics\nDiscussion: This study explored the relationship of career priority and family priority among nursing staff, an almost exclusively female occupation, with a variety of personal demographics, work situation characteristics, work motivations, work outcomes and indicators of psychological well-being.\n", "role": "user" }, { "content": "Purpose: This study explored the relationship of measures of career priority and family priority with a number of other variables including personal demographics, work situation characteristics, work motivations, work outcomes and indicators of extra-work outcomes such as affluence and psychological well-being. The paper aims to discuss these issues.\nDesign/methodology/approach: Data were collected from 290 nursing staff, the vast majority female, working in Ontario, Canada, using anonymously completed questionnaires.\nFindings: Career priority and family priority were significantly and positively correlated in this sample. Nursing staff also rated family priority significantly higher than career priority. Personal demographics were associated with levels of both career priority and family priority such that married nursing staff, nursing staff with children, and nursing staff working part time reported lower levels of career priority, while married nurses and nursing staff having children rated family priority higher. Nursing staff having higher levels of work motivation also rated career priority higher. Career priority was significantly correlated with several work outcomes. Nursing staff indicating a higher career priority were more satisfied and engaged in their jobs. Somewhat surprisingly, family priority was generally unrelated to these work and well-being outcomes.\nResearch limitations/implications: Recent writing on women in organizations has raised the question of can women \"have it all\", a successful and demanding career and a satisfying home and family life. The findings contribute to this debate.\nPractical implications: Suggestions for both women and organizations to facilitate career and family facilitation are offered.\nSocial implications: Increasing interest has been shown in women in the workplace, and whether they should \"lean in\" to advance their careers. The authors suggest that this strategy may be at odds with what women, and men, increasingly want.\nOriginality/value: The paper highlights differences in the antecedents and consequences of career priority and family priority in a predominately female sample bringing work and family issues into the forefront once again.\n", "role": "assistant" } ]
Beekeeping as a family artisan entrepreneurship business
[ "Family business", "Artisan entrepreneurship", "Beekeeping" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper focusses on ethnic Albanians and Macedonians who are beekeepers in the Republic of Macedonia, a transition economy which was formerly a part of the Yugoslavia federation. While the importance of entrepreneurship impacting the economic development of a region is well understood, the risks - uncertainty and issues in creating and growing new businesses - are so substantial that a significant number of all started businesses are bought, gathered into another company, or fail within the first five years (Hisrich et al., 2017). This is particularly the case in a developing economy like Macedonia.\nLiterature review: Artisan entrepreneurship\nMethodology: The research plan employed followed the mentor model, with authoritative local individuals assisting in the research process. The authors were first introduced to Tetovo's beekeepers by Jonuz M., a retired Teacher and the President of the Beekeepers' Association of Tetovo. Participants were selected using snowball sampling (Goodman, 1961; Muller-Wille and Hukkinen, 1999). As discussed by Goodman (1961), interviewing concludes when the last respondent can suggest only individuals already named by others. \"Asking questions and getting answers is a much harder task than it may seem at first\" (Fontana and Frey, 2003, p. 61). It was also important to know what questions that should not be asked (Dana and Riseth, 2011). It was suggested by Mr Jonuz to avoid questions about the amount of revenues, profits, and kilograms of honey produced, sensitive issues in Tetovo, and the entire Macedonian culture. The majority of participants in this study requested anonymity, and a few allowed their responses to be presented by first name only.\nFindings and discussion: In 2013, the majority of the Macedonia beekeeping, 61 per cent, started their business out of necessity, and 23 per cent did so because of perceived opportunities (GEM Macedonia, 2014). In total, 5 of interviewed beekeepers were motivated to become full-time beekeepers because they had no other alternatives; 16 were continuing the family tradition; and 9 became beekeepers because of the perceived opportunity for a successful, profitable business. Those who emphasised beekeeping to be a good opportunity for profitable business were Macedonians, while those beekeepers continuing the family tradition or viewed as a \"necessity\" were mostly ethnic Albanians. Macedonians dominate in all economic and social aspects, and can be employed in state and public institutions. Albanians had two alternatives: leave the country or start their own businesses. Some ethnic Albanians left Macedonia and others decided to do beekeeping as an entrepreneurial business. These differences in conditions are discussed broadly by Bexheti (2006) and Zeqiri (2005).\nConclusion: This research focussed on the beekeepers in Macedonia. The results have indicated that beekeeping entrepreneurship is very important to customers and plays a significant role in the transition economies. Beekeeping can provide additional income that keeps rural families financially stable. Or as in the case of Macedonia, beekeeping has become popular among citizens regardless of their ethnicity. Economic need forces Macedonian people to rely on self-employment to have enough family income. This self-employment engages other family members who are capable growing the entrepreneurial business. In order to improve and advance their business in the community, skills and knowledge are exchanged. Through this, the beekeepers are able to earn extra income for their families.\n", "role": "user" }, { "content": "Purpose: Throughout Macedonia, beekeeping is becoming popular regardless of ethnicity. Studying ethnicity, the purpose of this paper is to determine what beekeepers in Macedonia thought in their own words about their beekeeping entrepreneurship. The objective is to identify whether motivations of ethnic Albanian beekeepers in Macedonia were the same or different compared to those of ethnic Macedonians in the same country, and if different, how.\nDesign/methodology/approach: To accomplish this objective, in-depth interviews were conducted with 40 beekeepers in Macedonia. A total of 29 interviews were conducted face-to-face and the other 11 by phone. The first set of interviews took place between December 2016 and February 2017, followed by more interviews in June 2017. In total, 27 respondents said they were ethnic Albanians, and 13 identified themselves as ethnic Macedonians. Also, ten respondents were women. While eight were full-time beekeepers, 32 were part-time beekeepers.\nFindings: The results indicated that beekeeping businesses play a significant role in the transition economy of Macedonia. Beekeeping provides additional earnings that support rural families and keeps them financially stable. The majority of both Albanians and Macedonians understood that beekeeping on a part-time job basis provided a needed supplement to their income. Some part-time beekeepers are also working as auto-mechanics, locksmiths, medical doctors, restaurant/cafeteria owners, and tailors. A few in the sample were retired from their jobs or full-time beekeepers. An important difference between ethnic Albanian beekeepers and ethnic Macedonians in Macedonia is that the majority of ethnic Albanian participants see beekeeping as following in \"my father's footsteps\", while most Macedonians were motivated by the perceived opportunity of having a good business.\nResearch limitations/implications: Limitations of the research are twofold. First, financial data of family beekeeping are not available, which would be useful in determining the contribution made to economic development. It is common, especially in transition economies such as the western Balkans, that financial results are very sensitive to their owners. Second, unavailable databases for beekeepers make any quantitative approach difficult, if not impossible, resulting in most research using the qualitative research approach.\nOriginality/value: This paper is one of the first to treat beekeeping as a form of artisan entrepreneurship, which also contributes to the understanding of family business. As in other countries, the important and operation of the family business among family members in Macedonia is passed from generation to generation. The results of this research revealed the value of networking, which was found to be very important to income. For beekeepers to develop, grow, and be branded in the community, networking is an important ingredient.\n", "role": "assistant" } ]
Age and work-related stress: a review and meta-analysis
[ "Meta-analysis", "Aging workforce", "Irritation", "Life-span psychology", "Work stress" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: As a consequence of ongoing demographic changes, the workforce is becoming older in many industrialized countries. In the USA, for instance, the employment rate of older workers rose from 32.4 percent to 37.2 percent between 2000 and 2005 (US Bureau of Labor Statistics, 2006). Similarly, in Europe, employment rates of older workers increased from 36.2 percent in 1998 to 45.6 percent in 2008 (European Commission, 2010). These developments have motivated many researchers to examine age differences of work-related variables (e.g. Baltes and Finkelstein, 2011). As a result, substantial knowledge has accumulated regarding the relationship between age and several important work-related outcome variables, such as job performance (McEvoy and Cascio, 1989; Ng and Feldman, 2008; Sturman, 2003; Waldman and Avolio, 1986), job attitudes (Ng and Feldman, 2010), job satisfaction (Brush et al., 1987), organizational commitment (Cohen, 1993), voluntary turnover (Healy et al., 1995), or absenteeism (Martocchio, 1989).\nLiterature review: Age differences in the transactional stress model\nMeta-analysis: In the current meta-analysis, we focus on irritation (e.g. Mohr et al., 2006) as a measure of short-term consequences of stress at work. In contrast to long-term consequences such as depression, psychosomatic disorders, or burnout (e.g. Dormann and Zapf, 2002; Ng and Feldman, 2010), short-term effects of stress can be detected earlier and might therefore provide more opportunities for timely interventions. Furthermore, while many other measures of strain experience are not work-specific (e.g. anxiety, depression, psychosomatic complaints, subjective health; for a recent review on work stress research, see, for example, Hausser et al., 2010), the concept of irritation has been developed within the work context and reflects both cognitive and emotional reactions to work stressors (Mohr et al., 2006). According to its definition, irritation is caused by the perceived non-attainment of work-related goals (Muller et al., 2004). Irritation includes both ruminations (cognitive irritation) about the work as well as irritability (emotional irritation). Rumination indicates \"unintentional preservative thoughts in the absence of obvious external cues\" (Cropley and Millward Purvis, 2003, p. 197). Irritability describes \"a state of losing the incentive to achieve a certain goal after intense efforts toward goal achievement have not led to diminished goal discrepancy\" (Mohr et al., 2006, p. 199). Irritation was shown to be positively related to various task-related stressors (e.g. Grebner et al., 2003), social stressors (e.g. Frese and Zapf, 1987), and emotional stressors (e.g. Zapf et al., 1999).\nMethod: Literature search\nResults: The meta-analytical results for the relationship between age and irritation are presented in Table I. Consistent with H1, age was unrelated to irritation (r=0.02). Regarding the facets of the irritation construct, emotional irritation was also uncorrelated with age, (r=0.01), whereas cognitive irritation was significantly positively related to age (r=0.10), indicating that older workers reported higher cognitive irritation than younger workers. Testing for curvilinear relations, no quadratic effect of age on irritation (r=0.02) or emotional irritation (r=-0.01) was found. However, a significant quadratic effect was observed between age and cognitive irritation (r=0.06), showing that older and younger workers reported higher scores of cognitive irritation than middle-aged workers. Since all of the weighted correlation estimates were non-homogeneous (cf. Table I), moderator analyses were conducted (cf. Borenstein et al., 2009).\nDiscussion: In order to examine age differences in short-term stress and strain reactions, the current meta-analytic review connects theory-based and differentiated process analysis with a data-driven approach. Both the literature review and the meta-analysis of the available empirical data support our first hypothesis that age per se is unrelated to overall reports of strain at work. This is consistent with the idea that different age-related losses and gains might compensate each other. However, the cognitive facet of irritation was related to age differences, reflecting that young and - in particular - older workers reported higher levels of cognitive irritation than middle-aged workers. Thus, compensation of age-related losses via higher emotion regulation skills might be more relevant for irritability.\n", "role": "user" }, { "content": "Purpose: The ongoing demographic changes in many industrialized countries affect managerial decisions in many ways, and require sound knowledge of systematic age differences in central work-related variables. The current paper aims to address age differences in the experience of work-related stress. Based on life-span approaches, the authors focus on age differences in different components of the work-related stress process and meta-analyze existing empirical studies on the relationship between age and short-term indicators of work-related stress (i.e. irritation).\nDesign/methodology/approach: The authors conduct both a literature review and a meta-analysis of age and indicators of work-related stress.\nFindings: The literature review revealed that age might affect several components of the stress process at work. However, as these effects are partly conflicting, they might nullify each other in the overall relation between age and stress. Indeed, the conducted meta-analysis showed no general correlation between age and irritation as a short-term indicator of work-related stress. Instead, this relationship was significantly moderated by type of occupation and gender.\nResearch limitations/implications: The meta-analytic results are limited to short-term indicators of stress. Based on both the literature review and the meta-analytical findings, the authors introduce a research agenda for future research, including a call for more thorough research on the whole work-stress process and the integration of life-span theories.\nPractical implications: A more differentiated understanding of age differences in different stages of the stress process at work facilitates the implementation of age-differentiated stress prevention and intervention strategies.\nOriginality/value: This study is the first meta-analysis on the relationship between age and short-term consequences of work-related stress.\n", "role": "assistant" } ]
Innovation is good, fitness is better
[ "Business improvement", "Design", "Corporate strategy", "Innovation" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The struggle to succeed in business spawns any number of principles, beliefs, new notions, and adherence to past great ideas. Over time, how many of these principles persist? Is there a link between survival and the quality of an idea?\nCould it be that simple?: Growth is indeed an appealing topic. Companies have worked hard to build lean enterprises so that if they grow on this base of operational improvements, the profits flow right to the bottom line. There is great incentive to grow, but the challenge is how, and where, and with what. These are tough questions.I want to argue in this paper that while two of the imperatives a business faces are to reinvent itself and to be more innovative, that business does indeed have a higher purpose.A hit in business can be a monumental moment - establishing its success for a long time, but is that hit, or series of hits, an indication of how long a company will survive?To survive, it has to stay \"fit.\" Innovation is a key atom - but not the only atom - inside the molecule of fitness. By understanding this higher purpose, a business can incorporate innovation into its culture while understanding the other dynamics that may dull the senses to emerging innovation.\nSurvival of the fittest is not a new idea: Finding the broader concept of fitness appropriate for this discussion has a serendipitous connection. We are about to reach the 150th anniversary of The Origin of Species by Charles Darwin. This landmark work still generates controversy today. What is indisputable is that species come and go, thrive and die. In fact, it may be helpful to consider the full title of Darwin's work to understand its predatory cycle: The Origin of Species by Means of Natural Selection, Or, The Preservation of Favoured Races in the Struggle for Life.Thinking of fitness as the framework, we can acknowledge that business is also a complex system, so Darwin's work applies: a company either will thrive and survive or be dominated by a better system. This is why people like Steve Jobs are present-day business heroes. Apple is regularly producing \"hits\" around innovation while mastering the ability to embed a fitness mentality into the corporate culture. It is as though the struggle to compete with them is not even fair.Innovation can be an altruistic principle that is problematic when the innovative idea fails to take hold. Investing in innovation can be attractive but off-track if not coupled with design thinking. We have to do more than imagine the new ideas. We have to design them to interact in a harmonized way that, once executed, they are judged as a sustainable, competitive advantage. Otherwise, as the cycle of fitness suggests, you will become fodder for someone else's concept.Beyond the well-used example of Apple, you can actually see the fitness of new models overtaking other well-honed models that preceded them. For example:*FedEx inventing a faster and more reliable system for movement of important mail when it absolutely had to be there overnight.*Southwest Airlines building a profitable point-to-point flying model in the world of the hub and spoke.*Amazon using the community of users not only to sell and deliver books via the web, but also to provide customer reviews and suggestions for further reading.In each of these cases, the system that these companies built was more fit than the one that predated it, with the parts and pieces harmonized in a new way to address complexity.History shows us that this has always been the truth. Just as the US Postal System demonstrated fitness when it replaced the Pony Express, it, too, was eventually overcome by a system that was more fit. Darwin's model suggests that complex systems go through these epochs of domination, making incremental improvements for survival at the top, before a better model of fitness arrives. In the natural world, these eras can last millions of years. In business, we see that the cycle is getting shorter all the time.The competitive environment is relentless and there will always be the threat of something stronger lurking around the corner. Complex systems are fundamentally programmed to thrive by fighting for advantage. There is no other choice.Companies that rise to the level of being dominant enjoy the fruits of being number one, but they are also locked into a predictive cycle that is hard to shake. It is a perversity of fitness that can be depressing. Dominant systems have a hard time seeing that they need to change, now that the virtues that made them successful threaten to weigh them down and prevent their progression. It is as though the intoxication of what made us great obscures us from seeing what is needed for continued dominance.\nDesign thinking as a path to fitness: If I were to stop here, this article would sound much too pessimistic. I am a huge believer in investing in opportunity, and I think we can consciously make the systems we represent more fit - and able to thrive for a very long time. It requires more than just \"cheering for the home team\" or hoping that the good times continue. That won't be enough to overcome the essential truth that the system loses its advantage over time.We have access to frameworks and principles that allow us to tune complex systems and actually take advantage of their complexity. One of these approaches, design thinking, is the business equivalent of the biological virtues in nature that Darwin explored.Design thinking requires that we consider every possible dimension and every affected constituency in designing a solution, whether the end result is a product, an experience or a new business model. It also requires that we embrace the complexity within our systems because the complex elements add richness that cannot be found in simpler models. If you drive too hard for simplicity - and it is a natural tendency to do so - you may miss some of the elements that lead to ultimate fitness. By using design thinking, we can grasp the potential within the complex systems we have created and in which we compete, and then drive for innovation that helps our business thrive, grow and, most importantly, survive.When organizations fail to move beyond innovation to fitness, it is because they did not adopt design thinking as a way to get there. The fundamental gap is not the lack of appreciation for what looks good. Design thinking has everything to do with elegance, including, but not limited to, the visual representation. Finding the elegance in the harmony of the elements of a complex system is vital.I can speak about my experiences in running Steelcase since 1994. Our current board chair, Rob Pew, introduced me to design thinking about 20 years ago, when I thought that design thinking meant making things more beautiful. Rob and I dreamed about how we could use this approach to grow the business. He encouraged me to visit the Illinois Institute of Technology Institute of Design in Chicago to build my understanding of the concepts. In fact, I joined their board of overseers and learned a great deal. This board was made up of individuals who had built better products and better business models and developed more fit competitors in their chosen industries - and I would certainly say they were innovative, as well.Only later did I realize that design thinking gave me the tools, or frameworks, to work with very complex problems - problems that seemed to have existed for years with no hope for change or evolution.At one time, the Steelcase industrial model was the envy of the office furniture industry. We had the capacity to respond to customer requests of any magnitude. The true story of how Steelcase - truckload after truckload, day after day - filled the newly constructed Sears Tower in Chicago continues to be part of corporate lore. However, increased competition and the demands of a global marketplace required that we make major changes in this vertically integrated model. We were faced with an urgent need to reinvent our industrial system.To say it is a complex system is an understatement. For example, because of the range of product lines, sizes, colors and other options, our customers could configure our products in over 22 quadrillion standard offerings plus the nearly unlimited number of special requests! Over a period of five years, we established a global supply chain and reduced manufacturing floor space by more than 30 percent - while reducing manufacturing lead times.We had to overcome the thought that it was our manufacturing system which made Steelcase strong in the first place, and we probably did not move as quickly as we might have while overcoming this inertia. W.H. Auden depicts the human challenge in understanding the impact of systems fighting for fitness:We would rather be ruined than changed;We would rather die in our dreadThan climb the cross of the momentAnd let our illusions dieThus, we have to believe that even as we might be dominant today, we have to continue to seek the next level of fitness. At Steelcase, this means applying the lessons we have learned in the redesign of our manufacturing model to the front end of our business - including the way we design and market our products and administer our global company.Steelcase is the global industry leader, not an upstart trying to catch up with the big players, and we see examples of leaders in other industries that have demonstrated their willingness to create a more fit system:*Marriott Corporation once believed that they should own their real estate and develop the hotels on those properties. They refocused on the core aspect of running successful hotel brands while reducing the emphasis on land acquisition.*The benefits are not fully realized yet, but Boeing has created a new lean model of production in order to compete with a European model that disaggregated the supply chain. Boeing's system for building the 787 will create, I believe, the most commercially successful airplane in history.*The Dayton Hudson department store chain, one of the more successful mall anchor stores, was reborn under the Target brand. By focusing on thoughtful design of mainstream products, Target redefined fitness in traditional retailing.\nDesign thinking delivers frameworks for solutions: The quest for fitness also requires that companies pay homage to the exercise of deep thinking. At Steelcase we have adopted a design framework for critcal thinking (see Figure 1), and I wrote about it in greater detail in the April 2007 issue of the Harvard Business Review (Hackett, 2007). It is not in the normal DNA of companies to leverage the collective power of their organizations to think intensely about the complexity of the system that surrounds that business.As Steelcase we know that fit product design comes from knowledge about use. We have to be willing to scour for insights into what the users of our products and other key constituents need. Nothing as simple as a questionnaire or focus group will give us the necessary level of understanding. Instead, we must use a variety of intense ethnographic tools to develop insights into how people work and how they interact in the workplace.Systems design also comes from knowledge. The \"think\" phase - the first of four steps in our critical thinking framework - requires that we consider the impact on every dimension of our organization. This should lead us to the second phase, establishment of a point of view. Too often in companies, the process is reversed. The thinking that is done is in support of a previously established viewpoint, and the complexity of the system is not fully addressed.The final two steps of our critical thinking process - developing a fully realized plan to implement, followed only then by implementation - is intended to help people have confidence as they wade through the complexity of any problem. The aggressiveness of competition can blur our sense of the bigger picture and cause us to lose confidence that we can raise our level of fitness.Now the circle is complete: having a culture that values knowledge and embraces design, which will allow it to be more innovative and can then fuel fitness - and fitness is the key to survival in a flat and highly competitive world.\n", "role": "user" }, { "content": "Purpose: - Academic journals and the popular business press are filled with articles praising innovation. But innovation is not enough. This paper proposes a higher purpose - that survival of our business systems relies on achieving a new level of fitness - and aims to suggest that design thinking is one of the keys to becoming more fit.\nDesign/methodology/approach: - Ever since Darwin argued \"survival of the fittest,\" social scientists have worked to determine winners and losers in a wide range of complex systems. The paper asserts that there is value in applying the concept to the survival of business systems and provides examples of how successful companies have worked to overcome inertia and reinvent portions of their business model.\nFindings: - The paper finds that dominant systems have a hard time accepting the need for change, even though the very attributes that made them successful now threaten to drag down their long-term fitness. Design thinking is a framework that provides the opportunity to study all aspects of complex issues, and instead of driving for simplicity, design thinking requires that we embrace the richness that complexity provides.\nOriginality/value: - By considering the need to move beyond innovation, the paper provides insight into the critical thinking needed to drive for a higher level of fitness in a business system and offers a brief look at the model used by one industry-leading company.\n", "role": "assistant" } ]
Personal networks and knowledge transfer in inter-organizational networks
[ "Knowledge transfer", "Best practice", "Learning organizations", "Information management" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Some work carried out on the topic of inter-firm knowledge transfer show that interpersonal relations are of great importance to inter-firm exchanges (Hansen, 1999; Ingram and Roberts, 2000; Reagans and McEvily, 2003; Rodan and Galunic, 2004; Uzzi and Lancaster, 2003). However, one notes that these studies assume a direct impact of social networks on knowledge transfer, supposing implicitly that social networks generate resources that are useful to knowledge transfer. Consequently, the effects of social networks on knowledge transfer are assimilated to a \"black box\" (Chollet, 2005) that we propose to open in this study.We seek to answer the following question: what are the intermediate mechanisms through which personal networks influence the transfer of good practices between firms member of a network? It is suggested that personal networks influence the transfer of good practices between member firms of an inter-organizational network (ION) by allowing key individuals access a set of strategic resources on the one hand, and improve their absorptive capacity of external knowledge, on the other hand.We collected data from firms member of export consortia in Tunisia. Using a questionnaire, a face-to-face survey was conducted with individuals that are the most involved in their firms' participation in the consortium. The collected data were analyzed by the structural equations approach based on partial least squared-path modelling (PLS-PM) techniques.Results show that the transfer of good practices is primarily explained by the individual's absorptive capacity. Our analyses also show that the dimensions of a personal network affect differently the access to strategic resources and the individual's absorptive capacity. Thus, the strategic resource access is explained positively by the size of the personal network and negatively by the strength of ties established among network members. Absorptive capacity is explained negatively by indirect ties and by the strength of ties; and positively by the range of ties. The results thus support that the absorptive capacity constitutes the principal mechanism through which personal networks influence the inter-firm transfer of good practices. The absorptive capacity has not only a direct influence on the transfer of good practices but also a potential indirect impact through its effect on strategic resource access.This study hopes to reduce the gap between the strategic perspective and the social network perspective when studying IONs. It is a question of finding the means of cross-fertilization of the two perspectives and participating in the founding of an ION theory.The remainder of this paper is organized as follows. In the next section, we present a literature review on personal networks and the two mechanisms through which it is likely to influence good practice transfer. Section 3, we present some theoretical arguments and develop hypotheses on the ways that personal network variables are expected to affect strategic resource access and absorptive capacity development and the way these latter influence good practice transfer among firms. Section 4, we describe our sample, present our data collection process and our variables measures. Section 5, we present our data analysis and empirical results. Section 6, we discuss our findings, present some implications based on these findings, and provide some ideas for future research. In the final section, we draw up the conclusion of our research.\n2 Personal network and transfer of good practices in IONs: The transfer of knowledge between organizations becomes increasingly complex because of the dispersal of units over distance and the high costs of coordination among them. According to Huber (1991), this resulted in a new requirement: the coding and the transfer of the most useful knowledge. Thus, knowledge takes the form of \"good practices\" (Perrin, 2006) which refer to any techniques, methodologies or processes which were applied in an organization and have shown higher results. The identification and transfer of best practices is emerging as one of the most important and widespread practical management issues of the latter half of the 1990s. At an individual level, interpersonal relations appear of great importance for inter-firm transfer of good practices.2.1 Personal network\n3 Research framework: hypotheses and design: In a first step, our research framework relates the different aspects of the personal network to the strategic resource access and to the absorptive capacity development. In a second step, it relates resource access and absorptive capacity to the inter-organizational transfer of good practices.3.1 Personal network, access to strategic resources and absorptive capacity\n4 Research methodology: We describe the research sample, the data collection process and the variables' measures.4.1 The sample\n5 Analysis and results: To test our hypotheses, we used the structural equations approach based on the PLS-PM technique available in XLSTAT software (Tenenhaus et al., 2005). The techniques based on PLS enabled us to effectively treat samples from 30 to 100 observations. The results of the PLS-PM analysis include a first part for the study of the measurement model, and a second part for the study of the structural model.5.1 The study of the measurement model: description of the data\n6 Discussion: The present paper has examined the patterns of resources and absorptive capacity accessed via personnel networks of key individuals in SMEs members of export consortia in Tunisia. Our paper has also checked the effect of these resources and this specific capacity on the transfer of good practices between firms. Globally, findings show that the configuration of personal networks that lead to resource access is different from that leading to absorptive capacity development and that only the absorptive capacity significantly influences the inter-organizational transfer of good practices.6.1 Personal networks, resource access and absorptive capacity\n7 Conclusion: We have shown that absorptive capacity constitutes an important mechanism through which personal networks influence the inter-firm transfer of good practices. Through interaction with persons who hold diverse types of knowledge, individuals gain access to different ideas and mobilize these ideas by discussing and using them, ultimately transforming the promising ones into adapted good practices for better achieving their work (Mu et al., 2008).We also show that the absorptive capacity, like some aspects of personal networks (size and strength of ties) favour the exchange of strategic resources between actors, but these resources are not converted into good practice transfers. Thus, we cannot conclude as regards the role played by the strategic resources studied in mediating the influence of personal networks on good practice transfer.However, we can conclude that additional efforts, at the practical level, must be undertaken in order to direct resource exchanges towards the transfer of good practices objectives. These objectives remain misunderstood for a large number of SMEs employees. The reason is that, in general, the participation in consortia is justified by the access to subsidies granted by the ONUDI rather than by the search for synergy. For this purpose, our research emphasizes the need to provide the consortia directors and the managers of SMEs with the necessary tools to found a culture of division and external learning. The Tunisian experience as regards consortia is far from reaching its maturity and to the extent that managers can achieve a cultural shift, it will go a long way toward the eventual achievement of knowledge transfer goals.\n", "role": "user" }, { "content": "Purpose: - The purpose of this paper is to test the transfer of good practices under the effect of various aspects of personal networks. Whereas, the majority of former work considers a direct effect of networks on knowledge transfer, the authors examined two intermediate mechanisms: the access to strategic resources and the individual's absorptive capacity.\nDesign/methodology/approach: - A quantitative study was used to collect data, via a face-to-face questionnaire with key individuals in small and medium entreprises members of consortia in Tunisia. Data were analysed with a structural equations approach, based on partial least squared-path modelling techniques.\nFindings: - Results emphasize the impact of network size, strength of ties and absorptive capacity on the strategic resource access and the impact of indirect ties, strength and range of ties on the absorptive capacity. Moreover, while absorptive capacity positively impacts good practice transfer, no support was found to the effect of resource access.\nResearch limitations/implications: - The main limitation of this study concerns the measuring of good practice transfer, since in literature there are often imprecise proxies. Also, while the authors have investigated the global impact of strategic resources, future research needs to treat them separately. Finally, a longitudinal study allows better capture of the evolution of the phenomenon studied.\nPractical implications: - Top management and directors at a consortium level need to pay careful attention to the social context within which knowledge transfer efforts occur. Resources exchanged and the absorptive capacity developed through social interactions must be designed to increase knowledge flows between firms.\nOriginality/value: - The paper links two bodies of research often studied separately in inter-organizational research: literature on social networks and that on inter-organizational learning. It is hoped that the paper contributes to a cross-fertilization of the two fields.\n", "role": "assistant" } ]
Investigating the influence of TQM, service quality and market orientation on customer satisfaction and loyalty in the Indian banking sector
[ "Customer satisfaction", "Customer loyalty", "Market orientation", "Service quality", "Banking sector", "Total quality management" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In recent decades, due to rising global competition and to achieve better business performance, organizations are focusing substantially on total quality management (TQM) practices. TQM is seen as a broad set of management and control processes designed to focus an entire organization and all of its employees on providing products or services that do the best possible job of satisfying the customer (Talha, 2004). Due to rising importance of quality consciousness, the concept of TQM is no longer restricted to manufacturing sector, rather it is widely applied to service sector as well (Ooi et al., 2011; Samat et al., 2006; Saravanan and Rao, 2006). Though there is a dearth of studies in the service sector compared to manufacturing, a few researchers have tried to bridge the gap by doing intensive research focusing on TQM in service sector (Lemak and Redd, 2000; Sit et al., 2009; Sureshchandar et al., 2002).\n2. Theoretical background and hypotheses development: TQM is considered as a systems approach to management that aims to enhance value to customers by designing and continually improving organizational processes and systems (Kartha, 2004). Though TQM was first manifested in the manufacturing sector, researchers claimed that it is equally applicable to service sector (Brah et al., 2000; Sureshchandar et al., 2002). Researchers showed that there exist strong linkages of TQM and manufacturing objective as well as organizational performance (Lin and Chang, 2006). Woon (2000) suggested that several service sectors have similar model of operations or processes to manufacturing, termed as manufacturing-oriented service, which would make them possible to adopt TQM principles and practices. But researchers suggested that the critical success factors of TQM under manufacturing industries were different from service industries (Claver-Cortes et al., 2008; Sureshchandar et al., 2002; Saravanan and Rao, 2006).\n3. Research methodology: To test the hypotheses, the researchers used SEM. Data on TQM, SQ, MO, CS and CL were collected using survey method and analyzed. Figure 1 represents the conceptual model showing relationship between the variables and hypotheses about these pathways.\n4. Results and analysis: Exploratory Factor Analysis (EFA) with principal component analysis method and varimax rotation was performed separately on TQM factors, SQ dimensions, MO, CS and CL. The results of EFA are summarized in Table I. A minimum factor loading of 0.5 was taken as the criteria to retain items (Sit et al., 2009). All the items seemed loaded properly on respective variables and retained for future analysis. The variance extracted ranged from 69.14 to 80.41 percent for all the factors. The values were above 50 percent and this showed good reliability of the measurements (Hair et al., 2009). Reliability was checked using Cronbach's a and all the values were found to be above the acceptable threshold of 0.70 (Hair et al., 2009). The Cronbach's a value ranges between 0.844 and 0.931 for TQM constructs, 0.876 and 0.905 for SQ, 0.892 for MO, 0.893 for CS and 0.878 for CL.\n5. Discussion: The present research discussed about importance of TQM in Indian banking sector. The study examined the effect of TQM on SQ, MO and the subsequent effect on CS and CL. From, the analysis, it was shown that TQM contributes significantly and positively to SQ, MO and CS and thus, leads to CL. The findings established that TQM has a strong positive association with CS. The finding is similar to Terziovski (2006), who also revealed that quality management practice is the most significant and positive predictor CS. Some important goals of TQM programs are to understand customers, satisfy them, surpass their needs, and to make sure that employees of the firm are customer focused (Tsang and Antony, 2001). By ensuring proper TQM implementation, it is inevitable and obvious that CS would definitely improve in the organization.\n6. Implications: The research and its theoretical basis not only hold crucial implication for growing literature in TQM area, but also have a lot of implications for industries. Acting as a pioneer, the study tries to link five different variables in the context of Indian banking sector. Prior researches focused on impact of TQM on CS (Kristianto et al., 2012; Mehra and Ranganathan, 2008; Sit et al., 2009), SQ (Lam et al., 2012; Ooi et al., 2011) and MO (Lam et al., 2012). They studied only two or three variables together, whereas the current research tries to integrate the five variables in the context of Indian banking sector. This research has enriched our understanding of the TQM practices that should be implemented by banks and should have a positive effect on customers' satisfaction and later leads to achieve CL. The research is the first of its kind that is devoted to understand effect of TQM on SQ, MO and CL through CS in an emerging or developing country business environments. From theoretical viewpoint, this study has undoubtedly contributed to the existing literature by providing a better understanding on the proposed structural relationships between TQM, SQ, MO, CS and CL, where the structural relationship between constructs can be determined using robust SEM, rather than regression analysis. The research confirmed that TQM is a holistic management philosophy and can be applicable across different kinds of organizations, both manufacturing and service.\n7. Limitations and directions for future research: No matter how conceived a research effort is, there exist limitations. This study also has a number of limitations that provides future research opportunities. First, the present research was confined to banking sector only, which limited the generalizability of the study. Future research could reconsider this trade-off and research on more number of service sectors. The result can also be compared and differences can be seen. Second, the research was conducted in a single country India. This again restricted to generalize the findings. A cross-cultural study can be conducted and the causal links can be analyzed.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to examine the effect of total quality management (TQM) on service quality (SQ), market orientation (MO) and the subsequent effect on customer satisfaction (CS) and customer loyalty (CL). The research also analyses the direct effect of TQM on CS.\nDesign/methodology/approach: Data were collected from middle managers of retail banking sector in India using survey method. Structural equation modeling is used to analyze the data and to test the hypotheses.\nFindings: The results show that TQM is positively associated with SQ, MO and CS. SQ and MO are significantly and positively associated with CS. CS subsequently leads to improved CL.\nPractical implications: Enforcing TQM practices in their organizations as a change management tool, banks can achieve a greater degree of MO, improved SQ and higher CS to gain higher CL.\nOriginality/value: This paper empirically demonstrates that, by proper implementation of TQM in banks, the SQ and MO can be improved. Well implemented quality management practices, better customer service and market-focused attitude of banks will elevate the CS level and improve loyalty of customers toward the banks. With a little study has yet focused on India, the paper offers knowledge to banking professionals for increasing CL by effective implementation of TQM practices.\n", "role": "assistant" } ]
Library cooperation in Turkey: the results of a survey of ILL librarians in higher education
[ "Survey", "Higher education", "Turkey", "Resource sharing", "Document supply", "ILL" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The concept of \"cooperation\" can be defined as individuals acting together by uniting their strengths for a common purpose. We see that mankind has realized cooperation as an obligation of communal living since the beginning of history. Cooperation has enabled individuals and institutions to accomplish their objectives more effectively.\nThe document supply and interlibrary loan systems in Turkey: When we look at the history of sharing resources in Turkish libraries, we see that the first studies were conducted for public libraries in 1961 (Alkis and Yilmaz, 2008). Regulations governing the ILL of printed books were published in the official gazette in 1981. The legal regulation and standardization of all libraries was considered in 1988 but remained as a draft (GDLP, 1988, p. 25). In 2009, The General Directorate of Libraries and Publications prepared a cooperation protocol between Public and University libraries which was sent to all university rectors. However, this study remained as a project.\nThe survey conducted with ILL academic librarians: Purpose of the study\nQuestions about libraries and ILL librarians: The first seven questions concerned their job description, vocational training and their institutions. In all, 67 per cent of those responding worked at a state university, while 33 per cent were at a foundation university. In all, 48 per cent worked in small libraries with collections of up to 10,000 books, 21 per cent were at medium-sized libraries with 50,000-100,000 books, 19 per cent with 100,000-500,000 and 8 per cent with over 500,000. Of all, 89 per cent have at least an undergraduate or post graduate degree in librarianship, so it is pleasing that a high number of staff is trained. 72 per cent of participants were solely responsible for all ILL activities in their institution. One person may be sufficient for the institutions where the number of people working in the library is 1-10 (55 per cent); however, it will be insufficient for the institutions with 30 or above (11 per cent) library staff. That one person manages the ILL service in large institutions with a high number of staff might negatively affect the quality of service. The percentage of people answering \"yes\" to \"Is ILL service your only area of responsibility in the library?\" is quite low (4 per cent); 96 per cent of the participants work at different services besides ILL service. These percentages suggest that large-scale libraries are understaffed with ILL librarians.\nQuestions about the ILL service: The majority of the participants received ILL requests from users via e-mail (48 per cent). The percentage of the participants receiving requests through an integrated ILL module is 27 per cent. Only 10 per cent receive printed ILL requests. The majority of requests (66 per cent) are met within a good time of 1-3 days (Figure 1). Most institutions charge for ILL requests made from both their own users and other universities although a minority does not charge anything to the user (17 per cent). The criteria in choosing a supplier are determined first by the collection content, then by the speed of meeting the request, the proximity and, finally, the payment policy (Figure 2).\nQuestions about domestic ILL requests: The institutions participating in the survey mostly use KITS for domestic ILL requests (54 per cent). KITS is preferable because of its simple user interface; it is free of charge and has a high number of members (172). The second most used system is TUBESS at 35 per cent (Figure 3). A clear majority of institutions share e-articles through ILL (67 per cent) but a third do not. Users are mostly not charged for e-articles (91 per cent), and the speed of meeting domestic requests is quite high with just over half within 1-3 days (Table I). Articles are mostly sent to users through the file transfer system of KITS, but 25 per cent are via e-mail. Sending through e-mail is the most preferred way to send the article to the user (42 per cent), but 30 per cent prefer to give users the printed article and 26 per cent deliver in accordance with the license and copyright constraints. Thus, while ILL librarians choose the fastest methods to meet the needs of their users, they also pay attention to license and copyright issues (Figure 4). The vast majority of libraries receive and send less than 500 ILL requests annually (Table II).\nQuestions about ILL requests from abroad: A clear majority of the participants in the survey do not provide their users with publications from abroad (73 per cent), the reasons being expense, discouraging requests from patrons, lack of staff knowledge and commitment. Those that provide this service do so through OCLC World Share, IFLA Vouchers and the British Library (Figure 5). Conversely, the majority (72 per cent) of respondents do not send publications abroad (Table III). Those that do send via e-mail (16 per cent) and 5 per cent use OCLC's Article Exchange. This finding shows us that article exchange systems which are quite user-friendly and secure are not commonly used in Turkey compared to e mail. While 10 per cent of the institutions sending publications abroad through ILL do not charge a fee 12 per cent of them do charge a standard fee for shipment, photocopying, etc. The time taken to receive requested items from abroad is shown in Table V. However, the number of publications sent and provided from abroad is fairly low (Table IV).\nQuestions about the new trends in ILL services: E-resources\nFurther comments: We asked the participants to state their further comments on electronic document supply and the ILL service as practiced within the framework of traditional document supply in the \"Further Comments\" part of our survey. A summary appears below:\nRecommendations and conclusion: In today's world where information and information resources are increasing rapidly, it does not seem possible for libraries to buy whatever publication they want especially with their limited budget. Though ILL has been declining in some countries, in the case of Turkey, it is becoming increasingly important to use the service effectively, as libraries need to manage their resources carefully while supplying their users efficiently and effectively.\n", "role": "user" }, { "content": "Purpose: The purpose of this study is to investigate library cooperation among the libraries of Turkish higher education institutions in the changing global environment.\nDesign/methodology/approach: This study has been carried out with ILL librarians working in state and foundation university libraries in Turkey. The survey was sent to 140 inter library loan (ILL) librarians who were asked to answer 35 questions both open and closed. There were 100 responses which is an excellent response rate of 72 per cent. It was conducted with VETI (Data Collection and Statistics), which is a web-based form application developed by Istanbul Technical University.\nFindings: Turkish libraries are taking the opportunities offered by technical developments to serve their users with instant information by providing access to electronic resources and transferring their collections to electronic media. Reconstruction of ILL services to fulfill users' needs is inevitable because of the increasing expectations of users and emerging alternatives to access information.\nOriginality/value: This is the first systematic national investigation into the state of ILL in Turkish libraries.\n", "role": "assistant" } ]
The role of the university in accelerated learning and innovation as a regional ecosystem integrator
[ "Social entrepreneurship", "Innovation", "Accelerated learning" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper presents an adaptation of a program that is at the intersection of two dynamic force fields. The first relates to imperatives impinging upon and inherent in higher education. The second ties to the concept of ecosystems as spaces for aligning actors and resources to create value. As noted by educational thought leaders, the confluence of such forces may be an indication of an inflection point that signals disruption to long-standing constraints as to how we think about education as well as offering extraordinary strategic opportunities (Brandenburger, 2013).\nThe higher education landscape: As noted in a 2014 white paper from the American Council on Education, the landscape for higher education is in turmoil. Changing demographics, disruptive competition, eroding trust, reduced public financial support, and questions surrounding costs and productivity are a sampling of issues that complicate the context of higher education. Further, the one area where higher education institutions have traditionally grounded their value propositions - academic credentialing - is being challenged (American Council on Education, 2014).\nThe ecosystems metaphor: Coined by a botanist in the first half of the twentieth century, the term ecosystem referred to a localized community of living organisms interacting with each other and their environment. Noticing relevant parallels between the worlds of biology and commerce, Moore (1993) reported the concept to the increasingly interconnected and rapidly changing context of business. Moore noted the trend of successful business evolution involving attracting resources, creating cooperative networks, and co-evolving capabilities around innovation.\nThe context for the program: The context for the program is a relatively young university founded in 1965 to address the educational needs of an underserved region of the Midwest. The university's rapid growth attests to its value to the region in that it is now a comprehensive public university enrolling approximately 10,000 undergraduate students in 70 majors. Recognized by the Carnegie Commission as an \"Engaged University\" its vision is simple but powerful - \"Shaping the future through learning and innovation.\" Further evidence that innovation is in the DNA of the university is the vision for its college of business which emphasizes an entrepreneurial mindset including innovative thinking and openness to new ideas. The college has a defined initiative to foster entrepreneurial education, engagement, and outcomes in the region.\nThe regional development program: Originally funded by a grant from the Lilly Endowment, the program's purpose is nurturing citizenship to enhance the quality of life and to boost the retention of intellectual capital in the region. Through the program, participants expand their opportunities for creating new connections with other residents, with local and regional organizations, between communities, and between the university and the region.\nThinking innovatively and participants' pre-test-post-test changes related to the ideation process: Innovation in organizations is a competency capable of generating competitive advantage (Tidd et al., 2005; Barsh et al., 2008). As a result, understanding innovation processes has been identified as an imperative by business and educational realms (AACSB, 2010; The Chronicle of Higher Education, 2013). More specifically, it is crucial that we understand the situational processes that can engender innovative ideas to solve complex issues (Davis, 2000; Isaksen et al., 2009).\nResults: Tables I-III present pre-test and post-test means and p-values for the paired sample t-tests for the measures. Note that, as expected, post-test means are consistently significantly higher (or lower depending on item wording) for a shift in beliefs away from self-censoring and prejudging ideas during ideation and more toward greater openness in the ideation process. Further, perceptions relating to ideation self-efficacy and norms were also significantly strengthened.\nConclusion: As higher education stands at this inflection point characterized by blurring service boundaries and competitive disruption, the need for institutions to strategically create distinctiveness that leads to competitive advantage has never been greater. Historically, education is a service where \"customers\" do a good deal of the work to produce short- and longer-term outcomes. As a result, resources have been directed at understanding reciprocal communication and interaction as a means of engaging students. Yet like businesses, universities recognize that they operate in a broader ecosystem beyond the institution-customer/student intersection and can thus uniquely contribute and capture value from their broader systems. Thus, in order to create genuine value in educational service delivery of the future, there is a need for more highly developed understanding of the institutional-ecosystem intersection. The present paper aims to contribute to the higher education literature by prototyping a program which points to a broader conception of the role of the university in the ecosystem development.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to present an adaptation of a program that is at the intersection of two dynamic force fields. The first relates to imperatives impinging upon and inherent in higher education. The second ties to the concept of ecosystems as spaces for aligning actors and resources to create value.\nDesign/methodology/approach: Tables I-III present pre-test and post-test means and p-values for the paired sample t-tests for the measures.\nFindings: As expected, post-test means are consistently significantly higher (or lower depending on item wording) for a shift in beliefs away from self-censoring and prejudging ideas during ideation and more toward greater openness in the ideation process.\nOriginality/value: The paper examines the outcome of an educational program.\n", "role": "assistant" } ]
The antecedents of relational capital in key exporter-importer relationships: An institutional perspective
[ "National cultures", "Croatia", "Relational capital", "Institutional distance", "Export", "Buyer-seller relationships", "Relational bonding norms", "Reliance" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Globalization tendencies significantly influence a firm's desire to explore new and distant markets. By going international, small and medium-sized enterprises (SMEs) strive to achieve further growth, especially when the domestic market is saturated and hypercompetitive; however, international venturing endeavors are riskier than domestic ones (OECD, 2009).\nConceptual development: This study's framework builds on the theoretical premises of relational exchange and institutional theories. According to the relational exchange paradigm, competitive advantage results from the specific resources embedded within a dyadic relationship. Based on the ideas that originate from the theoretical legacy of social psychology, relational exchange activities aim to establish, develop, and maintain successful business relationships (Morgan and Hunt, 1994). The Industrial Marketing and Purchasing Group identified relationships and networks as a pivotal governance platform for interorganizational exchange (Hakansson and Snehota, 1995). With the goal of contextualizing these theoretical ideas, international marketing scholars have shown rising interest in the relational view (Katsikeas et al., 2009; Leonidou, 2003; Leonidou et al., 2006, 2014; Styles et al., 2008), suggesting that relationalism is the optimal governance mechanism in cross-border relationships (Ju et al., 2014; Zhang et al., 2003).\nResearch hypotheses: Importer's reliance and relational capital\nMethod: Sample description and research setting\nFindings: Analysis of the measurement model\nDiscussion and implications: Based on the tenets of relational exchange and institutional theories, this study offers novel insights into the field of managing E-I relationships. Previous international marketing studies have neglected the role of relational capital in developing key E-I relationships. Hence, our study responds to the recent call of international marketing researchers to pay more attention to the role of distance in the development and management of cross-border business relationships (Katsikeas et al., 2009). We address these gaps by empirically testing the theoretical framework of relational capital development under the influence of formal and informal institutional distance.\n", "role": "user" }, { "content": "Purpose: During their process of going international, small and medium-sized manufacturing firms seek to establish long-term relationships with key importers in order to minimize the risks of doing business in a foreign market. In the process of establishing long-term relationships, exporters aim to create relational capital with key importers. Yet, the body of international marketing literature that addresses the importance of relational capital in exporter-importer (E-I) relationships is still underdeveloped. The purpose of this paper is to examine the influence of relational norms on relational capital in key E-I relationships under the moderating influence of formal and informal institutional distance. The study's conceptual framework was developed by integrating relational exchange and institutional theories.\nDesign/methodology/approach: The study was carried out by using a survey methodology. Data were obtained by questionnaire from a sample of 122 small and medium-sized exporters from the manufacturing industry in Croatia. In order to test the hypotheses, the ordinary least squares technique was employed.\nFindings: The findings support the hypotheses, implying that the development of relational capital requires relational efforts in terms of reliance and relational bonding norms. Additionally, the empirical data suggest that the dimensions of formal and informal institutional distance significantly moderate the relationships between relational norms and relational capital.\nOriginality/value: The value-added of this study is embedded within the theoretical framing and empirical testing of the antecedents of relational capital in key E-I relationships in the context of the institutional distance between partners, which has been neglected by previous studies in the field.\n", "role": "assistant" } ]
Evaluation of MGNREGA: data envelopment analysis approach
[ "Governance", "Efficiency", "DEA", "Welfare policy", "Policy evaluation", "MGNREGA", "Efficiency benchmark" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: India, battling the twin problems of poverty and unemployment since the turn of the twentieth century, has relied heavily on policy directives as a course-correction action. Successive governments, faced with widening chasms of inequality, have undertaken welfare schemes to bridge the gaps. The focus of this paper is evaluation of the rural flagship scheme of the Government of India, the Mahatma Gandhi National Rural Employment Guarantee Act-2005 (MGNREGA). MGNREGA sharply changed the way government reached out to the needy. Aiming to give dignity to earn a livelihood to the beneficiary, it turned entitlements into work. The scheme made it mandatory on the government to provide upon demand, 100 days of guaranteed unskilled employment in a year at INR100 ($1.50) a day, to any one adult member of a rural household registered for the scheme. Along with providing a rural safety net, it also tried to usher in social, infrastructural and environmental change in rural India. The central government provides dedicated resources and the implementation rests with the states. Tremendous hope goes with MGNREGA, \"structured as a social safety net [with] the potential to transform rural India into a more productive and gender equitable society and become the driver of environmental regeneration\" (Kamath, 2010). Rural economies face challenges far removed from those faced by urban centers and thus require more robust and far sighted implementation and evaluation methods.\n2. Earlier studies regarding MGNREGA implementation challenges: Extending social protection on equity, empowerment, economic, social and cultural rights as a transformative social policy (Devereux and Sabates-Wheeler, 2004), MGNREGA has received critical scrutiny for the promises it made and the extent to which they have and have not been realized. Being a rural scheme, it faces challenges because both implementers and beneficiaries confront political and power obstacles (Raabe et al., 2010). The complex mechanics of the scheme involve the central, state and three tiers of local government. The three tier rural local administrative divisions of India are a nested hierarchy of sub-division of districts, blocks and villages. The village (gram) panchayat is an elected body at the lowest democratic level and controls five or six villages.\n3. Research gap: A thorough scanning of the literature points to the extensive work done on impact of the MGNREGA. However, there are very few studies on efficiency evaluation of its implementation. The two efficiency studies enumerated above (Datta and Singh, 2012, 2014), limited their work to state and district level evaluation. Thus the research gap that this paper hopes to fill is to evaluate MGNREGA implementing states on the basis of efficiency. In doing so, this paper is one of the first to provide efficiency evaluation of MGNREGA at the national level.\n4. Methodology for measuring efficiency: This paper has attempted to operationalize policy evaluation by encompassing policy objectives, policy implementation process and policy outcomes. To do this, it has borrowed from two sources first, Thomson and Perry's (2006) \"black box\" of \"collaborative processes\" and second Charnes et al. (1978), non-parametric \"data envelopment analysis.\"\n5. Measuring MGNREGA implementation efficiency: The paper now presents an efficiency framework to examine the input output ratio between factors used and output created in implementing the MGNREGA at the state level. The framework called the MGNREGA Implementation Efficiency Model (MIEM) (Figure 1) will help to quantify implementation quality and efficiency by measuring the output created for the inputs used. It also terms the implementation process at the state level as the black-box, of which not much is known in advance and perhaps leads to differential results in the output created by MGNREGA implementing states. The MIEM provides existing implementation quality and delivers efficiency measurement for benchmarking and categorizing the states. The framework will attempt to offer suggestions for better outcomes.\n6. Results, analysis and implications: Detailed DEA input and output numbers for Assam are presented in Table I as an illustration.\n7. Conclusions: The MGNREGA has been the biggest rural employment initiative of the Indian government and has generated 19,010.1 million person days at an implementation expenditure of $40,219.74 million (MGNREGA Operational Guidelines, 2014). Our analytical framework has benchmarked and categorized the states and presented factors contributing to implementation variation. This study argues that using the salient features of the MGNREGA as variables to evaluate efficiency will improve analysis and comparison, and prevent discrimination against any state with a downward bias in certain parameters.\n", "role": "user" }, { "content": "Purpose: How can efficiency of a welfare scheme be measured? The purpose of this paper is to develop an efficiency evaluation model, the Mahatma Gandhi National Rural Employment Guarantee Act (MGNREGA) implementation efficiency model (MIEM), to evaluate the rural employment guarantee scheme in India.\nDesign/methodology/approach: MIEM employs data envelopment analysis (DEA) to compare relative efficiency of MGNREGA implementing states. It uses policy implementation process as a central \"black-box\" about which not much can be said, to account for state-wise implementation differences.\nFindings: Based on administration, funds, expenditure, employment created, works executed and completed, women beneficiaries and households completing 100 days of employment, the MIEM captures current implementation efficiency and provides suggestions to propel inefficient states toward efficiency.\nPractical implications: DEA has operationalized MGNREGA evaluation. As a decision support system, MIEM assists evaluators to develop guidelines from better performing states. It is anticipated that it will facilitate scaling up MGNREGA in inefficient states.\nSocial implications: The model developed here can be applied to diverse evaluation conditions thus leading to better utilization of scarce resources.\nOriginality/value: This paper is one of few to use DEA to evaluate MGNREGA, and is one of the first to evaluate all India implementing states on efficiency.\n", "role": "assistant" } ]
Value relevance of voluntary disclosure: evidence from Turkish firms
[ "Turkey", "Disclosure", "Public companies", "Voluntary disclosure", "Capitalization", "Firm value", "Market capitalization", "Value relevance" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Voluntary disclosure in the annual reports and in other information media has been one of the rapidly growing research areas in the accounting field. In this, several factors have played important roles. Among them are development of communication tools, stakeholders' need for more transparency and accountability, and corporate governance practices. Earlier studies have evaluated the extent of voluntary disclosure made by corporations in many countries, and provided evidences of firm characteristics which have impact on voluntary disclosure level. However, there is one issue which has not been addressed sufficiently in previous studies; namely, value relevance of voluntary disclosure, particularly in the emerging capital markets. Emerging capital markets break from those of developed countries in that they have high-growth potential, relatively weak regulatory environment, weak corporate governance leading to expropriation of minority shareholders, and low information disclosure level causing high information gap between managers and investors. Thus, research studies about these markets are needed and are vital in improving the weak transparency and disclosure situation by attracting the attention of regulatory bodies and firm managers.Although there are two institutional mechanisms, namely accounting standards and auditors, that play a role in mitigating the information gap between managers and owners, they are far from being perfect (Healy and Palepu, 1993). Hence, voluntary disclosures can help decrease the information gap, improve the credibility of financial reporting (Healy and Palepu, 1993), and contribute to understanding the role of accounting information in firm valuation and corporate finance (Core, 2001). Earlier empirical studies have documented that more voluntary disclosure increases firm value and decreases cost of finance (Orens et al., 2009).Previously, some studies have focussed on the association between disclosure and firm value. However, as pointed out by Al-Akra and Ali (2010), there is little empirical evidence to support that association. In a recently published study, Hassan et al. (2009, p. 80) touched upon this issue and stated that \"there is little direct empirical evidence with regard to the relationship between voluntary disclosure and firm value in general and for emerging markets in particular.\" Thus, we aim at contributing to the literature by conducting an empirical study in the emerging capital market of Turkey.The remainder of the paper is structured as follows. In Section 2, literature review is provided. Section 3 develops the hypothesis. In Section 4, the research method is explained. Section 5 analyzes the results. Finally, Section 6 concludes the paper, explains limitations, and makes recommendations for further research.\n2. Literature review: Emerging markets have become the focus of international corporations, personal and institutional investors due to their high rates of economic growth (Millar et al., 2005). However, they suffer from low investor protection practices, especially expropriation of minority shareholders both by managers and controlling shareholders (Gonenc and Aybar, 2006). They have higher information asymmetry between managers and investors (Gul and Leung, 2004; Chau and Gray, 2010), and have lower level of disclosure than those in developed market economies (Salter, 1998; Tower et al., 2011; Wang et al., 2008), whereas high-growth firms need more voluntary disclosure than low-growth firms due to their need for external finance (Core, 2001). These two factors (i.e. low minority investor protection and low disclosure level) might be detrimental by keeping foreign investors away from the marketplace, and set an obstacle to the international capital flow toward these economies. Mitton (2002) states that disclosure quality and legal protection of minority shareholders are two key elements of corporate governance. Thus, firms have to respond positively to increasing demand for higher transparency and disclosure by investors in face of severe competition for international capital (Kohl and Schaefers, 2012). Young and Guenther (2003) found that disclosure of value-relevant accounting information has a positive effect on international capital mobility; therefore, they state that reducing information barriers may improve international capital mobility.Ugur and Ararat (2006) state that corporate governance quality and macroeconomic environment can affect each other inversely in emerging markets; and this type of reverse effect has been at play in Turkey in the 1990s and 2000s. They assume that the macroeconomic instability of the 1990s was associated with low corporate governance quality, whereas gaining macroeconomic stability after 2001 has contributed to improvement in the country's corporate governance regime. In order to ensure investor protection and to promote transparency, Corporate Governance Principles of Turkey was issued by The Capital Markets Board (CMB) of Turkey in June 2003 for the first time and amended in February 2005 (CMB, 2005). Recently, the CMB issued the Communique serial IV no. 56 dated December 30, 2011 on \"the determination and implementation of the principles of corporate governance\" (CMB, 2011).Similar to other emerging countries, ownership structure of Turkish companies is highly concentrated, families being the dominant shareholders (Demirag and Serter, 2003; Yurtoglu, 2000). The majority of publicly traded firms have only a small portion of publicly owned equity which causes information asymmetries between investors and the managers (Ugurlu, 2000). Therefore, Ararat and Yurtoglu (2006) point out the importance of transparency and good-quality disclosure of relevant information for Turkey where the biggest agency problem results from information asymmetry and expropriation by majority shareholders.Several earlier studies have pointed out that voluntary disclosure impacts firm value based on signalling theory. Comprehensive disclosure signals better governance mechanisms and fewer agency conflicts, thus leading to higher firm value (Sheu et al., 2010). Curado et al. (2011) point out that intellectual capital disclosure impacts firm value as perceived by markets. Gordon et al. (2010) also state that voluntary disclosures in the annual report send signals to the marketplace, and these signals are expected to increase a firm's net present value and, in turn, its stock market value. Bruggen et al. (2009) emphasize the importance of intellectual capital disclosure for investors' accurate assessment of financial statements and, thereby firm value. Drawing on the previous literature, Cormier et al. (2011) state that both social disclosure and environmental disclosure convey value-relevant information to stock markets. Gallego-Alvarez et al. (2010) provided evidence that social and environmental reporting practices have a positive effect on shareholder value creation. As there are incentives for disclosing more information, there are at the same time some disincentives as pointed out below (Beattie and Smith, 2010, p. 278):* avoiding giving away information which may harm one's competitive position;* reducing the costs of collecting and disseminating information;* avoiding attracting unwanted scrutiny by other stakeholders; and* avoiding attracting unwanted scrutiny by regulators.In this context, Abdolmohammadi (2005) states that although there is an incentive for firms to provide voluntary disclosure about their intellectual capital, he recommends firms to compare the benefits of information disclosure with the cost of accumulating and providing such information.The number of studies which investigated the impact of voluntary disclosure on market capitalization or firm value and market return is limited. Hassan et al. (2009) assert that more public information enhances firm value, drawing on empirical findings of previous studies. Murray et al. (2006) conducted a study in the UK's largest companies and found no direct relationship between share returns and social and environmental disclosures. In a Malaysian study, Anam et al. (2011) proved that the extent of intellectual capital disclosure has a significant positive effect on market capitalization. In addition, they found that there is significant positive impact of the control variables (i.e. book value, net profit, firm size, and leverage) on market capitalization. The results of Abdolmohammadi's (2005) study on a sample of Fortune 500 companies showed that there is a significant effect of intellectual capital disclosure on market capitalization. However, he found that neither the book value of the firm nor the difference between its return on asset and its industry average has significant explanatory power over market capitalization. Orens et al. (2009) analyzed the corporate web sites of firms in four continental European countries (Belgium, France, Germany, and the Netherlands), and investigated whether or not the level of intellectual capital disclosure on web sites has an association with firm value. They observed that firm value is positively associated with intellectual capital disclosure level. Furthermore, they found that profitability and analyst following have also positive association with firm value, but firm size has negative association, and leverage and ownership have no association. Al-Akra and Ali (2010) indicated the existence of association between voluntary disclosure and firm value using panel data for Jordanian listed companies over a nine-year period. Gordon et al. (2010) conducted a study in the USA about value relevance of information security disclosures. Their findings provided strong evidence that voluntarily disclosed information security items are associated positively with the market value of a firm.\n3. Development of the hypothesis: Earlier studies generally utilize signalling theory to explain why firms disclose voluntary information to stakeholders (Anam et al., 2011; Sheu et al., 2010; Gordon et al., 2010). Signalling theory purports that an organization will attempt to signal good news to investors and other stakeholders through voluntary disclosure (Oliveira et al., 2006). According to Francis and Schipper (1999), value relevance is measured by the ability of financial statement information to capture or summarize information which impacts share values. Previous studies reveal that increasing disclosure level results in decreasing misevaluation of share price, thus increasing firm's market capitalization (Anam et al., 2011). Anam et al. (2011) base their hypothesis about the association between intellectual capital disclosure and market value on resource-based theory and signalling theory. According to the authors, if firms have valuable resources and create value, they will let stakeholders know this by disclosing information. Furthermore, managers will also be willing to signal value creation process to the stakeholders. Thus, the following hypothesis is developed:H1. The level of voluntary disclosure of Turkish listed firms is positively associated with the value of these firms.\n4. Research method: 4.1 Data\n5. Results and analyses: 5.1 Descriptive statistics\n6. Conclusion: Although previous papers have investigated disclosure levels (either mandatory or voluntary) of the firms and the determinants of disclosure level, there is not a sizable body of research, especially in the emerging markets, which examine the impact of disclosure on firm value. Thus, there is a need for elaboration on the value that voluntary information has in these markets.This paper empirically investigates the association between voluntary disclosure level and firm value for Turkish listed manufacturing firms. The results of univariate analysis have indicated that there is a strong correlation between disclosure level and firm value as measured by market capitalization. This finding is further supported by multivariate analyses which confirm the significant positive relationship between disclosure level and firm value (i.e. market capitalization) at the end of the year. Later on, this association was tested and proven for the future firm value six months after the year-end. However, when we utilized the ratio of market value to book value of equity as the dependent variable, we could not obtain a significant association between disclosure and firm value. Thus, the significance of association between disclosure level and firm value varied depending on the model we used. In addition, we found that profitability and firm size has significant positive impact on firm value in all four models. On the other hand, the third and fourth models revealed that leverage has significant but negative impact on firm value.The results are particularly important for emerging markets since they need capital to finance high growth. In the face of intense competition for international capital, they have to make legal reforms to attract foreign investors. One of the obstacles faced by international capital flow is the insufficient transparency of firms. Thus, voluntary disclosure is an important means of overcoming this obstacle by which information asymmetry is reduced between managers and investors, and domestic and foreign investors. Moreover, the greater availability of corporate information can improve capital market efficiency and attract investors (Wang et al., 2008; Gul and Leung, 2004).The findings of the paper provide important implications for investors, managers, and regulatory bodies. The results tell investors that voluntary disclosures, in addition to mandatory ones, have positive effect on market value of companies. This might mean that managers who perform well let public know what they do, and how they manage their firms. Eventually, this affects firm value. Thus, it can be said that market participants value information disclosure, or that disclosure is value-relevant. There are managerial implications of the results as well. Managers who engage in good practices of information disclosure are recommended to continue doing so. For those who refrain themselves from providing information to the public, we can say that the results call for more transparency if they want their firms' shares to be more attractive. Regulatory bodies are the third important party for whom results have got some implications. They are expected to guide firms toward the best practices of disclosures since firms look for such guidance. They play a motivating role in this new era of information disclosure. Actually, Turkish regulatory bodies have indicated their support of transparency and better disclosure by enacting some laws and regulations regarding disclosure practices.There are some limitations of the study as almost every study has. The sample includes only Turkish listed manufacturing firms. Therefore, the findings might not be valid for other industries such as service or merchandising firms. Annual reports have been used as the sole source of data gathering, others such as web sites and press releases could be used in future studies. The data were gathered just for the year 2010, thus, the study may be replicated by using panel data covering multiple years to observe the trend in investigations. Finally, we have not investigated the possible endogenous relationship of firm value and disclosure level. While disclosure level has impact on firm value, it is possible that firm value may also influence disclosure level. Some earlier studies, which investigated the impact of firm characteristics on voluntary disclosure level, also found that firm size as measured by market value has significant impact on disclosure level (Aksu and Kosedag, 2006; Wang et al., 2008). Despite these limitations, it is hoped that the study inspires more studies in this field of research, and that it provides support in favor of promoting voluntary disclosure to present and potential investors, companies, and regulators.\n", "role": "user" }, { "content": "Purpose: - The purpose of this paper is to examine whether or not listed Turkish companies' voluntary disclosure practices are value-relevant in the capital market.\nDesign/methodology/approach: - The sample consists of 129 manufacturing companies listed in the Istanbul Stock Exchange (ISE) for the year 2010. Data regarding voluntary disclosure level and other variables have been collected by analyzing the contents of annual reports. Through multiple regression analysis, the authors investigated whether or not voluntary disclosure level impacts firm value.\nFindings: - The main finding of the study is that voluntary disclosure is value-relevant; i.e. impacts firm value. This implies that market participants value voluntary disclosure. The more information firms disclose voluntarily, the higher value they have in the eyes of investors. Therefore, this finding might be accepted as a signal to corporations to disclose more information to the stakeholders. However, the finding varied based on the dependent variable used; hence, the result was not supported by all models.\nResearch limitations/implications: - The findings of this paper are based on the study conducted on the ISE including manufacturing industry. Thus, the results might not be valid for non-listed and non-manufacturing companies.\nOriginality/value: - There is a scarcity of studies regarding how the marketplace perceives voluntary disclosure. The paper evaluates comprehensively the value relevance of overall voluntary disclosure as well as of 12 sub-categories of disclosure. It contributes to the relevant literature, particularly in an emerging market context.\n", "role": "assistant" } ]
A meta-synthesis approach to specify components of future banking
[ "Best practices", "Meta-synthesis", "Future banking", "Mega-trends" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Banking regarded as a canonical industry in every economy's financial system. Economists consider banks as mediators, which reduce transaction costs between depositors and borrowers (Mishkin, 2004). Bankers see modern banks different from the traditional entities and along with customers, call for more deregulation. They want to be more electronic, more connected, more authorized to provide more services.\n2. Modern to future banking: Banking as a leading activity in the financial system has had a definite function and definition. While a highly regulated financial system in every country is supposed to get funds of surplus spending units to deficit spending units (Mishkin, 2004) and make investments less expensive with decreasing transaction costs, financial mediators traditionally facilitate the above function.\n3. Methodology: Facing with diverse, extensive resources in the first part of reviewing future trends and best practices to have a general perspective of future banking, I found out that the best method to capture this diversity of resources and dimension is the meta-synthesis method.\n4. Concluding remarks: Banking regarded as a canonical industry in every economy's financial system. Mainstream economics considers banks as mediators, which reduce transaction costs between depositors and borrowers (Anyanwu et al., 2017; Mishkin, 2004).\n", "role": "user" }, { "content": "Purpose: This paper aims to depict an extensive and cohesive picture of future banking's dimensions and components.\nDesign/methodology/approach: A two-step qualitative approach has been applied. First, an extensive scanning has been carried out to identify megatrends and best practices. Second, applying meta-synthesis analysis, more than 186 up-to-date references were strategically scanned to elicit dimensions and components of future banking.\nFindings: This research has had twofold findings. The direct and explicit results were the main dimensions of banking in the future: information technology, employees, customers, diversified services, organizational structures and farsightedness. The implicit findings were also remarkable: many entities are thinking about future of banking, mostly in financial technology dimension; the departure from traditional banking has recently been accelerated; and more works need to be done to have a comprehensive map of banking in the future.\nResearch limitations/implications: As the research methodology was based upon a literature review, it lacks covering some hidden or less flashing dimensions such as future business models, merging between banks and other financial or technological firms in advance, the evolution of organizational structures, etc., which would be captured by applying other methods such as expert Delphi panels.\nPractical implications: Planners in the banking industry can benefit from the direct findings. They may extend the results, customize and prioritize the components to provide a competitive business model in the future market of banking.\nOriginality/value: The novelty of this paper lies in a cohesive representation of future banking dimensions and components, which is created by a systematic methodology and broad literature review.\n", "role": "assistant" } ]
Sensory evaluation of commercial ready-to-eat rice between trained panelist and consumer
[ "Consumer acceptability", "Descriptive analysis", "Brown rice", "Ready-to-eat cooked rice", "White rice" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Rice (Oryza sativa L.) is considered to be the most important crop and staple food in Korea. However, rice consumption per capita steadily decreased from 135.6 kg in 1979 to 72.8 kg and then decreased to 62.9 kg in 2015 (Korea National Statistical Office, 2015). White rice among different rice was identified as the most frequently consumed food per person per day (151.7 g) among the Korean population in 2015 (Ministry of Health and Welfare, 2015).\n2. Methods: 2.1 Sample preparation\n3. Results and discussion: 3.1 QDA\n4. Conclusion: This study was conducted to examine the descriptive sensory characteristics of commercial ready-to-eat cooked rice in the market; examine CA of ready-to-eat cooked rice; and investigate the relationship between descriptive characteristics and CA of ready-to-eat cooked rice. Descriptive sensory characteristics and CA of eight commercial ready-to-eat cooked rice samples were examined by 8 trained panelists and 50 consumers, respectively. A total of 24 descriptive attributes for appearance, odor/aroma, taste/flavor, and texture were developed. Three groups of rice samples were obtained from PCA. The overall liking score for the WR_C sample was highest (6.43) among the eight samples. The WR_C, WR_N, and WR_O samples characterized by intactness, starch odor, translucency, whiteness, and glossiness were located on to the positive PLS 1, whereas the WR_D and WR_E white rice samples characterized by scorched odor, cohesiveness, stickiness, and moistness were located on the negative side of PLS 2 in the PLSR analysis. Further studies on the improvement of sensory quality for brown rice are necessary to increase CA in terms of health functionality of brown rice.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to examine descriptive sensory characteristics and consumer acceptability of eight commercial ready-to-eat cooked rice samples by 8 trained panelists and 50 consumers.\nDesign/methodology/approach: A total of 24 descriptive attributes for appearance, odor/aroma, taste/flavor, and texture were developed. Also Consumer Acceptability (CA) was performed for overall liking, appearance, flavor, and texture liking. All statistical analyses were using analysis of variance, principal component analysis (PCA), hierarchical cluster analysis (HCA), and partial least square regression (PLSR).\nFindings: The overall liking score for the cooked white rice from C brand was the highest (6.43) among the eight samples. Three groups of eight commercial ready-to-eat cooked rice samples were obtained from PCA and HCA. The samples of cooked white rice from C, N, and O brand characterized by intactness, starch odor, translucency, whiteness, and glossiness were located on to the positive PLS 1, whereas the samples of cooked white rice from D and E brand characterized by scorched odor, cohesiveness, stickiness, and moistness were located on the negative side of PLS 2 in the PLSR analysis.\nOriginality/value: Further studies on the improvement of sensory quality for brown rice are necessary to increase CA in terms of health functionality of brown rice.\n", "role": "assistant" } ]
A cost-benefit case for RFID implementation in hospitals: adapting to industry reform
[ "Technology", "Operations management", "Facilities management services", "Tracking", "Healthcare", "Cost efficiency" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The healthcare sector in the USA is currently undergoing rapid growth and transition. Changes in reimbursement and incentives to improve quality of care by the Centers for Medicare and Medicaid Services are leading the transition to an emphasis on patient outcomes and satisfaction. The transition from pay-for-service to pay-for-performance is taking place in an industry hampered by rising costs and limited available resources. In 2006, healthcare expenditures accounted for 16 per cent of the US gross domestic product and is estimated to grow to 20 per cent by 2020 (Jaffe et al., 2006). Various factors contribute to rising expenses, including an aging population and increased chronic conditions resulting in higher demand for care, poor quality and operational inefficiencies.\nResearch method: A systematic method was used to guide the literature review for assessing RFID's impacts on adoption in a hospital facility. This method, depicted in Figure 2, is divided into the four phases of problem definition, literature identification, assessment and analysis.\nResearch background: Inventory management in hospitals\nRFID technology: RFID can be used to track the flow of mobile assets by improving the asset check-in process and control of inventory storage. Improving this process requires integration with the facility's existing technology infrastructure and the placement of RFID readers in hallways, doors and/or cabinets throughout the facility to automatically detect assets as they are taken into and out of rooms for use as well as returning them back into the correct storage rooms.\nCost-benefit analysis: A CBA compares the monetary costs of installing and operating a program to the monetary benefits provided by the program. By weighing the advantages and disadvantages of a new technology implementation, one can make a decision on whether the change produces enough value to make a business case for the capital expenses incurred through adoption.\nGeneral CBA framework: Proposed CBA methodology for RFID in healthcare\nDetermining cost-effective assets to track: While it is clear that automated asset tracking presents a competitive advantage for facilities, it is not cost-effective to track all mobile clinical mobile assets. Currently, hospitals that are implementing RFID systems are tagging high-value, frequently needed mobile assets. Item-level tagging appears better for high-priced items, whereas pallet-level tagging would be better for inexpensive everyday items (Bose and Pal, 2005). Significant savings have already been realized by tracking costly mobile equipment such as defibrillators, ventilators, microscopes, electrocardiograms and infusion pumps (Crooker et al., 2009). Studies have noted that items lower than $15 in price would be a poor use of resources (Angeles, 2005; Schwirn, 2006). Table I shows some of the commonly tagged assets.\nRFID asset tracking performance in hospitals: Numerous pilot studies have been conducted showing the feasibility of implementing RFID technology for the purpose of asset tracking and inventory management. According to Davis (2004), the cost of an RFID system ranges from $20,000 for smaller applications such as patient tracking to more than $1 million for tracking thousands of assets and people:\nCalibration and validation: Facility managers and staff should use performance metrics to measure whether the organization is fulfilling its goals. Comparing the performance of the hospital before automated inventory tracking with the performance of the hospital following implementation can identify the impact of RFID. These metrics help raise awareness to the strengths and weaknesses of the organization as well helping in benchmarking studies with other hospital facilities. Table II lists useful metrics for each of IOM's delivery goals. Cost metrics include savings from excess inventory purchases, rentals and underutilization as well as labor savings.\nThe need for continued development: The many advantages of RTLS can be seen through past pilot studies. However, there are implementation challenges such as interference with medical devices, privacy concerns and a lack of global standards that pose as current and future challenges to adoption. The current state of RFID's development and performance limits it to assisting inventory managers with location, utilization rates and inventory audits in specifically designed warehouses and storage facilities and care rooms.\nConclusions and future work: Based on current literature, using RFID to manage mobile devices increases the utilization rate, decreases annual spending, allows withdrawal of funds for underutilized assets and establishes confidence that equipment is readily available when needed. These benefits are realized through improved staff productivity, quicker patient turnover, higher quality of care and more cost savings by enhancing the visibility of equipment throughout a healthcare facility. RTLS technologies allow hospitals to be prepared for emergencies requiring the immediate use of medical devices without delay. RTLS allows staff to determine the status, condition and location of essential equipment, leading to a decrease in patient wait time. Other improvements include lower per-patient cost of service delivery, effective workflow and operational efficiency with support staff spared from manual searches. The return on investment comes from the ability to locate and return excessive and unnecessary rented medical devices and preventing the purchase of superfluous pieces of equipment. As expressed by Wicks et al. (2006), the goals of RFID applications should look beyond cost savings and legal compliance by addressing patient satisfaction and operational effectiveness (Wicks et al., 2006). While this paper focuses on RFID's impact on inventory management, additional benefits and a greater ROI are achievable through patient tracking. Mercy Hospital was able to cut their staff's continuous surveillance of critical patients by 60 per cent; roughly $30,000 in labor savings over four months in addition to an improvement in staff morale (Davis, 2004).\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to analyze a conceptual framework to identify significant benefit and cost attributes of a radio-frequency identification (RFID) system for asset tracking in healthcare facilities. Sources of value function for RFID in healthcare are categorized according to major improvement. Several cost functions are proposed to measure aspects of automated tracking implementation.\nDesign/methodology/approach: The four phases of problem definition, literature identification, assessment and analysis were used to begin the research. A cost-benefit analysis (CBA) was completed to identify the factors within healthcare with major benefits, and finally, a recommended group of items were identified to track with the CBA.\nFindings: RFID to manage mobile devices increases the utilization rate, decreases annual spending, allows withdrawal of funds for underutilized assets and establishes confidence that equipment is readily available when needed. These benefits provide improved staff productivity, quicker patient turnover, higher quality of care and more cost savings. Real-time location system technologies allow hospitals to be prepared for emergencies requiring the immediate use of medical devices without delay and allow staff to determine the status, condition and location of essential equipment, leading to a decrease in patient wait time. Additional improvements were also found.\nOriginality/value: The transition from pay-for-service to pay-for-performance is taking place in an industry hampered by rising costs and limited available resources. Healthcare expenditures are estimated to grow dramatically, and various factors contribute to rising expenses, including an aging population and increased chronic conditions resulting in higher demand for care, poor quality and operational inefficiencies. Therefore, understanding benefits of new technology use like RFID is critical for improvement and efficiency in healthcare.\n", "role": "assistant" } ]
Brave new world: service robots in the frontline
[ "Consumer behaviour", "Ethics", "Artificial intelligence", "Privacy", "Service robots", "Markets" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The service sector seems to be at an inflection point with regard to productivity gains and service industrialization similar to the industrial revolution in manufacturing that started in the eighteenth century. Rapidly improving technology that becomes better, smarter, smaller, and cheaper will transform virtually all service sectors (Wirtz and Zeithaml, 2018). Especially exciting are the opportunities offered by robotics in combination with cameras, sensors, speech recognition, big data, analytics, artificial intelligence (AI), mobile and cloud technology, geotagging and biometrics. The likely impact of the infusion of robots in conjunction with AI and machine learning on frontline employees across industries and settings has attracted significant attention from business practitioners (Lelieveld and Wolswinkel, 2017; Manyika et al., 2017; Microsoft, 2018) and recently also from service scholars (Huang and Rust, 2018; Marinova et al., 2017; Caic et al., 2018; van Doorn et al., 2017).\nCharacteristics and capabilities of service robots: Definition of service robots\nThe customer perspective on service robots: Research has demonstrated that a quality core service is necessary but not sufficient for achieving a competitive advantage (e.g. Bharadwaj et al., 1993). During the service encounter, customers often place a premium on pleasant relations with service employees - sometimes described as rapport, engagement and trust, and so providing emotional and social value. However, by 2020, it is estimated that 85 percent of all customer interactions will take place without a human agent (Schneider, 2017).\nEthical and societal implications of service robotics: Robots will not only impact customers but are likely to touch all strata of society (Delvaux, 2017), which makes it important to examine their ethical and societal implications. Ethical and moral principles guide human interaction and agency, and influence decision making and behavior. The purpose of ethics is the improvement of the general well-being of all participants in society (Kuipers, 2016). Ethics especially focuses on protecting and improving the personal integrity, and human dignity makes sure that the rights of the weakest in society are protected and aims at limiting possible inequalities caused by the advancement of robotics (Veruggio and Abney, 2012). As summarized in Figure 1, the increasing use service robots will provide a series of ethical and societal challenges across the micro, meso and macro level that require critical reflection.\nConclusions, implications and further research: This paper explored the role of service robots in three main parts. First, it provided a definition of service robots, described their key attributes, contrasted the features and capabilities of service robots and frontline employees, and provided an understanding for which types of tasks robots will dominate in service delivery, where humans will dominate, and areas where humans and robots will be likely to collaborate. Second, it examined the implications for consumer-adoption in a sRAM. Third, it provided an overview of the ethical and societal issues surrounding robot-delivered services at the micro, meso and macro level.\n", "role": "user" }, { "content": "Purpose: The service sector is at an inflection point with regard to productivity gains and service industrialization similar to the industrial revolution in manufacturing that started in the eighteenth century. Robotics in combination with rapidly improving technologies like artificial intelligence (AI), mobile, cloud, big data and biometrics will bring opportunities for a wide range of innovations that have the potential to dramatically change service industries. The purpose of this paper is to explore the potential role service robots will play in the future and to advance a research agenda for service researchers.\nDesign/methodology/approach: This paper uses a conceptual approach that is rooted in the service, robotics and AI literature.\nFindings: The contribution of this paper is threefold. First, it provides a definition of service robots, describes their key attributes, contrasts their features and capabilities with those of frontline employees, and provides an understanding for which types of service tasks robots will dominate and where humans will dominate. Second, this paper examines consumer perceptions, beliefs and behaviors as related to service robots, and advances the service robot acceptance model. Third, it provides an overview of the ethical questions surrounding robot-delivered services at the individual, market and societal level.\nPractical implications: This paper helps service organizations and their management, service robot innovators, programmers and developers, and policymakers better understand the implications of a ubiquitous deployment of service robots.\nOriginality/value: This is the first conceptual paper that systematically examines key dimensions of robot-delivered frontline service and explores how these will differ in the future.\n", "role": "assistant" } ]
Integrating search engine capacity and gender preference within a social media captured authority locus of control
[ "Social media", "Search engine optimization", "User behaviour", "Barnard’s theory of authority acceptance", "Search engine technologies" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: 1.1 Strategic and tactical importance of online search engines\n2. User engagement and search engine technologies: Google initially created the idea of ranking page searches on the internet based on how many other pages linked to it. This feature subsequently led users to the most prominent pages. Google exemplified the search industry goal of delivering information that goes through everyone's hands, eyes, and hard drives into the future. Search engines offer a significant amount of user-friendly integration. A number of examples from various search engines are used for illustrative purposes which can be found below. Search engine companies establish something that all businesses strive to achieve with users: a relationship. Search engines are the hub of all information of products and services in the marketplace (Smith, 2008, 2010).\n3. Search marketing technology: 3.1 Social media and search engine benefits\n4. Method: 4.1 Specific research hypotheses\n5. Results and discussion: 5.1 Descriptive comparisons\n6. General conclusion and implications: 6.1 General managerial results\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to provide useful insights underlying the popularity of search engine technologies within a social media-intensive environment.\nDesign/methodology/approach: The degree of social interaction for social media platforms that integrate search engine technologies as part of the homepage and related experience is very mixed on part of its users. Through Barnard' theory of authority acceptance, social media and its popularity may be examined by the ability of its users to create effective messages that can be broadcasted to many, yet controlled by individual. The hypotheses tested the interaction of social media and search engine with gender and technological ease-of-use factors.\nFindings: The statistical evidence suggested that significant technological and ease-of-use aspects of search engines are not meaningful, based on gender alone. Males may slightly be prone to take advantage of such technologies, but their search and use patterns are not much varied from their female counterparts. Social media, generally more fully captured authority in individual search patterns, and a number of interactions among gender status, search engine characteristics, and social media were found to be significant and profound. The testing of these hypotheses directly reflect the complexities of unique needs among users of search engines within a social media environment.\nPractical implications: Search engine technologies with a social media context has allowed for the development of a modern, user-driven internet experience that has been powered by users' imagination and is designed to at least partially satisfy users' need for self-directed engagement. Organizations are well advised to provide a mindful, less controlled, and more interactive presence of potential users, especially through an increasingly mobile presence.\nOriginality/value: Individuals as well as organizations are rapidly discovering that it is becoming easier to share and distribute their content, especially for more creative and innovative content, among all of its users. As businesses continue to focus on the quality of one's own content, individuals are increasingly taking advantage of some tools to exert more control over their experiences and what they are willing to share, resulting in more user-based partnerships will formulate. As the transition of traditional forms of marketing to newer forms of integrated marketing, the future for search engines as marketing tools by social media users appears to be very promising in adding contextual content within users' homepage.\n", "role": "assistant" } ]
Competitive intensity and inefficiency in European container ports: An empirical investigation using SFA
[ "Stochastic frontier analysis", "Inter-port competition", "Port policy" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Over the past two decades, development of a common policy framework regulating the governance of major European seaports has been in progress. Simultaneously, several governments of European Union member states have implemented own institutional reforms in their port sectors (Chlmoudis and Pallis, 2005; Baird and Valentine, 2007; Pallis and Syriopoulos, 2007; Valleri et al., 2006). A recurring objective of proposed common port policies has been to increase the autonomy and transparency of port finances, imposing stricter control on national subsidization to achieve greater competitive pressure between ports. In drafting the principles of a uniform port policy, the European Parliament stated that overcapacity has been a problem for the European container port sector, recognizing also that competitive investment, i.e. a \"spiraling\" effect of ports competing with infrastructure size to attract and accommodate a fleet of continuingly expanding vessels, was a potential cause of this overcapacity (European Parliament, 1999). Such concerns cannot be seen as unfounded, as various studies have pointed out that higher competition in container port service production is a potential inducer of excess capacity (Haralambides, 2002; Haralambides et al., 2002). At the same time, there is an obvious role for competition in promoting and producing efficient outcomes through the deterrence of monopolistic pricing and related practices, which is similarly emphasized by seminal work on the subject (Goss, 1990).\n2. Background and theoretical aspects: 2.1 Port competition and policy in Europe\n3. Methodology: The aim of efficiency analysis is to measure the extent to which a firm or some other decision-making unit (DMU) achieves a maximum level of output given a set of inputs, combined in an optimal way (Farrell, 1957). The measurement of efficiency is based on an estimated distance between the firm's actual level of production and an efficient frontier of maximum achievable production for given sets of inputs. Empirical analysis of efficiency requires some method of estimating this benchmark frontier. The two dominating methods are DEA and SFA (Lovell, 1993). The former is a non-parametric linear programming approach that has the advantage that it does not require any presupposition of the firm's production technology (Charnes et al., 1978). As a drawback, non-parametric methods complicate statistical testing of hypotheses, a problem which can be solved using a stochastic approach. In stochastic frontier analysis, a cost or production function is estimated with a two-part composite error term. One part is an inefficiency effect indicating distance from the frontier, and the other part is some well-behaved noise term (commonly assumed to be normally distributed) (Meeusen and van den Broeck, 1977; Aigner et al., 1977). In SFA, choice of functional form is crucial. In empirical applications, the chosen approaches are often that of the simple Cobb-Douglas, which restricts the returns to scale in production to be a constant (not to be confused with returns to scale being constant) or the generalized translog form, which adds cross-product terms for each input and does not restrict substitution elasticities to be unity (Coelli et al., 2005). For the N-input, single output case, the Cobb-Douglas function is:\n4. Empirical production functions and data: This study takes a production function approach to deriving estimates of technical efficiency of seaports. Before arriving at an estimable function, a few conceptual features of port services should be noted. Jansson and Shneerson (1982) note that port services are (as all services) non-storable. This implies that production and consumption need to occur simultaneously. From this obvious reasoning, two general statements can be made:\n5. Results: In Table III, the estimation results of the stochastic frontier production function are presented. The two columns are results for the Cobb-Douglas and translog models [equations (10) and (11)]. A Likelihood-Ratio test confirms that the Cobb-Douglas functional form can be rejected in favor of the translog at a significance level of 1 per cent. The output elasticities, which can be conveniently read from the Cobb-Douglas estimation results, show that an increase in 1 per cent in terminal area is roughly associated with a 0.46 per cent increase in throughput. The corresponding elasticities for berth length and number of cranes is 0.22 per cent and 0.47 per cent, respectively. The estimated output elasticity of maximum depth also has the expected positive sign but is not significant at a confidence level threshold of 5 per cent. For both specifications, an imposed restriction of CRS can be rejected. The sum of input parameters in the Cobb-Douglas function is 1.51, indicating increasing returns to scale in the production of port services. While the translog model parameters cannot be directly interpreted as output elasticities, partial elasticities can be calculated for each parameter. Such estimates are derived by calculating the percentage increase in fitted output resulting from a corresponding increase in a single input. The elasticities differ depending on the levels at which the variables are evaluated. Table IV (second column) shows these estimates when the median input values are used. This can be considered to approximate a typical port in the sample. While the elasticities can be interesting in themselves, they are not further discussed or analyzed in this study. Rather they serve as a robustness check to see that the estimated production functions appear to give reasonable results.\n6. Possible implications for port policy and research: The finding that the intensity of local competition is positively related to port efficiency could be seen to suppress some of the concern that inter-port competition is in itself a cause of overcapacity. Further, it implies that levelling competition between ports could be a suitable direction for policy aimed to improve efficiency. While the analysis does not show any distinguishable negative effect of competition (for any spatial level) on efficiency in a panel that comprises a decade's worth of observations, it is important to note that the somewhat crude nature of production frontier estimation makes it difficult to account for heterogeneity in port service production. With more refined data, particularly including the user side of production, the use of efficiency analysis techniques would be able to provide a higher level of accuracy and policy relevance in its results.\n7. Conclusions: Proposals for a common ports policy framework in the European Union have advanced during the past decades. A recurring objective of such proposals has been to increase the autonomy of ports and to achieve more competitive markets. A concern that is often voiced with regard to this objective is that competition may exacerbate perceived problems of excess capacity in container ports, which can be harmful to the efficiency of the maritime transport system. In this study, the relationship between intensity of competition and technical efficiency is analyzed using a stochastic frontier approach and a dataset of 77 large European container ports over the period 2002-2012. The results indicate that there is no significant negative effect of competition on efficiency. In fact, for ports within a proximity of 300 km, a higher level of competition is found to be associated with a higher level of efficiency. The analysis suggests that reducing the market concentration by 1 per cent in the local area of a port may yield an efficiency increase of roughly 1.2 per cent. This implies that while excessive capacity expansion may sometimes result because of fierce inter-port competition, this does not appear to have been the general outcome for large European container ports during the studied period. The policy-relevant conclusion is that focusing efforts to reduce monopolistic powers of ports in local networks could be a viable way to improve efficiency. It is notable that this study, similar to most studies of port efficiency, would benefit from a greater level of access to micro-level data in ports. This includes turnaround times of vessels, terminal-level outputs and labor data. Gathering and using such data for the purpose of port efficiency analysis and policy evaluation is a difficult but recommendable task for future research.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to empirically examine the relationship between intensity of competition and technical efficiency of large European container ports, accounting for regional diversities and spatial aspects of inter-port competition.\nDesign/methodology/approach: The analysis consists of applying a stochastic production frontier approach to a dataset of 77 large European container ports over the period 2002-2012, with inefficiency terms simultaneously modeled as a function of (among other factors) a constructed index of competitive intensity at different spatial levels.\nFindings: The results indicate that there is no significant negative effect of competitive intensity on efficiency. In fact, for competing European ports within a proximity of 300 km, a higher level of competition is found to be associated with a higher level of technical efficiency.\nOriginality/value: The originality of the paper stems from its particular focus on European port regions and its novel findings in this context, which have implications for the discussions regarding pro-competitive port policy and regulation in the European Union.\n", "role": "assistant" } ]
Serial commitments clearance in Rastin Banking
[ "Commitment", "Collateral", "Commitment clearing", "Guarantee", "Rastin Banking", "Serial commitments" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Rastin Banking[1] is a new operational Islamic banking system which, by studying theoretical and operational banking difficulties and on the basis of the latest scientific and technical innovations in the field, presents legal and operational solutions. Each of the complimentary systems[2] and financial subsystems[3] of Rastin Banking has been designed for solving specific problems (Bidabad and Mastorakis, 2014). Serial Commitments Clearance (SCC) system was designed to provide necessary arrangements for settlement of obligations of those who, on the one hand, serially owe someone and, on the other hand, are creditors to other persons.\nCommitment, a legal view: Commitment is created through a legal relationship resulted from a contract, cadence or by force according to which a person is obliged to transfer a property or wealth, or due to do or not to do a specific action. Terminologically, commitment means agreement and making a treaty, and the two words of commitment and agreement are used with their infinitive meanings as contract (Ghanavati, 2000, pp. 42-46). \"Commitment\" is used as meanings of \"condition\", \"obligation\", \"treaty\" and \"guarantee\" regarding the concerned case, but any kind of enforced guarantee, transfer of property or wealth, forcing to do something or preventing from doing something in general, can be considered as the subjects of a commitment (Langeroudi, 2003, p. 27). Commitment is the primary source of all legal relationships, which creates the rights of the two parties of a contract (obligor and obligee). Therefore, each contract or cadence which has been formed by a decision to do so (legal action), or happened without decision (legal event), has legal effects (Langeroudi, 2003, p. 9). The concepts of \"Zemme\" in Islamic Fiq'h (debt or due) and \"obligation\" in many Islamic countries are other expressions of the concept of commitment (Ghanavati, 2000, p. 67), which obligates the commitment of debts due to civil responsibilities and binding contracts (Imami, 1971). The Civil Law of Iran (Katouzian, 2003) expresses commitment as an initiated action from a contract and defines: \"contract is the agreement of commitment of one or some persons to one or more other persons to do something that is accepted by them\". Commitment is a duty of the obligor and a right of the obligee; nonfulfillment of commitment causes the enforcement of the obligor and compensation of the losses of the obligee in case of making loss, according to the legal rule of causality (Gorgi, 1986). Subjects such as proving, enforcement and termination of commitment are among the main subjects of commitment rights. The subjects such as suspension, substitution and clearing of commitments are among the secondary subjects, the latter is under consideration in SCC here in this study (Ghaemmaghami, 1999, p. 21). In addition to annulment option and canceling the contract, termination of commitment (Katouzian, 2003) includes subjects such as fulfillment of obligation (when the committed person fulfills his commitment, the commitment is terminated), mutual rescission (two parts of the contract agree to cancel the contract), release (obligee waives his own right), substitution (both parties agree to substitute the previous commitment), set off (clearance of two homogeneous debts between two persons in oppositeness) and acquisition of debt (unity in the attribute of the claimer and debtor of a commitment in one person)[4].\nSerial commitments clearance: Serials commitments is referred to the commitments of several persons each of them is a debtor to the previous person and is a creditor to the next person - except the first person, who is only a creditor to the next person, and the last person, who is only in debt to the previous one.\nSerial chain of obligations and claims: Theory of bankruptcy chain was explained about serial financial relationship of traders and how insurance hedges it (Bidabad and Mastorakis, 2014) Let us now adapt \"bankruptcy chain\" in the form of serial obligations and claims. Any firm (or person) at time t has some assets and liabilities. Its total assets (Wi) is equal to the value of all goods, physical capital and other acceptable items in firm's portfolio (Ci) plus its claims (Fi) to others. That is:\nSerial commitments recognition algorithm: Consider two n x n square matrices of D and F in which n is the number of economic activists:\nOperational considerations: From the operational perspective, the following notes should be considered regarding the explained algorithm.\n", "role": "user" }, { "content": "Purpose: This paper aims to propose a system for settlement of obligations of those who, on the one hand, serially owe someone and, on the other hand, are creditors to other persons. Serial Commitments Clearance (SCC) system provides the necessary arrangements for this purpose.\nDesign/methodology/approach: Theoretical discussions of serial chain of debits and credits were conducted by considering the laws and regulations of commitment clearance.\nFindings: By considering laws and regulations of commitment clearance and introducing theory of serial chain of debits and credits, this system proposes an algorithm for recognition of serial commitments.\nResearch limitations/implications: This process is set in connection to the Collateral Registration System and Mortgage Securitization System in Rastin Banking, while considering the legal and operational problems. Accordingly, banks, notary offices and other authorized authorities can clear serial obligations of persons due to their requests and agreements and release their collaterals and guarantees as far as the debts of the persons are equivalent.\nPractical implications: This system will cause financial release and efficiency of many economic firms. In addition, banks will gain commission for rendering this service.\nSocial implications: SCC is a model that can be used in all countries, especially those which have more uncertainties and traders need more pledges for their transactions.\nOriginality/value: This study fulfils an identified need to solve practical legal problems in vindication of rights.\n", "role": "assistant" } ]
An analysis of consumer reactions to green strategies
[ "Green marketing", "Green product", "Sustainable", "Green operations", "Environmental", "Ecological", "Environmental management" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\nIntroduction: Sustainability has become the mantra for companies seeking to create a competitive advantage in the global marketplace. Recent interviews with over 750 CEOs from around the world revealed that 93 percent feel that implementing sustainability programs that mesh with their core businesses would be critical to the future success of those businesses (UN Global Compact, 2010). These programs range from developing totally new green products, developing recycled, or refurbished products to engaging in greener processes. The primary sustainable business strategies pursued by companies can be grouped into two broad categories (Esty and Winston, 2006):1. Managing downside risk. These strategies are designed to ensure long-term resource availability, and are intended to enhance productivity, reduce environmental and regulatory costs, and lower supply chain costs. Many of these strategies have been relatively easy to adopt (e.g. changing manufacturing [methods] or distribution practices in order to reduce waste). However, some of these strategies require significant upfront investment (e.g. Frito-Lay's zero emission potato chip plant in Arizona, building a LEED certified manufacturing plant - www.fritolay.com/our-planet/working-toward-minimal-impact.html). This emphasis on productivity and cost in the application of sustainability strategies is popularly referred to as eco-efficiency.2. Building upside potential. This group of strategies addresses the potential for product expansion as part of sustainability efforts. Recent studies indicate that 93 percent of consumers say they participate in efforts to sustain the environment (Hartman Group, 2007), while 75 percent say it is important or somewhat important that they buy from environmentally responsible companies (WPP, 2010). Companies have responded to the shift in consumer behavior by focusing on the increasing environmental needs of consumers. This is accomplished through development of breakthrough products (e.g. the Toyota Prius hybrid auto), adding green components to existing products, reusing or recycling materials, improving corporate reputation through green brands, or other environmentally focused activities (e.g. reducing energy consumption, Intel changing the organizational culture by incorporating environmental metrics as part of employee compensation).Although there are a multitude of sustainable green strategies, consumer responses to these strategies are also varied. We seek answer to one important research question of interest, that is, how do consumers perceive different green strategies? This project focuses on three such popular strategies: new green product strategies, recycled or refurbished product strategies, and improved green processes strategies. As one primary strategy, a firm can choose to produce new green products. These products run the spectrum from complex or innovative new products such as Nissan's Leaf or the Clorox Green Works line of cleaning solutions, to enhancing existing products by introducing green qualities to their products, such as SC Johnson's incremental addition of safer ingredients in their product line. Consumers clearly identify these products as new and as green. A second strategic option for sustainability is the production of recycled or refurbished products. Application of this approach involves a sophisticated reverse logistics system, and requires the cooperation of both channel members and consumers. The newly produced goods that result are made of reclaimed/recycled/refurbished components. We propose that consumers evaluate these types of products differently than new green products. Third, strategies for developing green processes can be developed from firms' everyday business activities. These methods may include reducing water usage, plant/manufacturing emissions, energy consumption, solid wastes, and acid rain emissions (sulfur dioxide, nitrogen oxide, and ammonia).Consumer studies have analyzed a number of related issues: the \"ecological\" consumer (McDonald and Oates (2006), Straughan and Roberts (1999), willingness to pay a price premium for green products (Veisten, 2007; Vlosky et al., 1999), perceived quality of green products (D'Souza et al., 2007), and attributes that may motivate consumption of green products (Schuhwerk and Lekkoff-Hagius, 1995). Studies investigating corporate green strategies have primarily examined the effect of green strategies on life cycle costs, product differentiation, competitive management and new market development (Ginsberg and Bloom, 2004; Reinhardt, 1998).Unfortunately, few studies have measured consumer reaction to alternative green strategies. Though different segments of the green market (Hartman Group, 2007, Ginsberg and Bloom, 2004) are often used as a rationale or as a guide for the development of green products or corporate green processes, the impact of these strategies is often post hoc. This manuscript closes the literature gap by measuring consumer perceptions toward non-green options and different corporate green strategy alternatives. In addition, we consider the impact brand name and price have on consumers under the different green and non-green strategic possibilities.\nLiterature: Quality of environmental products\nMethod: Design\nResults: H1. There will be no significant difference in perceived quality between new green products, recycled/refurbished products, and products from a company that engages in green processes versus non-green products that come from a company that does not engage in green processes.A MANOVA was run on the complete model, using each of the dependent variables separately and all of the independent variables (brand name, pricing, and green factor). The model with perceived quality was not significant (Green Factor p=0.5779), thus supporting H1.H2A. There will be no significant difference in purchase intentions between new green products, recycled/refurbished products, and products from a company that engages in green processes versus non-green products that come from a company that does not engage in green processesSince H1 was supported, perceived quality was dropped from further analysis and we focused on purchase intention. In the survey described earlier, we included the eight eco-variables used by Borin et al. (2010). Factor analysis was performed on the eight eco-behavior variables. Three variables had a loading above 0.7 on the first factor. Variance explained dropped significantly for additional factors. The variables \"willingness to pay an additional 10 percent,\" \"prefer to buy from eco-friendly companies,\" and \"would support a required environmental course\" were combined into a new variable, which we refer to as eco-behavior. This variable was used as a covariate.The results of the regression using purchase intention as the dependent variable are presented in Table I. Green factor is significant (p=0.002) which supports H2A.Eco-behavior covariate\nDiscussion: As green segments continue to grow in the marketplace (Gfk Custom Research North America, 2009, Hartman Group, 2007), many firms see opportunities to develop products that will meet the needs of these segments (Esty and Winston, 2006). Even segments that do not identify themselves as strongly green indicate that they are willing to purchase green products if they are priced properly, if the quality is maintained, and if there is no sacrifice in convenience. Our project provides a number of important implications for companies that are pursuing sustainable product strategies.It's good to be green\nLimitations and future research: Our project focused on a durable product category. Prior results (Mobley et al. 1995; Essoussi and Linton, 2010) have found category-specific differences, though others have not. Future research needs to address a wider range of products as the stimuli. Additional categories such as foods or personal grooming products would be interesting to investigate. Consumers may be much more receptive to green products in these categories, though new green may be perceived as much stronger than recycled. In fact, for personal products recycled/refurbished may have lower consumer perceptions than regular products (e.g. recycled toilet tissue).Consumers appear to be unwilling to pay for the green benefits contained in products. Our study presented two different price options. However, the market has revealed that sustainable products have a wide variance in price premiums for different categories. Further research should investigate additional price/green interactions, and evaluate whether, for certain green products, consumers are willing to pay premium.Research should be conducted that further explores the brand name and green strategy interaction. Our study found no benefit for a well-known brand name that has a strong environmental record. This observation calls for research on more familiar consumer product goods, such as household items (e.g. toothpaste) or office products (e.g. copiers). There may be a benefit for strong consumer brands that have developed a reputation for developing sustainable products or processes (e.g. Johnson & Johnson).Because the consumers in our study responded favorably toward a green strategy but were indifferent to the type of strategy the company employed, it is necessary to study the costs of the different types of green strategies. With regard to the three green strategies discussed in this paper, it may be more cost effective for companies to focus on developing new green products or on engaging in green processes. These approaches, as opposed to developing recycled/refurbished products, may be more profitable from an operational point-of-view because recycling/refurbishing requires considerable reverse logistics and cooperation from channel members (customers and other channel members). While we found no statistical difference between recycling and the other strategies, this issue needs to be investigated for other products in future research.\nConclusion: The green movement and the managerial pressures to respond are here to stay, mirroring the pressure to adopt an Internet presence that businesses experienced in the 1990s. Increased concern for the environment on a broader social level has influenced consumer behavior, with the result that green aspects of products and companies are becoming as important a criterion in search behavior as price, brand, and product availability. Companies must now consider sustainability as part of overall business decisions, and once having decided to enter the green market, must evaluate the range of alternative approaches available. The current project found though it is beneficial to enter the green market, the specific choice of strategies is not as important for the consumer. However, specific choice of strategies is important from a company's perspectives. Developing new green product requires collaboration with supply chain members, R&D, universities, and customers. Recycled/ refurbished strategies require detailed reverse logistics planning and cooperation from channel members and customers. Green processes require significant changes in organizational culture. In this paper, we have shown that green strategies do matter to eco-customers and industry reports indicate that this group is growing quickly. Companies, therefore, can grow in this market provided they develop a well-defined road map.\n", "role": "user" }, { "content": "Purpose: - The purpose of this paper is to investigate the impact of three green strategies on key consumer metrics. More specifically, it aims to measure consumers' purchase intentions of new green, recycled/refurbished products, green company processes and a non-green product/process.\nDesign/methodology/approach: - Between subjects 2x2x4 experimental design with two levels of price (high and low), two levels of brand name (known and unknown) and four levels of green strategies.\nFindings: - Purchase intentions for green product and process strategies are significantly higher than non-green approaches. However, post-hoc analysis shows no significant advantage of one green strategy over another. Price and brand name do not have significant interactive effects with green strategies.\nPractical implications: - Although it is essential that companies develop green strategies for the eco consumer it is not important what specific strategy is selected i.e. going green is the key. Also, despite the continued growth in the demand for green products, price is still the most important driver for consumer purchase - even for the eco consumer. Lastly, despite continued improvements in functional performance, green products do not have a significant advantage in perceived quality. Companies cannot focus completely on the green nature of their products or processes.\nSocial implications: - Consumers believe that purchasing green products or products from green companies may be a way they can help the environmental problems society faces today. This project provides guidance to companies pursuing this market by evaluating different product and process approaches to this growing social trend.\nOriginality/value: - This project is one of the first to focus on the consumer impact of different corporate approaches to the green market.\n", "role": "assistant" } ]
Education and employment in two Chilean undergraduate programs
[ "Employment", "Student satisfaction", "Chile", "Inequality", "Psychology program", "Teaching program" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Recent, massive university student protests in Chile were motivated in part by doubts about the quality of teaching and learning (Espinoza and Gonzalez, 2013). The democratic governments that came into power after 1990 had promised that expansion of education would not only increase economic growth, but also turn Chile into a more egalitarian society. The economy did grow and enrollments in universities have more than doubled in 20 years. By 2015 the gross enrollment ratio for post-secondary education in Chile was over 50 percent (Espinoza and Gonzalez, 2015).\n2. Research on student ratings of teaching quality, student satisfaction, university ratings and employability: 2.1 Teaching quality\n3. Methodology: This study uses quantitative analysis of questionnaire responses to assess the relationship between participants' characteristics, their satisfaction with their university degree program and their employability experiences. Cross-tabular analysis and tests of differences of means are used to calculate the statistical significance of the relationships.\n4. Results: With respect to the first hypothesis, there is no relationship at all between the level of education of a graduate's mother, and the selectivity of the university attended. There is, however, a relationship between the type of secondary school attended and university selectivity. Graduates who attended municipal secondary schools, and those who attended private schools, were more likely to attend a selective university than those attending a state-subsidized (voucher) private school. This relationship is highly significant (p<0.000).\n5. Discussion: The data collected in this small study of three universities provide a complex view of how family background, gender, secondary school preparation and degree program quality influence (at least initial) salaries of psychology, and teaching, graduates. Some of the results are not surprising. Students in psychology and teaching programs had a higher employment rate, six months after graduation, than that of the general labor force. On average, the graduates' incomes were slightly higher than the national income per capita. Psychologists earned more than teachers; independent of occupation, men earned more than women.\n6. Conclusion: This study of the employment of university graduates confirms the results of prior research: graduates of more selective universities are hired more quickly and receive higher compensation for their employment and have higher opinions of their degree program. Both graduates and employers believe that these graduates will be more productive in their organization than would graduates from less selective universities.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to determine which of the two variables would be a more reliable proxy for quality of university training: graduates' satisfaction with their degree program, or institutional prestige.\nDesign/methodology/approach: Graduates of professional psychology and teaching programs from three Chilean universities responded to a questionnaire asking their perception of different aspects of their degree program and experiences in their first employment. The three universities differ significantly in the proportion of applicants admitted, and in their prestige.\nFindings: Salary levels are highly related to profession, but unrelated to graduates' ratings of quality of curriculum or teaching methods. Overall satisfaction with the university experience is not linked to job satisfaction. Job satisfaction, on the other hand, is highly influenced by salary and secondarily by instructional practices and perceived work relevance of the degree program.\nResearch limitations/implications: This study is based on data from 3 of Chile's 60 universities, and graduates of two programs. Most employment in the two professions is regulated by labor agreements. Generalizability of results is limited. Graduates may not have been employed enough to demonstrate their capacities.\nPractical implications: The findings offer more evidence that prestige ratings are an unreliable indicator of the quality of formation offered by universities. If the government seeks to reduce income inequality, public subsidies of higher education should be based on program quality rather than on institutional prestige.\nOriginality/value: The findings are directly relevant to the current debate in Chile about what might and what might not help to reduce severe economic inequality.\n", "role": "assistant" } ]
Brand evaluation, satisfaction and trust as predictors of brand loyalty: the mediator-moderator effect of brand relationships
[ "Brands", "Brand relationships", "Trust", "Brand loyalty", "Brand evaluation", "Consumer satisfaction" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this issue.\nIntroduction: Brands are complicated entities. They serve several valuable functions, including acting as marketers of the offering of the firm by helping customers in their choices and being assets in a financial sense (Keller and Lehmann, 2006). Brand can facilitate customers by reducing their search costs and companies by increasing the probability of repeat purchases that enhance the company's financial performance. Therefore, customer loyalty is one of the most fundamental constructs in marketing, and the investigation of factors that drive brand loyalty is a key concern for marketers, especially in an era where consumers do see supposedly different brands as being more and more similar and they increasingly express no preference when evaluating brands in certain product categories (Schultz et al., 2014).\nSetting the scene: empirical study variables: Brand loyalty\nSetting the scene: the proposed models: Understanding how corporate image and reputation impact on customer purchases is one of the key questions of brand management research currently (Keller and Lehmann, 2006). Customers tend to be more loyal to brands that they evaluate higher and have a good attitude towards, and brand reputation is one of the antecedents of brand loyalty (Gilbert and Hewlett, 2003; Gounaris and Stathakopoulos, 2004). The brand image plays a key role in the development of brand relationships (Esch et al., 2006). Brand loyalty only exists while brands project attractive images for consumers' momentary experiences, amplifying a subversive tendency, which modifies the distinction between daily life and aesthetic ideal (Thompson, 1997). The way that the brand is perceived and evaluated is one of the more important predictors of current purchases (Esch et al., 2006). Thus, one could argue that this link between the evaluation of the brand and brand loyalty exists and is positive. In this study, it is hypothesised that:\nMethodology: The study was primarily quantitative, with a qualitative phase, which helped the development of the quantitative phase. In total, ten interviews with students and three focus groups with five participants, each possessing different profiles (one with students, one with 25-40 people and one with over 40 people), were conducted. They aimed to identify the most suitable definition of the constructs when a relationship with the brand is investigated and to choose the specific product category for this study. In terms of the construct definition, alternative descriptions of the constructs provided in the literature were presented to the participants and they were asked to rate them for the brands that they seemed to build relationships with as individuals. This led to the final definition of the constructs for this research. In terms of the choice of the product category, it has been appreciated in the literature that, although consumer-brand relationships exist in many product categories and are not product category specific (Fetscherin et al., 2014), \"relationship-friendliness\" and the intensity of the relationship depends on certain characteristics of market segments and products features (Christy et al., 1996), as well as the product category (Fetscherin et al., 2014), while others suggest that the development of a relationship is feasible for high-involvement products characterised by inelastic demand, where regular interaction with consumers occurs (O'Malley and Tynan, 2000). From three broad product categories (clothes, personal care products and cosmetics), one of the products that the informants in the focus groups seemed to use and have a relevant preference and develop relationships with was lipstick. Cosmetics is one of the product categories that has been used to examine consumer-brand relationships (Francisco-Maffezzolli et al., 2014), and it is argued that consumers increasingly express no preference for it (Schultz et al., 2014), and therefore, studying the formation of brand loyalty in this product category is of importance. Using a product that customers form either a strong or weak attachment with is a practice that is used in the literature to determine an offer, which can help in the measurement of the strength of relationship (Thomson et al., 2005). This product was also convenient, as it was easier for the researcher to approach female respondents.\nResults: Table II reports the number of items, the descriptive statistics and the internal consistency of the perceptual constructs used in this study. All of the scales used in the study are reliable, as the dimensions of the constructs measured had Cronbach's alpha coefficients in excess of 0.80, values higher than the anticipated acceptable level of at least 0.70 (Hinkin, 1995) or very high and statistically significant at the 0.01 level Pearson correlation. To further test the internal consistency of the measures, the inter-correlations of the items loading in the same factor and the item-to-total correlations were also examined (Gerbing and Anderson 1988). A principal component analysis with Varimax rotation clearly separates the two dimensions of brand relationship (Appendix AII), while, for all the remaining constructs, all the items are loading on the same factor (Appendix AII). The means of the items that belong to each construct is the starting point of the analysis.\nDiscussion: Most of the existing research on branding focuses on familiar and well-researched concepts associated with branding, such as trust, satisfaction, brand reputation and loyalty. These concepts are always of importance. As we live in a relational era, transactions do not secure long-term benefits for the producers and the consumers (Valta, 2013). Consumers develop relationships with brands (Fournier, 1998; Veloutsou, 2007, 2009), which vary in terms of the communication they would like to have with the brand and the positive emotional bond they develop with the brand (Veloutsou, 2007; Veloutsou and Moutinho, 2009; Jurisic and Azevedo, 2011; Morgan-Thomas and Veloutsou, 2013; Chen and Lin, 2014). Researchers need to further examine the role of the various forms of relationships with brands and the contribution of these relationships to the formation of other well-researched constructs in the well-established branding literature.\nManagerial implications: Product and service providers should consciously try to have brands in their portfolio that individuals are able to develop strong relationships with. Given that the existence of a strong consumer-brand relationship is a better predictor of brand loyalty than brand evaluation, trust and satisfaction, producers should engage in activities that will enhance both the communication and the emotional element of the relationship. These are likely to lead to stronger brand loyalty, with all the expected outcomes such as increased purchase intention, sales and profits. Because companies' actions influence and can even dilute brand reputation (Puzakova et al., 2013b), managerial actions should focus on developing meaningful communication interactions that will be evaluated positively from the market. Producers should also focus on the emotional content of a positive brand relationship, which leads to enhanced brand loyalty.\nLimitations and further research: This study has certain shortcomings. Some of them are related to the nature of the sample, others to the methodological choices. In terms of the nature of the sample, the data were collected in a big city and the geographic area was limited. The sampling procedure was not probabilistic, with no quota. All respondents were women, while a great proportion of the respondents are younger consumers, who might have a different behavior than consumers with a different profile. Surveying a larger, more diverse pool of respondents would allow further generalization of the findings. The respondents chose their preferred brand of product, which may have resulted in having reports of stronger relationships between respondents and the selected brands. Therefore, the results may not be valid for contexts where weaker relationships with the brands are evident, or for brands that the consumer may not wish to include in their consideration sets. Researchers might want to replicate the study in other cultural contexts, as intercultural differences may lead to different relationship expectations.\n", "role": "user" }, { "content": "Purpose: This paper aims to examine whether the strength of positive brand relationship can either mediate between trust, satisfaction, attitude towards the brand and loyalty or moderate the link between these variables. Existing research has established that trust, satisfaction and the attitude towards the brand contribute to the development of brand loyalty. However, recently, there is a growing stream of research indicating that the brands are not only facilitating transaction but companies can also use them to develop and maintain links with their customers. The exploration of the role of brands in the development of bonds with the customers is still very limited.\nDesign/methodology/approach: Data were collected from 189 women who are using lipstick in Glasgow, Scotland. Respondents were asked to answer a questionnaire keeping in mind their preferred brand.\nFindings: The findings revealed that the strength of the consumer brand relationship is a very strong predictor of brand loyalty. They also suggested that brand relationship does not moderate the relationship between brand trust, satisfaction and brand loyalty, but it mediates the link amongst these constructs.\nResearch limitations/implications: One product category was examined. The data were collected from females in a big city with non-probabilistic sampling. Most of the respondents were younger consumers, who may have different behavior compared to older consumers.\nPractical implications: Managers are constantly looking for ways to increase brand loyalty. They need to appreciate the importance of consumer engagement with their brand through positive brand relationships. Therefore, they need to try to develop brand identities and project brand images that will make the brands look as appealing to the consumers as relationship partners.\nOriginality/value: This paper supports the view that brands are not only facilitating transactions but can also develop and maintain links with their customers, especially engaged customers, and the exploration of these bonds is still limited. This paper is adding to the literature on brand engagement and brand relationships from a quantitative perspective and is contributing to theory building, as there is no clear theoretical view on whether the brand relationship has a direct effect, i.e. either mediates or moderates the link between these variables.\n", "role": "assistant" } ]
Social media marketing and communications strategies for school superintendents
[ "Twitter", "Technology", "Stakeholders", "Social media", "Schools", "Blogs", "Social networking", "Effective communication", "Online video", "Podcasts", "Superintendents", "Social CRM" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The ever-expanding social media landscape has increased the opportunities for people to communicate with one another at any time and in any location. Smartphones, iPads, and other technology devices permit a constant flow of two-way communication and information at home, walking down the street, or vacationing overseas. As our interconnectedness continues to grow, our methods of personal and organizational communication have had to transform as well.\nPrevious research: In an era of higher levels of accountability for student achievement, tighter school district budgets, and increased options for parents regarding school choice, the importance of effectively communicating and marketing school programs and successes to the community as a whole has never been more important. Ferriter et al. (2011) found that branding is beginning to take hold in the education world as school administrators learn from the public relations blunders occurring in the business world. Administrators are leveraging expanded social media communications with their stakeholders into additional community investment to improve their overall quality.\nMarket orientation (MO), relationship marketing (RM), and social customer relationship management (SCRM): The fields of communications and marketing are filled with theories describing interactions between individuals. MO, RM, and SCRM each provide a framework upon which to examine the relationship that exists between a school administrator's use of social media as a communications tool as well as his or her ability to market the school's successes and mobilize continued community support for school programs.\nPurpose of the study: Scant peer-reviewed literature exists that describes the experiences of superintendents who have been utilizing multiple social media tools as part of their comprehensive communications efforts with their stakeholders. Additionally, little has been written about the reasons why school administrators have embraced these alternatives. The lack of literature led to the development of the two guiding research questions:\nMethodology: This study involved school superintendents from across the USA and Canada. It examined the comprehensive communication practices of superintendents in order to get data from schools where social media use by administrators is frequent. Urban, suburban, and rural school district settings were intentionally selected in order to expand the applicability of the findings. Conversations were held with superintendents from a cross-section of schools in four different regions of the USA and Canada in order to gain an understanding of how social media tools were being incorporated into the comprehensive communications practices of varied school settings.\nData collection methods: The data collection methods used in this research study included telephone and in-person interviews. The study also included an analysis of the administrators' social media tool usage. Interviews were conducted with currently practicing superintendents who used at least two of the following as part of their comprehensive communications with stakeholders: blogs, microblogs (e.g. Twitter), social networking sites (e.g. Facebook), podcasts, and online videos.\nData analysis: Yin (2011) recommended five steps to follow in the analysis of qualitative research: compiling, disassembling, reassembling and arraying, interpreting, and concluding. First, the primary researcher compiled the data. This step involved gathering the interview transcriptions, the field notes taken during the interviews, and the data on social media tool usage. The primary researcher followed the advice of Maxwell (2005) and began to organize and analyze some of the data immediately after conducting the first interview. Through the use of the Review tools in Microsoft Word, the interview questions were located in the transcriptions with New Comment bubbles and interesting quotes were highlighted for use in the write-up. Next, the primary researcher began to disassemble the data to get a sense of the emergent themes. This entailed searching for common responses to questions and key words that respondents used. The fragments were coded so that the data could be grouped in different ways in preparation for reassembling. After reassembling the data, they were interpreted based upon the themes that surfaced. Finally, the primary researcher drew conclusions and selected the ways in which to represent the findings, including detailed descriptions as well as charts, graphs, and tables.\nLimitations and delimitations: This study was limited to the perceptions of superintendents regarding their experiences using social media tools to communicate with stakeholders. It did not examine how employees, students, parents, or community members perceived the received information from the school administrators, nor did it provide an analysis of who read the blogs, followed the tweets, \"liked\" a school's social networking site, listened to the podcasts, or viewed the uploaded videos. Additionally, this study did not examine stakeholders' preferences for receiving school communication via social media tools vs the traditional forms of communication that school employees use. Each of those topics could be the focus of a different independent research study. This study focussed on the unique experiences of school superintendents in order to begin filling the void in the literature on this topic.\nFindings: Six themes emerged from the 12 qualitative interviews with the superintendents and from the internet data regarding their social media use. The six emergent themes were as follows:\nCounterarguments: One of the counterarguments to using social media and digital communication rather than traditional print forms has been the issue of access. It has been feared that stakeholders would be excluded from vital information due to the lack of access to social media tools. The majority of the participants in this study expressed that, due to devices such as smartphones, such concerns are mostly unfounded.\nImplications for school administrators: The advice from the participants in this research study to administrators who have not been using social media tools in their professional role was to \"jump in\" and start using them. The participants acknowledged that there might be fears about not having enough time to learn how to use the tools, not having enough time to devote to two-way communication with stakeholders, or not getting stakeholders to engage with their school's principal or superintendent through these new avenues.\nImplications for theory: SCRM, or CRM 2.0, served as a framework for this qualitative study. Findings from this study indicated that CRM 2.0 frameworks also apply to the field of education. As such, a more inclusive definition of social CRM is proposed:\nConclusion: \"I'm glad you're doing the study,\" declared a superintendent from the northeast as she finished her interview. \"I think it's an important study just to call attention to the topic. Sometimes we can be a bit of a Luddite in these jobs, so it's good to have people spark our thinking.\" That certainly was one of the goals of this research study: to spark some thinking. Unlike the Luddites of two centuries ago, school administrators can ill afford to ban social media or turn their backs on this important communication tool.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to describe, analyze, and interpret the experiences of school superintendents who use multiple social media tools with stakeholders as part of their comprehensive communications practices. Additionally, it examined why superintendents have chosen to communicate with their stakeholders through social media.\nDesign/methodology/approach: Qualitative interviews were conducted with a total of 12 school superintendents; three each from the northeast, south, midwest, and west regions of the USA and Canada.\nFindings: Six themes emerged from the 12 qualitative interviews with the superintendents and from the internet data regarding their social media use. The six emergent themes included: (1) more immediate and more frequent interactions between school superintendents and their stakeholders; (2) greater transparency regarding decision-making and budgeting processes; (3) impact on a school superintendent's personal and professional growth; (4) stronger connections to local stakeholders, to fellow educators, and to the world; (5) use is an expectation: it is no longer optional; and (6) accessing information from the superintendent in a multi-modal way.\nResearch limitations/implications: This study was limited to the perceptions of superintendents who currently use two or more forms of social media to communicate with their stakeholders. The study did not include the thoughts and perceptions of superintendents who are only experimenting with social media use, those who have not been able to engage their stakeholders in two-way conversations, or those who have not yet implemented social media as part of a comprehensive communications approach.\nPractical implications: Implications are provided for school administrators, boards of education, and educational leadership programs.\nOriginality/value: A unique study into the experiences of school superintendents who use multiple social media tools as part of their communications practices with stakeholders.\n", "role": "assistant" } ]
The importance of role sending in the sensemaking of change agent roles
[ "NHS", "Human resource management", "Organizational change", "Communication", "Qualitative research", "Organizational restructuring" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Healthcare reconfiguration is challenging, and the use of champions or change agents has been widely accepted as best practice (Soo et al., 2009; MacFarlane et al., 2011; Chreim et al., 2012). Change agent sensemaking, in particular, has been found to be a key factor in the success of organisational reconfiguration (Hendy and Barlow, 2012; Balogun and Johnson, 2004; Parry, 2003). Sensemaking is the process by which new meaning is created to rationalise the actions of individuals in relation to change (Balogun and Johnson, 2004; Gioia and Chittipeddi, 1991; Maitlis and Sonenshein, 2010; Weick, 1995; Weick et al., 2005). Interpretive schemes are socially constructed from the connecting of cues and frames by organisational groups (Maitlis and Sonenshein, 2010). This process is influenced by existing knowledge (expectations and existing logics) as well as attempts to influence sensemaking from others (i.e. sensegiving) (Gioia and Chittipeddi, 1991; Humphreys et al., 2012). Much of the focus in research has been on the importance of change agent's sensemaking of the change (Gioia and Chittipeddi, 1991; Naslund, 2009; Tucker et al., 2012; Rouleau, 2005) but there has been little focus on the how sensemaking plays a role in the interpretation of the change agent role itself. This sensemaking is important because it will help us to understand how change agent roles should be created, designed or managed. The current lack of understanding about change agent roles is argued to be one of the determining factors in determining the extent to which agents are found to improve change management processes in healthcare organisations (Greenhalgh et al., 2005; Locock et al., 2001). Our research will inform healthcare managers responsible for the management, recruitment and evaluation of change-related roles in how practice can be improved.\nChange agent role complexity and role ambiguity in healthcare: In healthcare research, the role of managers with a clinical background (e.g. nurse managers) has been highlighted as being very important for implementing cultural change (Bolton, 2005). It is suggested that giving nurses managerial roles will empower them as organisers of care and having involvement in managerial activities, such as allocation of resources, encourages initiative, self-reliance and self-development (Du Gay, 1996). Managers with a clinical background act as transformational leaders and agents of change due to their ability to adopt a pragmatic style: interpreting and implementing management policies into an achievable programme within the strict confines of their immediate healthcare setting (Bolton, 2005).\nSensemaking in organisational change: Sensemaking is a process by which a mental model of events is developed which acts as an interpretive scheme for the actions and behaviour of individuals in explaining past behaviour and influencing subsequent situations (Gioia and Chittipeddi, 1991; Gioia et al., 1994; Maitlis, 2005; Maitlis and Sonenshein, 2010; Weick, 1995; Weick et al., 2005). This framework allows organisational members to comprehend events, attribute meaning and predict future events (Weick, 1995; Starbuck and Milliken, 1988). The basis of this framework consists of information which is gathered from the organisational context in many forms; conversations, documents, storytelling, behaviours and actions (Balogun and Johnson, 2004). Sensemaking is a useful theoretical frame to advance our understanding of the creation of change agent roles because it seeks to unpack the process of interpreting stimuli from a cognitive perspective - for example, how change agents come to understand their role expectation and what may lead them to act in a particular way in a given situation (Balogun and Johnson, 2004; Rouleau, 2005; Weick et al., 2005; Evans and Baker, 2012). It takes into account the social nature and collective process which is present in the negotiation of new role expectations.\nMethods: Study context and approach\nFindings: We approached the data with our three research questions in mind. We first wanted to understand the creation of the role and how role sending influenced the role behaviour of change agents. The ambiguity which resulted from a lack of role sending enabled us to study in detail the sensemaking process at work. Below we explain how the change agents collectively made sense of the role in their own way, drawing on the known expectations of others and their own experiences and ideas. And finally, we identify the organisational outcomes which resulted from this freedom in sensemaking.\nDiscussion and conclusions: The aim of this research was to improve our understanding of how change agents make sense of their role in organisational change management. This paper explored the creation of a formal change agent role in a large scale hospital reconfiguration project and found that an absence of role sending by senior managers led to uncertainty about the expected role behaviour. This ambiguity enabled us to study in detail the sensemaking process at work. Our data helps to explain how change agents collectively make sense of a role in their own way, drawing on their own experiences and views as well as cues from other organisational members. We also identified the organisational outcomes which resulted from this freedom in sensemaking. This study demonstrates that by leaving too much flexibility in the definition of the role, agents developed their own sensemaking which was subsequently very difficult to manipulate.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to investigate what happens when a lack of role-sending results in ambiguous change agent roles during a large scale organisational reconfiguration. The authors consider the role of sensemaking in resolving role ambiguity of middle manager change agents and the consequences of this for organisational restructuring.\nDesign/methodology/approach: Data were collected from a case study analysis of significant organisational reconfiguration across a local National Health Service Trust in the UK. Data consists of 82 interviews, complemented by analysis of over 100 documents and field notes from 51 hours of observations collected over five phases covering a three year period before, during and after the reconfiguration. An inductive qualitative analysis revealed the sensemaking processes by which ambiguity in role definition was resolved.\nFindings: The data explains how change agents collectively make sense of a role in their own way, drawing on their own experiences and views as well as cues from other organisational members. The authors also identified the organisational outcomes which resulted from this freedom in sensemaking. This study demonstrates that by leaving too much flexibility in the definition of the role, agents developed their own sensemaking which was subsequently very difficult to manipulate.\nPractical implications: In creating new roles, management first needs to have a realistic vision of the task and roles that their agents will perform, and second, to communicate these expectations to both those responsible for recruiting these roles and to the agents themselves.\nOriginality/value: Much of the focus in sensemaking research has been on the importance of change agents' sensemaking of the change but there has been little focus on how change agents sensemake their own role in the change.\n", "role": "assistant" } ]
Lean manufacturing performance in Indian manufacturing plants
[ "India", "Manufacturing industries", "Lean production", "Lean manufacturing", "Manufacturing", "Manufacturing performance", "Manufacturing operations" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Lean manufacturing, often coined as Toyota production system (TPS) in academic literature, started in Toyota Motor Manufacturing Company after the Second World War when most Japanese organizations including Toyota were confronted with the challenge of managing production facilities with limited resources (Liker, 1998; Pavnaskar et al., 2003). This challenge motivated Toyota managers to develop various elements of TPS aimed at reducing waste. Thus, lean is about producing the same output with lesser resources (men, material, space, and machinery). Today, it has helped Toyota achieve the distinction of being the best car manufacturing company in the world (Stewart and Raman, 2007). Application of lean is not limited to the automotive sector only, but, it has also found acceptance in a wide range of manufacturing industries operating under a unionized or a non-unionized environment in the US (Shah and Ward, 2003) or elsewhere (Cua et al., 2001; Anand and Kodali, 2008), and is being applied in big as well as small organizations (White et al., 1999).This paper is organized as follows: Section 2 provides an overview of the relevant literature and the motivation behind this research. Section 3 describes the research methodology. Section 4 deals with the results of the survey. Section 5 presents a discussion on the research findings and ends with a conclusion.\n2 Literature review: Lean manufacturing is viewed by the scholarly community primarily at three levels. At the first level, which is the philosophical level, lean is about eliminating \"waste\" from the production system (Ohno, 1988; Shingo, 1989; Womack et al., 1990; Womack and Jones, 1996) and yet be able to produce products of the highest quality that satisfies the ultimate customers. As Shingo (1989) aptly remarked, 80 percent of lean is about waste elimination and the balance about system. Waste, often called muda, in Japanese, comprises seven types of common waste: over production, unnecessary motion, excess inventory, excess transportation, rejections/rework, waiting, and over processing (Cachon and Terwiesch, 2009). Apparently, elimination of these wastes looks simple and straightforward, yet their identification is often difficult in most organizations.At the second level, some scholars construed lean as a rule driven system (Spear and Bowen, 1999). Spear and Bowen studied 40 plants over a four year period in the USA, Europe, and Japan - some plants were operating according to lean and some not. They concluded that Toyota uses three rules for designing production system and another rule for systematic problem solving. Rule 1 states that all activities need to be specified in terms of content, sequence, timing, and outcome. Rule 2 suggests that every supplier and customer connection need to be direct and unambiguous. Rule 3 supports direct and simple pathways for every product and service. Rule 4 advocates small improvements done scientifically under the guidance of a teacher at the lowest possible level.At the third level, lean is viewed as congregation of tools and techniques (Shah and Ward, 2003; Pavnaskar et al., 2003; Li et al., 2005; Seth and Gupta, 2005; Hines et al., 1999; Lasa et al., 2008; Basu, 2009) aimed at eliminating waste. Shah and Ward (2003) investigated the application of 22 lean practices and categorized them into four \"bundles\", i.e. just-in-time (JIT), total quality management (TQM), total productive management (TPM), and human resource management (HRM), in a variety of industries in the US, and observed that 23 percent variation in operational performance attributed to the use of lean bundles. They also observed a strong influence of plant size on lean implementation and a less influence on plant age and unionization. Shah and Ward (2007) in a subsequent paper argued that lean is a multi-dimensional construct and developed ten distinct factors/dimensions to characterize lean production system.Thus, the philosophical perspective of lean provides the highest level of abstraction; the guiding rules providing the second level of abstraction, and, the tools and techniques providing the least level of abstraction. The tools and techniques residing at the lowest abstraction level are widely used for lean implementation in the industry (Shah et al., 2008).Investigation of the effect of lean was not just limited to the US. Scholars have tested this popular concept in the developing world as well. Lawrence and Hottenstein (1995) investigated the impact of JIT manufacturing on plant performance (quality, lead time, productivity, and customer service) in 124 plants in Mexico. Cua et al. (2001) studied 163 plants in Germany, Italy, Japan, UK, and USA and observed higher level of manufacturing performances (cost, quality, delivery, and flexibility) when TQM, JIT and TPM were jointly implemented.In the last decade or so, China has become the manufacturing or sourcing hub for many global organizations primarily due to its low manufacturing costs. China's resurgence as a global economic power started in the late twentieth century and has established itself as a fast growing economy in the world (Das, 2006). The focus on efficient production to compete in the global market has triggered many Chinese firms to implement lean. In fact, Taj (2007) reported that lean was introduced in first automotive works (FAW), the first Chinese automotive plant, in 1977, even before its adoption in USA and Europe. Taj observed the application of lean manufacturing in 65 manufacturing plants in China, encompassing a variety of industries (electronics, telecommunication, wireless, food/beverage, garment, pharmaceutical, etc.) and found significant benefits achieved in material flow, on-time delivery of finished goods, defect rate, and scheduling. In a subsequent paper, Taj and Morosan (2011) reported that lean operations practice (human resources and supply chains) and product system design had a significant and positive impact on three performance dimensions - flow, flexibility, and quality.In comparison to China's stellar economic performance, India's post 1991 economic performance showed discernible improvement. The average GDP growth rate between 1992 and 2002 was approximately 6 percent, which put India among the fast growing developing countries in the 1990s (Ahluwalia, 2002). India's GDP in 2010 stood at $1.3 trillion with 8.5 percent growth rate. Eichengreen and Gupta (2010) reported that share of industry in GDP grew rapidly between 1950 and 1965, modestly from the mid 1960s until early 1990s. Since post liberalization in 1991, the industry's share stagnated at 27 percent. On the contrary, service sector's share in GDP grew from 30 percent in 1950 to 55 percent in 2007-2008. After liberalization, many foreign players in the automotive, electronics and other sectors opened manufacturing facilities in India with better operating systems, which exposed Indian manufacturing firms to fierce competition. Witnessing intense competition, Indian firms did start implementing lean in the early 1990s, but the adoption process gained momentum very slowly due to various reasons.Thus, the academic literature, reports application of lean manufacturing in many countries across the world, yet little is known about its degree of adoption in India. In fact, many organizations in India have started implementing it recently (Vinodh et al., 2010). As a consequence, little is known about the current status of lean implementation in India except some case studies in select companies (Dhandapani et al., 2004; Seth and Gupta, 2005; Kumar et al., 2006; Singh and Sharma, 2009; Vinodh et al., 2010). Upadhye et al. (2009) empirical work is a notable exception. Upadhye and others sought the opinion of executives from 71 different organizations in India on different attributes of lean manufacturing system, quality, inventory and maintenance. Their research indicated that Indian executives are cognizant of the importance of lean on a firm's performance and big organizations are predominantly implementing lean. In a subsequent paper, Upadhye et al. (2010) presented a model for the implementation of lean. This paper, therefore, makes an attempt to report the degree of lean implementation in Indian manufacturing plants primarily from the manufacturing sector. It also investigates what set of lean management practices lead to improvement in operational practices - productivity, lead time, first-pass correct output and the like.\n3 Research methodology: Since lean manufacturing is a fairly new concept in India, the intent behind the survey was to measure the current state of lean implementation in Indian manufacturing plants located in the four regions (North, South, East, and West). Stratified sampling was used for the survey, as automotive companies (automobiles and auto ancillaries), the biggest users of lean system, are clustered in the Western, Southern, and Northern regions of India. Additionally, simple random sampling was used for manufacturing companies spanning a wide variety of sectors (e.g. metal products, metal processing and power equipment). The survey was carried out between May and September 2010. The total sample size used for the study was 79. Table I provides the distribution in percent of the 79 plants classified by industry.3.1 Measurement scale\n4 Results: 4.1 Descriptive statistics\n5 Discussion and conclusion: The academic and the industrial community in the Western world have long acknowledged lean manufacturing or TPS as the gold standard for operational management. Its success has been so stunning that its adoption has not remained restricted to manufacturing, but, has propagated to the non-manufacturing arena as well. Yet, its adoption in India in the industrial sector has begun very recently. This study makes an attempt to find the degree of lean implementation in India and the different lean manufacturing practices that have been embraced in automotive and other manufacturing plants in India. The study suggests that many Indian plants are at an advanced level of lean implementation and have achieved superior operational performance by implementing lean. The findings of this study in terms of improvement in operational metrics seem to be in agreement with the findings of other scholars (White et al., 1999; Detty and Yingling, 2000; Cua et al., 2001; Shah and Ward, 2003). Out of the six operations metrics considered for the study, first-pass correct output, reduction in manufacturing lead time and productivity increase are found to be the three key drivers for lean implementation. However, as the sample size for this study was not large enough, generalizations from this study to the population need to be made with caution.\n", "role": "user" }, { "content": "Purpose: - The purpose of this paper is to examine the current state of lean adoption in Indian manufacturing plants and its impact on operational performance.\nDesign/methodology/approach: - A survey questionnaire was developed and adapted based on work done in US industries. The survey questionnaire was sent to 400 firms in four geographic regions in India. In total, 79 usable responses were received for the study.\nFindings: - Since lean manufacturing is a multi-dimensional construct, the results demonstrate that approximately 80 percent of the respondents have implemented many dimensions of lean - focus on customer needs, pull system, setup time reduction, total productive maintenance, supplier performance, statistical process control, and cross-departmental problem solving. The operational metrics have improved on all accounts: high productivity, reduced lead time, improved first-pass correct output, reduced inventory and space requirement. Interestingly, respondents have indicated that first-pass correct output, reduced manufacturing lead time, and increased productivity are the three main drivers of lean implementation.\nResearch limitations/implications: - Since the sample size is not very large, the results need to be considered with caution.\nOriginality/value: - Lean manufacturing is a very popular concept in the developed world and in some countries in the developing world as well. Yet, little is known about its current status in India, except for a few case studies. The study investigates the degree of lean production implementation in Indian manufacturing plants and its impact on operational metrics. The study also indicates the relationship between lean dimensions/practices and operational outcomes.\n", "role": "assistant" } ]
Exploring the relationship between financial inclusion and assets accumulation in South Africa
[ "Social welfare", "Poverty", "Quantile regression", "Financial inclusion and poverty", "Assets building", "Ordinary least square" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Recently asset building and financial inclusion have been used as a strategic intervention to achieve multiple outcomes: consumption smoothing and capacity development of the poor through asset accumulation. This innovative approach to social policy is gaining support as scholars argue that income transfer alone is not a sufficient condition to assist the poor out of poverty (Sherraden and Boshara, 2007, p. 3). Elsewhere in the literature, evidence suggests that financial inclusion encourages individuals to accumulate assets (Demirguc-Kunt and Klapper, 2012). Consequently, since assets generate income, there is a growing consensus that savings and asset-building opportunities should be extended to everyone, both the poor and rich for two reasons. First, from a moral point of view, everyone should have a fair chance and opportunity to develop his or her capacity to lead a stable and satisfactory life. Practically, everyone will be better off if individuals, communities and societies accumulate assets, invest in themselves and their children and become productive and engaged in meaningful economic activities (Sherraden and Boshara, 2007, p. 3). Hence, individuals and families need to build savings and assets which they can invest to improve their economic well-being over time.\n2. Conceptual definition of asset building: Sherraden (1991) proposed the term assets building as a strategy to assist individuals at the bottom of the income pyramid to build assets. Sherraden (1991) advocates asset building for the low-income families to complement the government income transfer assistance to enable the poor to break out of poverty. He argued that the government should subsidise the poor and low-income families in the same way it supports working class to enable them save for children's education, down payment of house, pension and retirement savings. Since the 1990s, asset building has expanded and now means physical or man-made capital, financial capital, human capital, social capital and natural resources capital (Moser, 2005). During a three-day symposium organised by the Global Assets Project in June 2007, asset building was consistently used to refer to efforts to build savings and productive assets, including education, homeownership, livestock, investment and pensions. Elsewhere it refers to asset accumulation, asset development and wealth building (Sherraden and Boshara, 2007, p. 2). For this study, asset building will refer to man-made capital such as television sets, cell phones, microwaves, fridges and many others.\n3. Theoretical perspective of financial inclusion and assets building: The theory of asset-based social welfare is based on the argument that income transfer alone is not a sufficient condition to enable the poor break out of poverty. Borrowing heavily from psychology, sociology and economics, Sherraden (1991) argues that assets provide a variety of social, economic and psychological effects that income alone cannot provide. Sherraden (1991, p. 148) postulates that assets improve economic stability, connect people with a viable and hopeful future, stimulate development of human and other capital, enable people to focus and to specialise, provide a foundation to take risk, yield personal, social and political dividends and enhances the welfare of offspring.\n4. Empirical review: Assets-building social welfare programmes are growing fast in advanced economies - the USA, the UK, Canada, Korea and Singapore (Li and Sherraden, 2010) but are still at the early stage in Africa. Nevertheless, evidence from developing economies and Africa is gradually growing (See Han and Hong, 2011; Kim et al., 2010; Deng et al., 2014; Austrian and Muthengi, 2014; Chowa et al., 2010; Chowa and Masa, 2015; Curley et al., 2016; Ssewamala and Ismayilova, 2009).\n5. Empirical strategy and data: 5.1 Asset index\n6. Results: 6.1 Descriptive analysis\n7. Conclusion: This study examined the relationship between financial inclusion and individual assets ownership in South Africa. Financial inclusion is captured by monthly savings and insurance whereas asset ownership is measured by a composite assets index derived using MCA from indicators of individual asset possession. The intention is to test the feasibility of asset-building social welfare in South Africa.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to explore the feasibility of asset-building social welfare in South Africa using the FinScope (2014) consumer survey data set. This is achieved using quantile regression technique to examine how financial inclusion influences asset ownership of individuals at the bottom of the assets distribution.\nDesign/methodology/approach: This paper test the feasibility of asset-building social policy for poor families in South Africa by examining the relationship between financial inclusion and asset ownership using FinScope 2014 consumer survey for South Africa. Financial inclusion is captured by monthly savings and insurance whereas asset ownership is measured by a composite assets index derived using multiple correspondence analyses from indicators of individual asset possession. Quantile regressions are used to examine how financial inclusion influences asset ownership of individuals at the bottom of the assets distribution.\nFindings: Evidence from mean and quantile regressions showed that the relationship between financial inclusion and asset ownership is positive and statistically significant at 1 per cent level across the entire assets distribution. However, across the distribution, the change in asset ownership varies: higher at the lower tail (10th) quantile, lower at the median (50th) quantile and higher at the upper tail from the 60th quantile. Thus, the poor and low-income families, some of whom may be gaining formal access for the first time, may derive more satisfaction than frequent users such as the working class.\nResearch limitations/implications: This evidence provides a good case for progressive asset-building social welfare programmes for the poor and low-income families in South Africa. With 11.9 million children currently receiving child support grants, the puzzle is whether income transfer alone can assist these children to break out of poverty. The results should be interpreted as association as the analysis is based on cross-sectional data.\nPractical implications: The implications of this study are that social welfare in South Africa needs to extend beyond transfer and invest in capacity development of the poor. Asset-building social policy that combines income transfer and asset building such as child development/saving accounts will help to provide a sustainable pathway out of poverty.\nSocial implications: Financial inclusion and asset-building social welfare is a crucial issue as it has the potential to improve welfare of the poor. That is, it acts as a complementary strategy to the income transfer approach to poverty alleviation by enabling the poor to find a sustainable pathway out of poverty by building assets.\nOriginality/value: Financial inclusion and asset building is a rare area of research particularly in South Africa. This study therefore is timely and its findings will be handy for policy makers in South Africa. Furthermore, the findings will stimulate future research and debates on how financial inclusion and asset-building social welfare can be used to close the gap between the rich and the poor in South Africa.\n", "role": "assistant" } ]
Assessing the current state of information security policies in academic organizations
[ "Security policies", "Qualitative analysis", "Mixed methods", "Policy analysis", "Academic institutions" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Data breaches in the corporate sector have been on the rise for several years with companies such as Yahoo (Fahey and Wells, 2016), Target (McGrath, 2014), MySpace (Perez, 2016) and many more suffering highly publicized and damaging data breaches. However, corporate institutions are not the only ones affected by this rise in security incidents. Though discussed substantially less by mainstream news sources, universities and colleges have become an increasingly popular target for cyber-attackers. Notable university data breaches include the University of Maryland (Svitek and Anderson, 2014), ND University (Greenberg, 2014), UC Berkeley (Gilmore, 2015), MI State University (Simon, 2016), Stanford University (Hayward, 2013), among many others. According to the Privacy Rights Clearinghouse (2017), 789 academic institutions in the USA have suffered data breaches since 2005. To put this in perspective, an estimated 5,436 data breaches are estimated to have occurred across all domains (academic, corporate, government, etc.) since 2005. That is, academic data breaches have represented 14.5 per cent of reported data breaches since 2005; a non-trivial percentage.\n2. Related work: As literature in the space of security policy analysis is surprisingly sparse, we will ground readers by first providing some background on policy analysis and organizational policy design, prior to transitioning into a discussion of the general impact that strong policy has within organizations and our strategy to address the problem space of security policy.\n3. Methodology: 3.1 Policy selection\n4. Results: 4.1 University information security policies\n5. Discussion: With this work, we collect and analyze in-practice information security policies by using a sizable sample of universities in the USA. We position this paper as an initial work in this space, which seeks to inspire additional needed research in this area. The research questions of this work focused on what content is included within information security policies at universities, as well as how certain factors within these policies may or may not relate to each other. The simple answer is that information security policies across organizations are anything but standardized, with a wide variety of covered features included (or not included) in each.\n6. Conclusion: In this paper, we collected and analyzed a corpus of 90 distinct information security policies from universities within the USA. We find that the prevalence of these high-level policies across the sampled population surprisingly is low, with a 54 per cent occurrence rate. For those universities that maintain these policies, we find that the high-level policy types are evenly split between what would be traditionally-called enterprise information security policies, which focus more on organizational details, and issue-specific information security policies, which may include not only organizational details but also technical ones. Future work should determine, which of these strategies is more appropriate. In an analysis of these policies, we find that they are generally difficult to read, in particular, if the policies are short in length. Despite the brevity of a number of policies, we also found that a vast majority of these documents shared no similar content to their policy peers indicating a lack of reputable source material for which to base these policies on. This finding brings into question the relevance of many of the policy-writing textbooks or standards that are less effective in practice within many organizations. As a final takeaway of our work, we argue that, more generally, there have been few successful attempts to study information security policies on a wide scale. Further work in this space is necessary to understand how these documents are designed, implemented, enforced to ensure that universities and other organizations are more effectively protected from rising external security threats.\n", "role": "user" }, { "content": "Purpose: Colleges and universities across the USA have seen data breaches and intellectual property theft rise at a heightened rate over the past several years. An integral step in the first line of defense against various forms of attacks are (written) security policies designed to prescribe the construction and function of a technical system, while simultaneously guiding the actions of individuals operating within said system. Unfortunately, policy analysis is an insufficiently discussed topic in many academic communities with very little research being conducted in this space.\nDesign/methodology/approach: This work aims to assess the current state of information security policies by analyzing in-use policies from 200 universities and colleges in the USA with the goal of identifying important features and general attributes of these documents. The authors accomplish this through a series of analyzes designed to examine the language and construction of these policies.\nFindings: To summarize high-level results, the authors found that only 54 per cent of the top 200 universities had publicly accessible information security policies, and the policies that were examined lacked consistency with little shared source material. The authors also found that the tonal makeup of these policies lacked a great deal of emotion, but contained a high amount of tentative or ambiguous language leading toward policies that could be viewed as \"unclear.\"\nOriginality/value: This work is an extension of a paper that was presented at ECIS 2018. The authors have added additional analyzes including a cross-policy content and tonal analysis to strengthen the findings and implications of this work for the wider research audience.\n", "role": "assistant" } ]
Could formalized State Business Relations lead to economic growth? Business participation, government representation and institutions
[ "Institutions and institutional analysis", "Economic growth", "Formal State Business Relations", "O43", "D73", "D85", "P48" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Much in the literature has been said on state and business failures. Works such as Hausmann et al. (2008), Hausmann and Rodrik (2003), Leftwich and te Velde (2010), te Velde (2009) and United Nations Industrial Development Organization (UNIDO) (2000) highlighted both sides' shortcomings. The collaboration between both sides in the context of State Business Relations (SBR) was suggested as bearing the seeds of success, leading to positive economic and developmental outcomes. This optimistic view was often contested, however, by concerns on rent seeking and cronyism and their possible impact on long-run economic performance. There was also a major distinction brought often to attention demarcating a line between informal State Business Relations (I-SBR) and formal State Business Relations (F-SBR).\n2. Theoretical discussion: 2.1 Formal SBR and growth\n3. Hypothesis: Trying to explore which institutional variables are responsible for helping F-SBR in bringing more investment and economic growth, in this section I propose six hypotheses:\n4. Empirical testing: 4.1 The regression model\n5. Results and discussions: The results of the conducted regression are listed in Table IV. Other than the control variables, a number of variables are significant, whether individually or in interaction terms or both. The following lines will focus on F-SBR and its interactions with various institutions in the three regressions, reporting only statistically significant results.\n6. Discussion and conclusion: The results of the conducted regression provide evidence for many suggestions presented in the theoretical part of this paper, but they also contradict some others. In many instances, contradicting results with regard to the effect of the various included independent variables were witnessed. This makes it important to start with differentiating what the results of each regression mean.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to investigate whether formal State Business Relations (F-SBR) lead to economic growth and which institutional variables are responsible for their success in achieving this.\nDesign/methodology/approach: Following a theoretical analysis, several hypotheses are presented, and a number of different regression models are being used to test the hypotheses.\nFindings: The paper finds that F-SBR with high government representation and business participation are likely to increase investment flows as well as help overcome some burdensome institutional settings. This could be done thanks to building trust, enhancing information exchange and better policy formulation. Despite of its suggested trust-building capacity, F-SBR seem to fail to function in cultures where trust in governments is low, since trust is a key issue for cooperation as that expected in F-SBR. On the other hand, collective but also non-hierarchical cultures that foster cooperation seem to help in the success of F-SBR. Accountability to the public and openness to various societal groups as well as good regulatory quality would help F-SBR in leading to better economic outcomes.\nPractical implications: This paper, thus, suggests supporting the establishment of F-SBR with high official representation and business participation because of their positive economic value. It also suggests that voice and accountability and regulatory quality are important institutions for the realization of better economic outcomes from F-SBR.\nOriginality/value: The theoretical and empirical approaches used in this paper are what make it a novel endeavor in this field. Much of the suggested hypotheses were hardly discussed thoroughly elsewhere in the literature on F-SBR. Moreover, the proxy used for F-SBR, constructed by the interaction term of government official representation and the level of business representation, made it possible to conduct empirical testing on a global level, something barely done in works on F-SBR experiences worldwide. The obtained results supported some of the arguments uttered in the literature while raising questions on others, opening the field for future research on the matter.\n", "role": "assistant" } ]
What happiness at work is and how to use it
[ "Performance", "Productivity", "“Happiness at work”", "Mindset", "Potential" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Happiness at work. That short phrase can sound pretty flaky taken in isolation. But research and scholarship are showing time and again (Boehm and Lyubomirsky, 2008) just what an important subject this for driving top line growth and bottom line results. And how deeply underestimated it has been (Fisher, 2010).\nWhy happiness at work matters?: The iOpener Institute based in Oxford UK has run an extensive, rigorous and robust research program since 2005 into what everyone is increasingly called the science of happiness at work. And our empirical data has found that a happy worker is a high-performing one (Pryce-Jones and Lutterbie, 2010). The happiest employees:\nWhat is happiness at work?: So how do we define happiness at work? Our definition is that \"it is a mindset which enables action to maximize performance and achieve potential\", so it is a simple yet practical definition which applies to an individual, a team or an organization. But the key focus here is this mindset: everyone knows that approaching something in a positive manner is more likely to get results than doing the opposite. It is common sense. Yet common sense seems not to be so common when it comes to organizations.\nWhat is the model we use?: For us happiness at work is based on a researched-based (Lutterbie and Pryce-Jones, 2013) and practitioner-driven model. The performance-happiness model is made up of five important components, which we know as the 5Cs. They are:\nHow is this different from engagement?: While we acknowledge that there is some over-lap between both the science of happiness at work and engagement (whatever definition you use), we can see from our data that there are some important differences. For example senior leaders and general managers frequently report that they are highly engaged at work but they are not happy. We hypothesize that this is because they often also show high levels of conscientiousness (McCrae and John, 1992) (NEO-PI-3) and therefore invest heavily in their work regardless of mindset.\nHow to use the approach? A case study: An organization came to us with a big strategic problem about 18 months ago. They were having trouble retaining business critical employees and this was having a devastating effect on their ability to grow. They simply could not take on more client work and were in danger of over-trading. Internally there were problems scoping projects, meetings milestones and delivering quality outcomes for their clients. The business was simply unable to expand or grow because they were losing talent fast. That meant that every team was pretty much in permanent crisis, so our goal was to help them improve this turnover number.\nWhat issues did we face during implementation?: During this time the organization went through significant change in terms of ownership with all the knock-on consequences in terms of delivering a large-scale intervention. In addition there were, as always, a few leaders who are not ready for embracing new ways of working or leading others. Senior leaders who are not prepared to get behind an initiative create risk which needs managing and addressing because employees always take their cue from the top.\nWhat was key to overall project success?: Access to, support from and the attitude of the CEO was a critical success factor. This CEO had an intrinsic and unshakeable belief that the approach was absolutely the right thing for this knowledge-based professional services organization. And we see that in our data too: knowledge-workers who unhappy at work create much more instability and risk because they can so easily jeopardize projects they are involved with.\nSo what were the outcomes?: When the project started, turnover of business critical employees was running at 25 percent. Over 15 months, this halved to 12.5 percent. Not only has this reduction created much more stability and a platform for growth, but of course the recruitment costs to the organization have fallen dramatically. What matters more is the intangible effect on the organization's social networks. Real-time relationships and therefore trust within and between teams has increased greatly because there is a much greater sense of stability and progress.\n", "role": "user" }, { "content": "Purpose: Change management, organizational, team and leadership development is often conducted through frameworks that do not meet expectations. This paper aims to look at how the science of happiness at work delivers strategic outcomes when used in a bottom-up and top-down approach.\nDesign/methodology/approach: While the idea that happiness is an important concept outside work is now prevalent, there is little research or practice to show how it can be used to drive organizational success or deliver return on investment. The case study shows how to deploy the approach in an organization-wide strategic intervention.\nFindings: The science of happiness at work delivers return on investment and strategic outcomes when properly implemented.\nPractical implications: This new approach is a powerful methodology which offers traction at individual, team and organizational level: it opens up a different evaluation and development methodology that positively fast-tracks change and growth.\nSocial implications: There is huge potential for making teams and organizations better places to work by using a simple, practical and aligned framework rooted in something that matters intuitively to everyone.\nOriginality/value: This paper provides an understanding of the theory and application of the science of happiness at work. It shows how strategic issues can be addressed within a short space of time and highlights the benefit of an approach which encompasses personal, social and organizational issues simultaneously.\n", "role": "assistant" } ]
A perspective on gender in management: The need for strategic cross-cultural scholarship on women in management and leadership
[ "Gender", "Leadership", "Management", "Women", "Equality", "Cross-cultural" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Midway through the second decade of the twenty-first century, gender-based challenges continue to be a topic of great interest around the globe. Joshi et al. (2015) recently reported that \"gender equality appears to be at the forefront of the global humanitarian agenda\" (p. 1459), as they highlighted progress being made for women, particularly in some developing countries. It is also clear that women in many developed and in some developing countries are completing undergraduate degrees in higher numbers than males, and that the percentage is nearing equality in graduate degrees in many areas (Burke and Vinnicombe, 2013). In addition, we know that women globally are entering the workforce in higher numbers than even a decade ago, and that workforce gender diversity is increasing in regions and countries around the world (Ali et al., 2011). Adler (2015) also highlighted the global progress made in recent decades in terms of the number of women in top positions in government, business, and education. Yet, Joshi et al. called this progress \"both promising and problematic\" (p. 1459), as these and other researchers are quick to point out that there are ongoing, persisting challenges with efforts to strengthen women's impact around the world.\nLiterature review: Over a decade and a half ago, a few articles in a special issue of Cross Cultural Management (now re-titled as Cross Cultural and Strategic Management), summarized the areas of research that were surfacing in cross-cultural research on women in management. First, Burke (2001) noted that the current body of research focused on country comparisons, non-North American countries, broad conceptual pieces, and edited collections for the following themes: career patterns, career strategies, career priorities, gender earning differences, sex role stereotyping, media and advertising, gender and power, work values, managerial styles, work-family conflict, career priorities, occupational sex segregation, intergenerational correlations, female entrepreneurs, women's experiences, employment patterns, education, adjustment programs, cultural dimensions, labor markets, representation of women, expatriates, social economic changes, life stress, and women on corporate boards. Second, Omar and Davidson (2001) summarized the key differences across cultural and national boundaries in four categories of influencing elements: cultural (noble and aristocratic heritage; centrality of family, marriage, and motherhood; and influence of religion - particularly Confucianism and Islamic beliefs), access to education, political and social policies (equal opportunity legislation, new population policy, and family policy), and organizational processes (geographical mobility, forced retirement, and military connection).\nResearch methods: To capture the most recent scholarship in the area of gender and cross-cultural management, we conducted a content analysis of peer-reviewed articles published in 15 journals between January 1, 2010 and March 1, 2016. We selected the five-year time frame to get the most current state and scope of published literature and to see what gaps may exist for both future research and conceptual/theoretical progress. Nine of the 15 were recommended by cross-cultural management experts and were ranked among top international and cross-cultural management journals (i.e. Cross Cultural & Strategic Management, Global Strategy Journal, International Journal of Cross Cultural Management, International Journal of Human Resource Management (IJHRM), Journal of International Business Studies, Journal of International Management, Journal of World Business, Management International Review, and Thunderbird International Business Review), while six are widely accepted as influential periodicals more generally in the field of management (i.e. Academy of Management Journal, Academy of Management Perspectives, Academy of Management Review, Administrative Science Quarterly, Journal of Applied Psychology, and Journal of Management), even though they too tout an international presence and readership.\nFindings and discussion: Six research questions (RQ) were used to explore the current trends in cross-cultural and international management scholarship on women/gender. The first section below - Journals and articles - reports descriptive statistics related to articles published in the 15 journals, as well as other details (RQ1). The second section presents statistics related to the Gender of authors (RQ2). Countries and regions data appear in the third section (RQ3), while findings on the core constructs (i.e. topics) explored in articles are reported in section four (Core constructs, RQ4). Answers to RQ5 are found in section five (Tone of study). The final section, Theoretical frameworks, provides data to answer RQ6. A discussion has been integrated within this result section as well.\nConclusion: This study shows that, while some relevant research is being conducted, much more theorizing and research are still needed to figure out what accounts for the slight progress that has been made thus far, why women still have less access to management and leadership positions, and what needs to be done to change this and by whom. This includes continued research within all types of organizations (e.g. corporate, nonprofit, government, community, and political) regarding their policies, practices, processes, and systems.\n", "role": "user" }, { "content": "Purpose: There is still a lack of understanding why there is little progress when it comes to women seeking and obtaining top management and leadership positions in organizations today, and this is particularly true within the cross-cultural and international management and leadership contexts. One step forward, however, is to understand current work and trends in research and theory to identify these gaps. Hence, the purpose of this paper is to provide an overview of the most recently published literature on the role of gender in management teams within and across cultures.\nDesign/methodology/approach: This content analysis has examined the most recent literature (i.e. January 1, 2010 to March 1, 2016) in 15 influential academic journals within the cross-cultural and international management field. The study has analyzed 152 primary and 85 secondary articles that met the strict criteria of the study.\nFindings: Results include findings on journals/articles, gender of authors, countries included in data collection, constructs measured, tone of manuscripts (i.e. adverse outcomes associated with gender compared to the neutral/mixed or positive effects), and the theoretical frameworks utilized in the articles.\nResearch limitations/implications: This analysis will be useful for researchers, theorists, and practitioners in understanding the current knowledge base and in discovering the emerging gaps and needs.\nOriginality/value: This is the first study of its kind within gender and cross-cultural/international management. The findings clearly show gaps in research and theory that will help guide future work.\n", "role": "assistant" } ]
Agricultural structure and farmers' interconnections with rural communities
[ "Farmers", "Agricultural structure", "Rural communities", "Social connectivity" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Processes related to rural development are tied to historical developments. At the local level of single farm households, however, existing interrelations and networks can be seen as representations of the connectivity of farmers to their local communities (Van der Ploeg and Roep, 2003). Together with reciprocity and trust, social connectivity is significant for the economic sphere of a business (Sage, 2003). Thus, social connectivity can be seen as one important element on the path toward sustainable development in rural areas.\nBackground and theory: Farms and farmers\nHypotheses: Two sets of hypotheses emerge from the literature reviewed above. The first one builds on the effects of scale and type of farming on community well-being and on the social functions of farming. Drawing on Mann (2004) and Land and Willisch (2002), we hypothesize that:\nMethods: The survey\nResults: The regression results for both study regions are shown in Table III. In order to test for significant influences of structural variables on the connectivity and attachment variables in both study regions, we re-coded the region dummy variable and re-ran the models so that for both samples the significance of the main terms could be assessed (see the top part of Table III). The bottom part of Table III (interactions) shows the results focused on the Swiss sample (region dummy: Swiss farmers =1). Therefore, significant interaction terms indicate general differences in the effect of the structural variable on the social connectivity indicators and the effect's direction when comparing the Swiss with the NE German sample. All models additionally were controlled for covariates on farmers' personal data (i.e. age, gender, education, and children/no children).\nConclusion: With respect to our hypotheses and especially to the Goldschmidt debate, we found diverging results. On the one hand, our main explanatory variable representing scale of farming, namely farm size, had negative influences on the local social networks. On the other hand, the number of employees, our second scale variable, positively affected farmers' networks and farmers' attachment to the community. Especially for NE German farmers social bonding with employees seems to be common. Also, we could show that a farmer's personal connection with the farm is in both agricultural systems more relevant than the actual legal form of the farm. Although not too surprising, this result stresses the significance of locally rooted farm businesses for social cohesion in rural areas. Other findings of this study highlight the importance of off-farm work for farmers in NE Germany. For the Swiss context, there seems to be no need for off-farm work as an agent for social interaction. Farm diversification was equally positively associated with social network and involvement indicators in both study regions. Additionally, we saw that organic farmers in NE Germany seem to be less integrated within the local community than conventional farmers. This might be a result of a higher polarization in the NE German agricultural sector between extensive and intensive farming systems. In the Swiss system, where small-scale farming is the norm, such a polarization is not given and organic and conventional farming are ideologically not that far apart from each other.\n", "role": "user" }, { "content": "Purpose: In the context of rural development, the question how farmers are interconnected with local rural communities is crucial, as farmers historically have played a key role in rural areas, always shaped by the cultural-systemic context in which they acted. The purpose of this paper is to explore this connection in North-East (NE) Germany and Switzerland, two countries whose agricultural systems can be seen as diametrically opposed to each other with respect to their structure.\nDesign/methodology/approach: The authors conducted a survey on NE German and Swiss farmers to assess the connectivity between farms and rural communities in terms of the farm managers' perceptions of their social networks, social support, sense of belonging, and active involvement in organizations.\nFindings: The results show commonalities and differences between both study regions. Smaller farms are characterized by strongly locally based networks and a higher sense of community belonging, whereas larger farms rather have networks with strong ties outside the local dimension. Moreover, farmers' local origin and farm diversification are positively associated with strengthening the connection between farms and local communities. Off-farm work is a means for this connection only in NE Germany.\nOriginality/value: This paper contributes to the discussion about adverse effects of farming scale and corporate farming on community well-being by simultaneously delivering insights into two structurally different agricultural systems. At the same time the approach allows for a comparison between the systems.\n", "role": "assistant" } ]
The perceived importance of core soft skills between retailing and tourism management students, faculty and businesses
[ "Training", "Soft skills", "Education" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: The twenty-first century workforce has experienced tremendous fluctuation due to the changing workplace environment, increased education of employees, and advances in technology; consequently, the old way of preparing students for success in the workplace may be effective but not as efficient as needed (Redmann and Kotrlik, 2004). As stated in the NRC Report (National Research Council of the National Academies, 2009), the largest problem employers facing today is whether or not there is an ample supply of qualified individuals to work in current work world. The National Business Education Association states the shortage of skills confronting today's dynamic workforce goes beyond academic deficiencies and lack of hands-on occupational skills (Mitchell et al., 2010). Historically, technical skills, also known as hard skills, were the only skills necessary for career employment; but in today's work environment, technical skills are no longer enough to keep people employable (Robles, 2012). The result is that when a workforce has proficient technical skills but has an absence of soft skills, a soft skills gap emerges (Slade, 2014). Soft skills are defined as a combination of personal qualities, interpersonal skills, and additional skills/knowledge (Pandey and Pandey, 2015) that help an employee better perform their job. Soft skills are also characterized as a cluster of personality traits, social graces, and facility with language, personal habits, friendliness, and optimism that individuals might possess to varying degrees (Pandey and Pandey, 2015).\nLiterature review: Core soft skills\nMethodology: Conceptual framework\nResults: Demographic data\nDiscussion: The future of many workplaces is dependent on the education and professional qualifications of college graduates. As the work world continues to change, the qualifications of needed workers also changes. Because of these constant shifts, it is increasingly important that college students are prepared to start entry-level jobs with the right skills to be successful. The retail and tourism industry provides consumers with products and services that accommodate their pleasures, leisure's and business activities. That said, the workforce in this industry needs employees with people focused skills to create a positive consumer experience. To create this type of workforce in this industry, faculty, and industry leaders should work together to make sure RTM students have not only the necessary technical skills but also have the recognized soft skills to be competitive. This study examined and compared the perceived importance of soft skills by students, faculty, and industry leaders in retailing and tourism management. Variations between and within each identified soft skill were presented. Among the seven core soft skills (experience, team, communication, leadership, decision making/problem solving, self-management, and professional), all three groups reported communication as the most important soft skill needed with the leadership skill being the least important to faculty and industry leaders. Experience was the skill listed as the least important to the student sample.\nImplications: The need for employees that can meet the challenges of the consumer-centered retailing and tourism management industry is growing rapidly. In order to prepare students to better meet the needs of consumers in these industries, there should be some consistency among students, faculty, and industry leaders as to what students require to be competent in the entry-level career positions. This study found a difference in the rank order of the importance of soft skills among these three groups. The question now is how can this importance gap be narrowed so that students are more prepared to be successful employees? The challenge for faculty and industry leaders is to develop a robust, high-quality, and long-term relationship that enables appropriate curriculum development to occur. This relationship should seek to pair pedagogy with experience to create advantageous opportunities of success for students. Both industry leaders and faculty should meet regularly for discussions and exchange of ideas on matters related to the preparedness of graduates for the workforce. This process may involve collaborative research to uncover appropriate soft skill focus. Clearly, the need for further research is evident from the findings with the importance gap between the three groups existing. Beginning with Crawford et al. (2011), there are indications of variations in the types of soft skills by various sample groups. From these findings, it is evident that different industries must try to understand what soft skills are required by entry-level employees that are specific to their industry. Working together on research projects can help uncover these specific needs.\nRecommendations for future research: The purpose of this study was to determine how students ranked the importance of soft skills and compare their rankings to retailing and tourism management faculty and businesses rankings. The findings suggest the importance of some soft skills differ between students, faculty, and businesses. Because of the variation in sample size further research is needed to clarify these differences. As researchers and faculty, we need to further examine the priorities of the students compared to what faculty are requiring of them, and contrast that with what industry leaders expect of them when students take entry-level positions in particular. More research is needed to understand why each group has a different idea about which softs skills are the most important and which of the soft skills are the least important. A qualitative research project would allow individuals to actually express specifically what they feel in regards to the importance of each of the soft skills discussed. Furthermore future research would benefit from studies which had larger sample sizes from more diverse populations as well.\n", "role": "user" }, { "content": "Purpose: Soft skills which are a combination of personal qualities and interpersonal skills that help an employer perform their job are an increasingly important concern to businesses and academia, the purpose of this paper is to determine how students ranked the importance of soft skills and compare their rankings to retailing and tourism management faculty and businesses.\nDesign/methodology/approach: A cross-sectional survey of students, faculty and industry leaders was conducted using an existing survey instrument validated by Crawford et al. (2011). Faculty who were members of retailing and tourism management professional organizations were solicited to participate in the study. Retailing and tourism management students from the researchers' university were sent a link to complete the survey. All participants were asked to rank the order of importance of the soft skills and their characteristics.\nFindings: Variations in the importance of soft skills were reported between the three groups. Variations in the importance of the soft skills characteristics were also identified between the students, faculty, and industry leaders. While communication was identified as the most important soft skill by all three sample groups, experiences was the least important for students and leadership was the least important for faculty and industry leaders.\nResearch limitations/implications: A limitation of the study was the variation in the sample sizes between the student, faculty, and industry sample. The strength of this study lies in the ability to provide evidence for the need to compare soft skills research results for retailing and tourism management students. Soft skills are found to be important to all three groups, but differences indicate faculty and industry need to work together to clarify exactly what soft skills students need to successfully compete for employment in the retailing and tourism management field.\nOriginality/value: As the work world continues to change, employers seek workers who have soft skills that support their knowledge base. While technical skills are a current part of educational curricula, soft skills need to be emphasized at the university level so that students gain expertise that prepare them to be successful in this changing workplace.\n", "role": "assistant" } ]
Strategic IT management: how companies can benefit from an increasing IT influence
[ "Business performance", "IT business value", "Empirical results", "Strategic IT management", "IT department’s influence" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The value of information technology (IT) and its contribution to corporate performance has been one of the most discussed topics in the literature for years. Both scientists and practitioners agree that effective information provision and management can contribute competitive advantage to companies. While competition in the market is significantly increasing, the pressure on companies and their management also increases, and they, in turn, put enormous pressure on individual departments to perform better. Departments must not only justify their existence, but also their position within the company. From a strategic management perspective, this includes a continuous evaluation of the individual business units and an adequate cost-benefit analysis of each department, including IT, to ensure that there is \"a source of value creation instead of a cost\" (Tippins and Sohi, 2003). Caused by the rapid technological development, the IT department is increasingly moving into the focus of performance evaluation by delivering crucial intangible assets through knowledge and core competencies (Sampler, 1998). A clearly defined area of responsibility and a strategically defined orientation of the IT department in the company seems indispensable, not only for the established \"dot-com companies\" of the new economy such as Amazon, Google and Facebook, but also for newly emerging digital business models that focus on web-based services.\n2. Conceptual model and hypotheses development: 2.1 Conceptual model\n3. Methodology: 3.1 Measure development\n4. Measurement model and research results: 4.1 Measurement model\n5. Discussion: After systematic consideration of previous findings from the literature, scientists do not agree on the role and influence of the IT department as well as the resulting importance of the IT for the company. The study therefore provides information on the extent to which the IT management has a right to exist in the company. The results of the PLS-SEM analysis to what degree an influence in the company can be attributed to the IT department today. It turns out that the IT department has a direct influence on the performance of the company. Thus, it has been empirically proven that the decision-making influence of the IT department has a positive impact on both monetary and non-monetary business performance. The study therefore has interesting implications for research and practice that are discussed below, together with the limitations of the study. Finally, an outlook for future research and a conclusion follows.\n6. Conclusion: An increasing competition and the associated cost pressure require an increase in efficiency and are one of the main motives for companies to equip IT for the age of digital change (Singh and Hess, 2017). Important findings of the study are the knowledge of significant influencing determinants of the IT department that offer enormous potential for a sustainable success of a company. In a growing digital market environment, investments in the IT department, specifically in its management and relative size, the degree of digitization and the value assessment of IT, create advantages over competitors and have positive effects on business performance. Hence, the tasks of an IT department have changed considerably over the last few years, not least due to digitization and Industry 4.0. Through social networking and the Internet of Things, the IT department has established itself as an elementary business area that no modern company can do without in order to survive on the market in the long term. Especially new requirements due to the progressive digitization have led to an organizational change in the IT department. The competencies of pure system support have receded into the background; the focus is on the integration of new services and technologies. The involvement of the IT department in strategic decisions and corporate planning is a critical success factor in terms of monetary as well as non-monetary performance and more than ever one of the key issues for IT executives (Luftmann and Kempaiah, 2008). It is already clear today that the IT department, as part of a growing complexity of the company, must be given the necessary decision-making leeway to be able to contribute to increasing business performance. As the majority of companies still have some catching up to do in anchoring IT knowledge in the TMT and increasing the degree of digitization, the topic of the organizational and strategic orientation of IT will continue to be relevant for both research and practice.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to analyze the value of a strong decision-making information technology (IT) influence within organizations. Although research and managerial practice has repeatedly shown the importance of IT departments within firms and has commented on the influence of IT departments on business performance, prior research has still no evidence about the value of a strong decision-making IT influence yet.\nDesign/methodology/approach: To prove the influence of the IT department within the company, this study identifies and operationalizes a formative construct determined by four main specifics of the IT department: IT department size, IT department value assessment, IT experience of the top management and degree of digitization. A questionnaire was used to collect the data of 124 experts from companies that could be assigned mainly to the sectors manufacturing, trade as well as information and communication. The data were analyzed using exploratory and confirmatory factor analyses. Further, partial least squares structural equation model (PLS-SEM) was used to test the proposed model.\nFindings: The results show that both investments in the size of the IT department as well as in top managers with IT experience, and a high degree of digitization in a company positively influence the role of the IT department. It also shows that a higher general appreciation of the IT department goes hand in hand with a higher influence of IT in the company. The measures are significant as companies, which do have an influential IT department, actually have higher monetary as well as non-monetary business performances.\nPractical implications: The study is aimed equally at science and practice, as it provides information on the extent to which more importance should be attached to IT management in the future and what organizational adjustments need to be made.\nOriginality/value: Despite the ongoing discussions on the importance of IT management for business performance, no existing studies have delivered evidence that there is a significant direct link between the decision-making influence of IT and the extent of corporate performance. The present work therefore has two objectives. The theoretical goal is to clarify the impact of the IT department on business performance and to identify the factors that make up an influential IT department. The practical objective of the research is to provide recommendations on how firms could establish or expand the IT department.\n", "role": "assistant" } ]
Patient perceptions of quality: analyzing patient satisfaction surveys
[ "Quality", "Regression", "Healthcare", "Analytics", "ACA", "Patient satisfaction surveys" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Healthcare facilities are entering an era of increased oversight and heightened expectations concerning both reduced costs and measureable quality. The US Affordable Care Act aims to provide more healthcare coverage while reducing costs and elevating care quality. In order to monitor and improve the quality of healthcare, increased emphasis has been placed on evaluation metrics. These metrics, including patient assessments of quality, are important components of the pay-for-performance initiative adopted by healthcare insurers and the Centers for Medicare and Medicaid (Medical Group Management Association (MGMA), 2013). Consumers also have access to this data, which they can use to select healthcare providers and health insurance plans. According to the US Health and Human Services Administration, patient satisfaction surveys are valuable tools for elucidating patient perceptions of the care and services they receive. The results of these surveys are used in quality assessments, improvement efforts and healthcare provider comparisons. Capturing the voice of the patient is critical to providing managers with the data and analytic capabilities required to make well-informed decisions about healthcare policy and procedures (Carlucci et al., 2013). Effective use of patient satisfaction and quality outcomes can create a competitive advantage for healthcare organizations (Urden, 2002). Competition for healthcare dollars often rests on how satisfied patients are with their provider. Therefore, it is essential to draw as much valuable information as possible from patient satisfaction surveys.\n2. Methodology: This research uses a case study to illustrate the usefulness of a more rigorous analysis of patient survey data. The data for this study were obtained from Portage Health[1], a multi-facility healthcare system located in the western Upper Peninsula of Michigan. The system includes a 36-bed acute care hospital, several family practices, numerous specialty medical practices and multiple hospital-owned clinics; including a university center and express care clinics. Portage Health Medical Group has recently received designation as a Patient-Centered Medical Home (PCMH), which means, among other things, that team-based healthcare is emphasized and patients are encouraged to be actively involved with healthcare professionals in both treatment plans and quality assessment.\n3. Model results: Table III shows the results of the regression analysis for the three overall measures of patient satisfaction. An analysis of variance for each model showed a Prob\n4. Implications and limitations: While gender and provider type do not appear to have an impact on overall patient satisfaction, statistical analyses suggest there may be real differences in overall patient satisfaction based on age range, specialty of the physician, and clinic type. In each overall satisfaction rating (and with the exception of children), older respondents gave higher ratings on average than younger ones. The smallest family practice center, in particular, received higher ratings than other clinics, and the EXPRESS and OBGYN clinics received poorer scores than most of the other clinics. Family practice and rheumatology received the highest patient satisfaction scores among medical specialties.\n5. Conclusions: This study has examined patient satisfaction surveys of medical practices administered by a community-based non-profit healthcare group. The healthcare organization studied receives consistently high-quality evaluations, but sought to explore more fully which factors contributed to those evaluations and which factors needed improvement. The study differs from most academic and industry research because it focusses on medical practices and clinics (rather than hospitals); it examines clinic type, medical specialty and physician type as part of the analysis; and it incorporates the CAHPS PCMH item set in the survey instrument.\n", "role": "user" }, { "content": "Purpose: Healthcare facilities are entering an era of increased oversight and heightened expectations concerning both reduced costs and measureable quality. The US Affordable Care Act requires healthcare organizations to collect certain metrics, including patient assessments of quality, in order to monitor and improve the quality of healthcare. These metrics are used as a basis for graduated insurance reimbursements, and are available to consumers as an aid in selecting healthcare providers and insurance plans. The purpose of this paper is to provide healthcare providers with the analytic capabilities to better understand quality of care from the patient's point of view.\nDesign/methodology/approach: This research examines patient satisfaction data from a multi-specialty Medical Practice Group, and uses regression analysis and paired comparisons to provide insight into patient perceptions of care quality.\nFindings: Results show that variables related to Access, Moving Through the Visit, Nurse/Assistant, Care Provider and Personal Issues significantly impact overall assessments of care quality. In addition, while gender and type of care provider do not appear to have an impact on overall patient satisfaction, significant differences do exist based on age group, specialty of the physician and clinic type.\nOriginality/value: This study differs from most academic research as it focusses on medical practices, rather than hospitals, and includes multiple clinic types, medical specialties and physician types in the analysis. The study demonstrates how analytics and patient perceptions of quality can inform policy decisions.\n", "role": "assistant" } ]
Satisfaction of learning, performance, and relatedness needs at work and employees' organizational identification
[ "Organizational identification", "Learning needs satisfaction", "Performance needs satisfaction", "Relatedness needs satisfaction" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Organizational identification (OID) concerns the perception of psychological oneness with one's organization (Ashforth and Mael, 1989). It is believed that identification is a process in which an employee comprehends the signals sent by the organization and establishes a link with the organization after becoming aware of various similarities between self and the organization (e.g. O'Reilly and Chatman, 1986; Cheney and Tompkins, 1987). In the opinion of Ashforth et al. (2008), identification with the organization reflects a fundamental connection that other attachment constructs (e.g. commitment, psychological ownership, and engagement) lack. OID for an individual is closer to self-definition of an individual as compared to other similar constructs (Ashforth et al., 2008) and is a root construct that, in turn, may foster commitment, psychological ownership, and engagement (see Ashforth et al., 2008; Hansen et al., 2014; He et al., 2014).\nLiterature review: Satisfaction of learning and performance need and OID\nMethod: Sample and procedures\nData analysis: To begin with, we determined the extent of common method variance by conducting Harman one-factor test using confirmatory factor analysis (Podsakoff et al., 2012). All the items were included in a one-factor model and estimated using IBM(r) SPSS(r) AMOSTM 22. If the results indicate that one-factor model fits the data well, then common method variance is a powerful force in the study. Otherwise, the problem is not prevalent.\nResults: We first specified the one-factor model in which all the items used in the study were loaded on a common factor. This model revealed a bad fit to the data in absolute sense (see Table I). Next, we specified a two-factor model in which all the items pertaining to various needs loaded on one factor and items measuring OID loaded on another factor. This model revealed a fair fit to the data in absolute sense (see Table I). In comparison with Model 1, Model 2 was a better fit to the data on account of a significant kh2 difference test (Dkh2 (1)=313.12, p\nDiscussion: Through this paper, we address the dearth of research on the relationship between need satisfaction and OID particularly in the Indian context. In this study we found support for the positive relationship of growth need satisfaction (second-order factor of LNS and PNS) and RNS with OID of members.\nConclusion: An attempt was made in this work to integrate some of the learnings of SIT and needs theories. As espoused in SIT, OID is expected to have personal and social antecedents which are core to self-definition of an individual. Employee needs and motives as personal level variables, in a way, are likely to manifest the notion of \"who am I\". However, individual motive is a nuanced concept. Therefore, a selective set of three needs; performance needs, learning needs, and relatedness needs were considered in this study. Satisfaction of individual needs were found to be related to OID.\n", "role": "user" }, { "content": "Purpose: In today's scenario when loyalty can no longer be demanded from employees, the extent of organizational identification (OID) of employees predicts outcomes of organizational interest. It is therefore essential for organizations to foster the sense of oneness in employees. Since, need satisfaction lies at the core of human motivation, the purpose of this paper is to test if satisfaction of learning, performance, and relatedness needs at workplace could influence employees' identification with the organization.\nDesign/methodology/approach: For this study, the authors collected data from 365 professionals representing diverse work backgrounds and analysed it using structural equation modelling approach.\nFindings: Results suggest that OID is positively influenced by satisfaction of these three needs and about 45 per cent of its variance is explained by these needs.\nOriginality/value: The study corroborates the relevance of the three needs in organizational context. Although not a part of design of the study, nonetheless the study sheds light on methodological and theoretical possibilities of salience of needs. Therefore to the best of awareness, this study stands apart from other studies exploring relationship of disparate needs with OID. The authors believe the study enriches social identity theory literature in indicating that researchers should not only consider focusing on relationship of needs variables with OID but also the dynamics of interaction between various needs among themselves as well as on their differential relationship with OID.\n", "role": "assistant" } ]
2006 Poetry Publication Showcase: selected titles
[ "Poetry", "Literary criticism" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Poets House, it needs to be said, is not a warehouse of past dreams but an ongoing experiment in cultural imagination (Stanley Kunitz).\nIndividual authors: Adamson, Robert, The Goldfinches of Baghdad, Flood Editions, Chicago, IL, 2006 (paperback, $13.95, ISBN 0974690287)\nAnthologies: Bargallo, Luis Cortes, (Ed.), Connecting Lines: New Poetry from Mexico, Sarabande Books, Louisville, KY, 2006 (paperback, $16.95, ISBN 9781932511208)\nProse on poetry: Davis, Peter, (Ed.), Poet's Bookshelf: Contemporary Poets on Books That Shaped Their Art, Barnwood Press, Seattle, WA, 2005, (paperback, $14.95, ISBN 0935306501)\n", "role": "user" }, { "content": "Purpose: - The purpose of this paper is to highlight a selection of poetry titles from the Poets House Showcase of 2006.\nDesign/methodology/approach: - This article provides reviews of selected titles from the 2006 Poets House Showcase.\nFindings: - This review represents a wide-ranging selection of contemporary poetry collections and anthologies.\nOriginality/value: - This list documents the tremendous range of poetry publishing from commercial, independent and university presses, as well as letterpress chapbooks, art books and CDs.\n", "role": "assistant" } ]
The selection of sharing ratios in target cost contracts
[ "Target costs", "Risk management", "Sweden", "Contracts", "Construction industry" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: A target cost contract reinforces collaboration between the contractor and the client by letting them share economic gains or losses (Bresnen and Marshall, 2000). In the broader partnering literature, such arrangements have many names, including cost-plus-incentive-fee contracts in the USA (Al-Harbi, 1998) and financial incentives (Bresnen and Marshall, 2000). The formula for a target cost contract used in this study is that modelled by Scherer (1964): Equation 1 where:PC = the contractor's new fee;CT = the target cost;CA = the actual cost;PT = the target profit (basic fee);a = the contractor's sharing ratio 0<= a <=1.Despite this simple formula, the use of target cost contracts still presents problems (Perry and Barnes, 2000). Some of the difficulty may be associated with the fact that there are three negotiable terms in target cost contracts: a target cost CT, a target profit PT, and a sharing ratio a (Gandhi, 1979). The selection process is further complicated by the fact that the sharing ratio may vary as gains are often shared differently from pains (Perry and Barnes, 2000; Broome and Perry, 2002). It has been argued that the selection of a sharing ratio is affected mainly by the contracting parties' perceived level of risk (Al-Harbi, 1998).Although there is an extensive body of literature about financial incentives in contracts, there are few empirical studies of the negotiation of a target cost contract (Al-Harbi, 1998; Broome and Perry, 2002). Fewer still focus on the factors that influence the selection and negotiation of sharing ratios in construction contexts. Although contactor motivation and risk assessment are subjects that are well suited to empirical research, most research on them is highly theoretical. Broome and Perry (2002) identified this limitation and conducted empirical research in an attempt to give some guidance on how to select a sharing ratio in a target cost contract. They concluded that there should be more study of the interaction of risk and the selection of a sharing ratio. Yet, despite the identification of risk sharing as a fundamental issue in target cost contracts, many researchers neglect the analytical potential of incentive theory within the framework of agency theory.The purpose of this study is to provide a broader understanding of the design process by empirically studying the choice of sharing ratios in Swedish construction contracts from the perspective of both clients and contractors. Agency theory is used in the discussion of the impact of asymmetry of information and long-term relationships on the design of a target cost contract. The following research questions were asked:RQ1. What key factors influence the selection and negotiation of a sharing ratio in a target cost contract?RQ2. What underlying aspects of perceived level of risks can be identified from an agency perspective?This paper begins with a theoretical framework setting out the basic elements of agency theory and theories about the factors that impact on the selection of a sharing ratio in a target cost contract. Empirical findings from an exploratory interview study and a case study of a target cost contract are then presented. Finally, the theoretical framework is applied to answer the research questions.\nTheoretical framework: Many economic arrangements that involve risk sharing and goal alignment may be described in terms of a principal-agent relationship (Shavell, 1979; Weitzman, 1980). Agency theory generally focuses on tasks that are too complicated or too costly to do on one's own. Thus the principal is obliged to hire an agent with specialized skills or knowledge to perform the task in question (Sappington, 1991). In construction, the client is the principal who has the right to write a contract with a contractor (agent). In drawing up this contract, the principal assumes that the agent is opportunistic and will seek his own benefit (Petersen, 1993).Agency relationships involve two critical problems. The first is that the principal and the agent are likely to have conflicting goals. The second is that different perceptions of risk may lead the agent and principal to prefer different actions. It can also be expected that many of the arguments put forward by the contracting parties in the process of negotiating a target cost design will be affected by what agency theory describes as asymmetric information. For instance, when only one of the parties has critical information about some aspect of the contract, that party can use this information to gain an advantage. Even if both the principal and the agent can observe the outcome, only the agent observes the hidden action taken for the accomplishment of a certain outcome.The selection of the right agent can be seriously affected by problems related to information asymmetry. This asymmetry may involve a moral hazard - for example, the possibility that the chosen agent may be a shirker. Or it may involve adverse selection that is, choosing the wrong agent. Adverse selection occurs because the principal has difficulty in verifying information about the agent's skills or abilities, and the moral hazard is due to the asymmetry of information among individuals (Holmstrom, 1979). According to Holmstrom (1979) and Eisenhardt (1989) the solution to the moral hazard problem is to reveal the agents' behaviour with the help of an information system such as open-book accounting in a construction project. Another solution is to have contracts based on a long-term relationship, since such contracts tend to mitigate dysfunctional behaviour because the client-contractor interaction is repeated over time, thus reducing uncertainty and thereby alleviating the problem of moral hazard (Holmstrom, 1979). In terms of project risk, moral hazard may appear as either performance risk or relational risk (Das and Teng, 2001a).Relational risk reflects the level of cooperation between the contracting partners and is mainly concerned with the probability that an agent may demonstrate opportunistic behaviour (Das and Teng, 2001a; Elmuti and Kathawala, 2001; Ireland et al., 2002). Further, since risk sharing may be used as a bonding tool in more collaborative arrangements, it is natural to question what will happen if one of the contracting parties is successful while the other experiences a failure (Elmuti and Kathawala, 2001). In many explicit collaborative arrangements a displeased partner may blame the partnering company, and thus increase perceived relational risk (Elmuti and Kathawala, 2001). On the other hand, performance risk is grounded in the interactions of the contracting partners and regards those factors that may get in the way of achieving the project objectives, given that the partners cooperate fully (Das and Teng, 2001b). Further, while performance risk is common to all strategic decisions, relational risk is a characteristic of more collaborative arrangements (Ireland et al., 2002), where the decision makers are concerned with the partner's opportunistic behaviour in cooperative efforts (Das and Teng, 2001a). In the present context, performance risk is understood as referring to the decision maker's perceived risk that conflicting goals may have a negative impact on project performance.Sharing ratios\nMethod: To gain a broader understanding of the process of developing a target cost contract, eight Swedish construction clients and eight contractors were interviewed. The clients came from both the public and private sectors and had many years of experience in procuring, coordinating and administering large construction projects. Six were employees of the client organizations and two were project management consultants working for the clients. The eight contractors were project managers for the three largest construction firms in Sweden and one smaller regional construction firm.The interviews were followed by a case study that was intended to yield a deeper understanding of the negotiation process and of the relationship between the client and the contractor. The exploratory nature of this study justifies the use of a case study approach (Yin, 1994). The participants in the particular case studied followed the traditional target cost arrangement used in the construction industry, in accordance with Scherer's (1964) formula, and both had previous experience of working together. The focus of the study was on the selection and negotiation of a sharing ratio. Data were gathered through four semi-structured interviews with key informants, two representing the client and two representing the contractor, who were all directly involved in the contract negotiation and execution process. Further data were collected from contract documents and non-participant observation. To avoid misinterpretation of data, the interviews were transcribed, reviewed and sent to the interviewees for corrections.\nEmpirical findings: The findings from the broader interview study will be presented first, followed by those for the case study. When reading the comments made about sharing ratios during the interview, it should be remembered that, as in all semi-structured interviews, the number of respondents who explicitly state an opinion is likely to be less than the number of respondents who actually hold that opinion.Client views on sharing ratios\nDiscussion and analysis: All the key factors mentioned in the literature review were represented in the empirical data. However, additional factors were also identified: issues of fairness, the level of knowledge about target cost contracts, and the existence of long-term relationships.The tendency to introduce concerns about fairness when selecting and negotiating a sharing ratio introduces another type of complexity, especially since the precise determinants of what is considered to be fair or unfair are still largely unknown and usually vary between individuals. For instance, one contractor interviewed perceived 50/50 sharing as very fair, while a client in the same study thought that they were entitled to the larger share of any cost savings. Some interviewees regarded the latter position as evidence of a lack of generosity on the part of the client, and felt that it could jeopardize the climate of cooperation.The findings in regard to the decision makers' knowledge or, more precisely, skills relating to target cost contracts were interesting. Many of those interviewed had only vague general ideas about risk sharing and contractor motivation. They did not seem to base their thinking on any advanced theories or sophisticated mathematical models.The lack of clear preferences regarding sharing ratios can also be seen as an indication that an important function of target cost contracts is to provide openness and signal trust (an opinion that was voiced in both the interview study and the case study). Thus the details of the economic arrangements are perhaps not particularly important, provided the basic requirements of fair dealing are met. However, since the sharing ratio, the target cost and target fees are interrelated and interact with each other (Perry and Barnes, 2000), the selection and negotiation of a proper sharing ratio should not be thought of in isolation, for there is a risk that inexperienced clients and contractors may end up with a contract design that makes them bear a greater risk than they are prepared to handle.Looking at the empirical data in the light of agency theory reveals a variety of underlying risk issues. For instance, the interviews indicate that there is a trade-off between setting a high sharing ratio and a perceived negative effect on quality. A contractor may prefer to keep costs down at the expense of quality, especially if there is a chance that the target cost may be exceeded. Clients who are aware of this performance risk may aim for a low-sharing ratio when quality concerns are high. However, since quality may be hard to verify (Lewis and Sappington, 1991), the client cannot be sure that the size of the sharing ratio will affect the project's quality.The interviews also reveal that clients think that there should be a maximum limit to how much a contractor can gain from a project because the contractor is believed to have a considerable information advantage in setting target costs. This belief brings out the relational risk implicit in asymmetric information. The contractor may behave opportunistically, especially when possessing exclusive information about material costs, estimated labour-hours, new techniques and ground conditions, etc. However, it is also possible that the client has advantageous information about similar matters based on previous experience with similar projects.If a client is known to behave unfairly, a contractor may evaluate this as a risk that the target cost may be exceeded and may prefer a lower sharing ratio. Therefore, when negotiating an efficient sharing ratio, one must not overlook the type and character of the client. Moreover, while it is quite easy for the client to explore the contractor's organization by collecting references, the contractor has less access to information about the client, a reality that is often overlooked in principal-agent analyses. This factor is especially significant if the client is a first-time buyer with no previous experience.\nConclusion: A sharing ratio may be selected on the basis of previous experiences of working together, indicating that long-term relations and perceived relational risk have an important impact on the size of a sharing ratio because they decrease the asymmetry of information between the contractor and client. This position is supported by the case study data in which the contractor based the sharing ratio on earlier experience of negotiating and working with the client.Previous researchers have regarded the selection and negotiation of a sharing ratio as a one-time event, and thus they have omitted the possible impact of long-term relationships on the perceived relational risk in the design of a target cost contract. The findings reported here should bee seen as an exploration of the impact of long-term relationships. The qualitative approach used obtains unique and in-depth information about sensitive matters, such as information about personal insights and experiences that provide a stronger base for further investigations than a broad quantitative study would have given at present. Moreover, the current study provides inspiration for future studies of how the selection process of a sharing ratio in a target cost contract can be explained by factors such as relational risk which have traditionally been excluded from the analysis. However, in order to create normative knowledge, further qualitative and quantitative research is planed, focusing on questions related to the risk allocation process and to the extent that the result of a target cost contract impacts on relational risk. Another extension of this line of thought is to investigate possible actions to be taken by the contractor in response to perceived costs of involving additional specialist input or for developing more sophisticated purchasing systems. In this manner, supply change consequences may arise from incentive arrangements.To conclude, this paper stresses that in order to come up with a proper sharing ratio, as well as an effective contract design, the contracting parties need to collect trustworthy data, not only about the project, but also about each other. This data should cover areas such as skills, reputation and knowledge about target cost contracts. Hence, the impact of long-term relationships on the design and outcome of target cost contracts should not be neglected, especially since they reduce the relational risk and other risks associated with asymmetric information.\n", "role": "user" }, { "content": "Purpose: - The purpose of this paper is to provide a broader understanding of target cost arrangements by empirical study of the choice of sharing ratios from the perspective of both clients and contractors.\nDesign/methodology/approach: - Eight Swedish construction clients and eight contractors were interviewed. These interviews were followed by a case study of a large construction project with a target cost contract. The data for the case study were gathered through interviews, contract documents and non-participant observation. The impact of perceived risks on the selection process is discussed in terms of agency theory.\nFindings: - Key factors influencing the selection of a sharing ratio included perceptions of fairness, knowledge of target cost contracts, and long-term relationships. The perceived level of risk is affected by the perceived performance risk and the perceived relational risk. Consequently, attention should be paid to the impact of long-term relationships on the design and outcome of target cost contracts since such relationships reduce the risks associated with asymmetric information, such as relational risk.\nOriginality/value: - This study identifies factors that previous research has ignored in relation to the negotiation of target cost contracts.\n", "role": "assistant" } ]
Assessing the benefits of B2B trade cycle integration: a model in the home appliances industry
[ "Order processing", "Invoicing", "Activity based costs", "Electronic commerce", "Payments" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1 Introduction: In the last decades, supply chain collaboration has been extensively studied by academics and its relevance in boosting company performances has been widely recognized. Integration of the B2B trade process (or order-to-payment cycle) is often considered the first necessary step in the implementation of collaborative processes among a company and its supply chain partners (Lee et al., 2000, Fawcett et al., 2009). The order-to-payment cycle can be defined as the process by which the trade partners create, exchange and process the documents in a commercial transaction, e.g. purchase orders, notifications of acceptance of a purchase order, delivery notes, notifications of acceptance of goods, invoices, credit/debit notes and notifications of payment (Shin and Leem, 2002), as well as all the documents exchanged through the banking system, e.g. payment orders and bank statements (De Boer et al., 2008). Integration can be obtained by exchanging structured electronic documents through technologies such as electronic data interchange (EDI) and the more recent internet-based applications so that they can be directly processed by the information system of the receiving company (Mackay and Rosier, 1996), and by managing the order cycle as a single inter-company process through an effective workflow that links all the involved functions (Chung et al., 2007).Despite the importance given to this topic in the literature, and the apparently clear benefits that trade process integration can bring in terms of efficiency (Fredenhall and Hill, 2001; Boyson et al., 2003), i.e. a reduction in the costs to operate the process, and effectiveness (Lee and Whang, 2001, Araujio and Easton, 2003), i.e. a faster and more accurate process, the degree of adoption of these solutions is still low compared to their potential. This is especially true in industries like home appliances and apparel, but it also applies to industries that were the pioneers in the adoption of these solutions, e.g. the groceries and the automotive, where many companies are still non-adopters or exchange only a very limited set of documents. There are various reasons for this, but one of the most important is that the trade process is typically not managed as a process, but parcelled out among different functions. As a matter of fact, very few companies measure trade process key performance indicators such as the cost of the overall order-to-payment cycle or the order-to-cash time. The benefits of integration are therefore often perceived as local and insufficient to repay the investment, while the fact that inaccuracies generated in one phase can have a significant impact on other functions is disregarded. Moreover, the literature currently offers no benchmark regarding the cost of the order-to-payment cycle in conventional (i.e. paper exchanges) and integrated (i.e. exchange of structured electronic documents) conditions. Such a measure would help companies to assess the benefits that might derive from integration.The objective of the present paper is to assess the costs in the home appliances industry of the conventional trade process and the benefits achievable through information and communication technologies (ICT)-based integration. In particular, we have analyzed the processes involving the manufacturers (or importers) and the specialized retailers, and by using an activity-based costing model, we have been able to estimate the process cost (including the usually neglected non-conformity costs) when companies use paper documents and when they adopt an integrated process. Besides, providing a numerical example of the benefits and a useful benchmark for this particular sector, the paper presents a methodology that has already proved valuable in other industries.This study has been carried out together with the EDIEL technical committee. EDIEL is an initiative, promoted by some of the main manufacturers and specialized retailers in the home appliances sector in Italy, born in 2007 with the objective of improving supply chain performance through ICT. The initiative focuses on three main work streams: the master data alignment, the integration of the trade process through the exchange of structured electronic documents (EDI), and the radio frequency identification, with the first two areas considered fundamental and preliminary to the third one. A technical committee, involving the IS directors of 19 companies in the industry covering more than 60 per cent of the market in terms of sales, was formed with two goals:1. to tackle the operational issues necessary to the implementation, such as process and document standards; and2. to drive the companies to adoption, identifying the actions to support the change management.The assessment of the potential benefits has been considered a key issue to promote the adoption (as suggested by many authors, Becker, 2008): a plan was therefore jointly prepared by the authors and the technical committee in order to build a business case assessing the benefits of the trade process integration for an average manufacturer and an average retailer in the industry.Five industry associations were also involved in the EDIEL project and in the benefits assessment: ANIE ICT-CE, the association of the consumer electronics manufacturers; ANDEC, the association of the consumer electronic importers; CECED, the association of white goods manufacturers; AIRES, the association of specialized retailers; and Indicod-ECR, the GS1 representative for Italy.The paper is organized as follows: the next section presents the most significant contributions in the literature, Section 3 illustrates the research scope, and Section 4 introduces the methodology and the model. Finally, Section 5 reports the results of the application of the model, while Section 6 provides some concluding remarks.\n2 Literature review: 2.1 Trade process integration benefits\n3 Research scope: 3.1 Trade process in the home appliances industry\n4 Research methodology: In total of 19 companies (seven retailers and 12 manufacturers) - all the participants to the EDIEL technical committee - and four industry associations were involved in the research together with Indicod-ECR (Italian representative of GS1). We selected only the companies belonging to the EDIEL technical committee in order to have the highest level of understanding and commitment. Table I lists the companies and associations taking part in the initiative.To assess the trade process costs in the different technological scenarios, a three-step methodology was followed:1. description of the trade process in the home appliances industry both for the conventional, the intermediate and the integrated scenarios;2. construction of a model following the ABC approach; and3. data collection and validation.4.1 Trade process description\n5 Results: 5.1 Basic case\n6 Conclusions and further research: The paper presents an assessment of the benefits of trade process integration in the home appliances supply chain, exploring the interactions between manufacturers and retailers. The reduction in trade process costs in the baseline scenario, including only the benefits from efficiency and greater process accuracy, is about EUR74 per cycle, a process cost reduction of about 80 per cent. Moreover, the distribution of the benefits is almost equal between the manufacturer and the retailer. The general conclusion that can be drawn from the application of the model to home appliances is that substantial benefits can be obtained from the integration of large parts of the trade process. The exchange of just a few documents without integration of processes results in more limited benefits, making investments in these solutions less attractive.The proposed model has some limitations in terms of benefits considered and data collection. Assessment of operational benefits could be enhanced through the evaluation of the impact of the faster order-to-cash process. For a complete evaluation, the strategic benefits should also be considered. A second limitation is that the study has mainly been based in Italy. Nevertheless, as most of the companies who participated in the research are multi-nationals, their processes are replicated in exactly the same way in other countries, meaning that results can be easily extended.The same methodology has been applied to other supply chains (i.e. groceries, electrical equipment and pharmaceuticals), in collaboration with the respective industry associations and companies. The first significant variation to point out is the entity of the trade process costs in the conventional scenario: this is lower in all the other industries considered, reaching a minimum of EUR36 in the pharmaceutical sector. The reason for this difference is essentially because of a leaner trade process compared to the home appliances sector, especially in the order phase, and because of fewer non-conformities. In fact, the pharmaceutical industry presents a high level of regulation (e.g. item code unification at national level, fixed prices, etc.), which significantly reduces misalignments. Anyway, depending on the context (sector and company type), these benefits can range from EUR25 to EUR100 and represent between tenths of a per cent and several percentage points in turnover. Bearing in mind that in many sectors the supply chain can be long and complex and therefore involve many customer-supplier relations in the path from raw materials to finished products, the resulting value for the companies involved and the final customer is considerable.\n", "role": "user" }, { "content": "Purpose: - The paper aims to give a quantitative assessment of the benefits obtainable by using information and communication technologies to integrate the order-to-payment process involving manufacturers and specialized retailers in the home appliances industry.\nDesign/methodology/approach: - Standard business processes were defined and activity base costing was applied to develop the cost model with the support, i.e. validation and data provision of most of the companies operating in the Italian home appliances market.\nFindings: - The bottom line is EUR96 per order-to-payment cycle in the conventional, paper-based scenario, compared to EUR23 in the integrated scenario involving the exchange of structured electronic documents. The potential savings amount to about 80 per cent of the costs, almost equally shared between the retailer and the manufacturer. A significant part of the benefits comes from the reduction in the management costs of non-conformity issues.\nResearch limitations/implications: - The assessment considers only tangible benefits and does not include more intangible advantages, such as cycle time reduction, compliance with regulation and easier and faster accessibility to documents.\nPractical implications: - The paper measures the benefits of using trade process integration technologies and presents a methodology that can be applied to other industries and to particular supply chains. The difficulties in assessing the benefits have been claimed to be one of the main barriers to adoption.\nOriginality/value: - The paper provides a model to assess the order-to-payment cycle costs, considering both the costs of the \"perfect cycle\" (where no errors occur) and the cost of managing non-conformities. At present, there is no benchmark available in the literature for the order-to-payment cycle costs in conventional and integrated conditions that could help companies to assess the expected benefits of integration.\n", "role": "assistant" } ]
Innovation in Swedish hotels
[ "Innovation", "Hotels", "Hospitality", "Sweden" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Innovation in tourism and hospitality is vital for business survival, competitiveness and growth (Orfila-Sintes and Mattsson, 2009; Ottenbacher, 2007; Sundbo et al., 2007). Its positive impact on customer preference, service quality, employee productivity, firms' market value and share and customer retention is widely recognized (for a review, see Gomezelj, 2016; Hall and Williams, 2008; Hjalager, 2010).\nLiterature review and hypotheses: Defining innovation\nMethod: Sample and data collection\nResults: To explore the degree of service/product, process, organizational and marketing innovations in the hotel sector in Sweden, respondents answered the question whether their hotel had produced any of the four innovation types in the previous year. In total, 62 per cent produced service/product innovations, 46 per cent process innovations, 50 per cent organizational innovations and 60 per cent marketing innovations. Incremental innovations constitute the majority of the produced innovations (73 per cent compared to 27 per cent). Table I shows the percentage of hotels based on their involvement in the four innovation types. To test H1, we used the chi-square test for each innovation type. This test compares if the percentages of radical and incremental innovations are equal. The statistical results of the chi-square test provide support for rejecting the null hypothesis (p < 0.001). The proportion of incremental innovations is higher than the percentage of radical innovations (for the four innovation types, Table I). Thus, H1 was supported.\nDiscussion: The results of this study reveal that more than 60 per cent of responding hotels produced service/product and/or marketing innovations. Lower percentages were reported for organizational (50 per cent) and process (46 per cent) innovations. These rates reflect innovation-involved organizations, supporting previous hospitality studies (Hall, 2009; Martinez-Ros and Orfila-Sintes, 2009; Mattsson and Orfila-Sintes, 2014).\nConclusions, implications and limitations: Conclusions\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to assess innovation in the hotel sector in Sweden and to investigate how structural and organizational factors influence hotel's likelihood of producing service/product, process, organizational and marketing innovations.\nDesign/methodology/approach: This paper is based on responses from 174 hotels with membership in the Swedish hotel association. Responses were collected via a web-based survey.\nFindings: This paper provides insights about the nature and extent of innovations in the hotel sector. Although traditionally considered rigid and non-innovative, around half of the responding hotels produced at least one type of innovation. Most common are service/product and marketing innovations. A hotel's likelihood of innovating depends largely on structural independence (non-chain), having an explicit innovation strategy and investing in non-traditional R&D.\nResearch limitations/implications: Because of the chosen design (convenience sampling), the results of this paper may lack generalizability. Therefore, future research is encouraged to test the hypotheses further.\nPractical implications: Managers in the hospitality industry can influence the production of innovations in the hotel sector. By promoting flexibility, defining and communicating an innovation strategy, and engaging in non-traditional R&D activities, practitioners can better respond to the changing business environment.\nOriginality/value: This paper presents a systematic, and internationally recognized, method for assessing four types of innovation in the hotel sector. Its originality stems also from its approach to investigating how key structural and organizational factors, when considered in the same analysis, predict service/product, process, organizational and marketing innovations.\n", "role": "assistant" } ]
Oral health status and behaviour of Mauritians visiting private dental clinics
[ "Personal health", "Personal hygiene", "Dentistry", "Health education", "Mauritius" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Dental caries and periodontal infections are the most common oral diseases. The former is a multifactorial condition, which affects people worldwide and of all ages (Jawed et al., 2006) representing a major cause of tooth loss. International reports indicate that nearly 20 per cent of the young children, almost 80 per cent of the young adults and approximately 95 per cent of the older adults have experienced dental decay (National Institute of Health, 2001). In the elderly population, literature has shown that mouth or teeth problems are also quite frequent and are due to factors such as impaired dentition, changes in chewing ability, dry mouth due to medication, untreated tooth decay and compromised nutritional status. In addition, with increasing age, decalcification results in less resistant enamel and erosion of tooth surfaces. However, the highest incidence of caries formation is during childhood and adolescence. Caries are more likely to occur in fissures and prevail mostly among children, who have less resistant enamel as compared to adults (Truswell, 1999). In teens and young adults, the eating patterns associated with the lifestyle also contribute to caries risk (Majewski, 2001), and increased intake of sugar-sweetened beverages has been identified as a definite risk factor for dental disease (American Dietetic Association, 2007). To establish and maintain good oral hygiene and to prevent the development of dental caries, tooth brushing and flossing are recommended at least once per day. Regular visits to the dentist, topical fluoride exposure, and elimination of prolonged exposure to simple sugars in the diet is also necessary (American Academy of Pediatrics, 2003; Hayward et al., 1989).Transition in the standard of living, lifestyle, dietary habits and dental hygiene practices of the population in the island of Mauritius over the last 40 years may be linked to the prevalence of dental diseases and their adverse consequences. The island is a multiethnic state of 1,900 square km located in the south west region of the Indian Ocean about two-thirds of the way to South Africa from the western coast of Australia. It is one of the most densely populated countries of the world together with Bangladesh, Hong-Kong and Singapore. Since its independence in 1968, its economy has grown steadily and is now classified by the World Bank economists as a middle income developing country. The mortality and burden of diseases statistics of the country highlight the increasing challenges which the public health sector is facing with rising prevalence of obesity, diabetes and other systemic diseases. Few published data on the oral hygiene and oral health status of the population are available and statistics exist for the dental health service of the Ministry of Health and Quality of Life (MOHQL) only (Ministry of Health and Quality of Life, 2006). In its latest report for the year 2006, the public health services indicate that people from all age groups of the population are affected by oral problems and attended public dental clinics, which are free of cost. The reasons reported for acceding to these health services were either for permanent dental fillings or for teeth extractions.For the year 2006 only, less than 3 per cent attendance at public dental clinics concerned babies and pre-primary schoolchildren, 20 per cent primary schoolchildren, and 3.3 per cent pregnant women and nursing mothers (Ministry of Health and Quality of Life, 2006). The percentage of cases for permanent teeth extractions decreased amongst the schoolchildren (from 11.8 per cent to 9.8 per cent in 2001 and 2006 respectively) and the number of permanent dental fillings increased from 43.8 per cent to 48.5 per cent in 2001 and 2006 respectively. The same trend was observed for adults: permanent fillings increased (from 23.3 per cent to 27.7 per cent) whilst permanent teeth extractions decreased (from 52.5 per cent to 41 per cent) over the same five-year period (Ministry of Health and Quality of Life, 2006). For the period 1998 to 2007, the number of private dental practitioners increased by nearly 66 per cent i.e. from 99 to 164 in Mauritius (Ministry of Health and Quality of Life, 2007) and the number of people attending private dental clinics which are fee paying, has followed suit, but no formal statistics are available regarding the clientele nor the specific services sought.Mauritius is a country undergoing rapid transition. Knowing that hygiene practices, lifestyle and nutrition have direct influences on the progression of tooth decay which in turn could be implicated in the development and progression of systemic diseases in later life (US Department of Health and Human Services, 2000; Sgan-Cohen and Mann, 2007); our study was designed to focus on risk factors associated with oral health, emphasising hygiene practices.\nMethodology: Respondents and recruitment\nAnalysis of collected data: Oral health variables\nResults: Demographics\nConclusion: Oral health status and behaviour are dependent on several factors and are country-specific. In many developing countries, oral health education and intervention programmes are limited due to inadequate funding or due to priorities being laid on other health problems. Knowing that in Mauritius, the prevalence of chronic and systemic diseases is on the rise, and knowing the relationship between poor oral health and systemic diseases, the need for researching the status of oral health and its determinants in the country was pertinent and timely. This study demonstrated that socioeconomic disparities influenced oral health status as well as oral health related behaviours. The high prevalence of DMFT scores in children and older adults aged more than 45 years highlighted the need to strengthen dental caries control and prevention activities as well as to develop preventive programmes that are targeted at high risk groups. The better health status of oral health behaviour in females as opposed to males draws attention to the fact that males may need more motivation and encouragement to improve oral health behaviour in terms of time spent for tooth brushing, number of toothbrushes used yearly and frequency of visits to the dentist. Overall, a higher level of attention devoted to the transfer of appropriate dental health education in all the groups studied is warranted to enable patients to become oral-hygiene conscious and active partners within the management of oral health and prevention of adverse effects such as associated systemic diseases.\n", "role": "user" }, { "content": "Purpose: - This paper seeks to assess the oral health status and behaviour of a sample of the Mauritian population visiting private dental clinics.\nDesign/methodology/approach: - Oral health status was determined using the World Health Organization (Decayed, Missing, Filled Teeth (DMFT) index indicating the prevalence of caries, and factors associated with oral health behaviour were investigated. Cross-sectional data on oral health behaviour and demographic characteristics were collected through a structured 22-item self-designed survey questionnaire administered to a randomly selected sample of 250 Mauritians visiting private dental clinics. The data were statistically analysed to examine associations between oral health variables.\nFindings: - Analyses revealed significant correlations between DMFT scores, age and socioeconomic status. Over 80 per cent of the children below the age of 11 had DMFT scores higher than the acceptable WHO value of 3, and most adults older than 45 years had DMFT scores higher than 7. A large majority (85 per cent) reported using both fluoride toothpaste and tongue scrapers, but none reported using dental floss. Those who brushed teeth only once daily had high DMFT scores. Among males, smokers had significantly higher mean DMFT than non-smokers. Females demonstrated better rates of good oral health behaviour than males. Respondents from higher socioeconomic classes were between three to five times more likely to visit dental clinics than those from lower classes.\nPractical implications: - This study reveals the need for intervention programmes to raise awareness of good oral health practices among the Mauritian people.\nOriginality/value: - The results identified the high risk groups and highlighted the need for dental education, taking into consideration socioeconomic status and sociodemographics in order to enable Mauritians to become oral-hygiene-conscious and active partners for the management of oral health and prevention of systemic diseases.\n", "role": "assistant" } ]
Development of Lean Six Sigma training module: evidence from an emerging economy
[ "Manufacturing industries", "Factor analysis", "Lean Six Sigma (LSS)", "COARSE approach" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In this era of global economic integration and technological advancement, there is a rising demand in customers' needs, with different tastes in choice for product quality (George, 2003). With the increased competition and the introduction of new products, manufacturing firms have a compelling need to handle these new challenges for both survival and to outperform the competition. In the past two decades, organizations began to realize the importance of product and service quality, and started adopting quality excellence programs like ISO, total quality management (TQM), Six Sigma, Lean, etc. (Palo and Padhi, 2003; Talib et al., 2011; Gupta et al., 2016). Lean Six Sigma (LSS) is the latest among these quality excellence practices and recognized as an organizational change strategy for multiple benefits (Edward and John, 2005; Sunder, 2016). A few of these include cost reduction, increase in revenue, risk management, quality and productivity improvement, customer and employee satisfaction, error reduction, improved competitiveness, enabling effective decision making, continuous improvement culture building and enabling disruptive innovation (Sunder et al., 2018). Furthermore, literature shows evidence that LSS improves both operational effectiveness and efficiency in organizations (George, 2003; Brett and Queen, 2005; Caldwell et al., 2005; Sunder, 2013; Sreedharan and Raju, 2016). There is an extensive literature available on the emergence and application of LSS. Most of the studies on LSS are concerned with the practices, awareness, relationship development, implementation models, success factors and development of frameworks (Sreedharan and Sunder, 2018).\n2. Literature review: Research shows evidence of LSS philosophy, methods and success. Surprisingly, only a limited number of studies have addressed the employee perception about LSS in the manufacturing industries. Arnheiter and Maleyeff (2005) asserted that for attaining successful implementation, various aspects relevant to LSS should be understood clearly by the employees before the implementation. This also establishes a strong need for the first step toward organizational change management. Even the erstwhile studies endorsed that managers' basic understanding and perceptions about LSS are vital for its success (Naslund, 2008). Furthermore, a majority of the studies are found to be qualitative in nature (Arumugam et al., 2012; Pepper and Spedding, 2010; Scherrer-Rathje et al., 2009). Though one study conducted by Jayaraman et al. (2012) focused on impacts of LSS amongst the electronics manufacturers, it was inconclusive in terms of LSS training and development for the employees.\n3. Research questions and methods: After meticulously going through the literature, it is inferred that studies that attempted to develop training modules in the LSS context are limited. Furthermore, none of these studies were focused to improve the LSS awareness through the training. Hence, there is a vital need for a tailored training module to understand the tools and techniques of LSS, as a resource for managers and LSS champions to create awareness of quality excellence in their respective firms. If this gap of awareness among employees is not bridged, this may lead to several destructing effects including poor project selection, lack of quantifiable results, incorrect perceptions about LSS, ineffective learning and change process, etc., that may retard their quality excellence journeys. Filling this gap could help firms to realize positive ROIs with regard to their quality programs, in specific to LSS practice.\n4. Results and discussion: This paper presents inferences and discussions that have emerged from the survey results collected from 183 manufacturing firms in the Indian sub-continent. The context of the questionnaire was to measure the awareness of the LSS practice and the various related factors concerning it. Erstwhile studies by Vouzas and Psychogios (2007) and Talib et al. (2011) highlighted the areas where TQM awareness can be enhanced. However, they did not suggest training as an approach to increase the employee awareness, which brings in novelty to this paper. Even earlier, Salem et al. (2016) conducted a study on lean awareness in Qatar and found that lean awareness is low in the companies that participated in the survey. The main causes they identified are lack of knowledge on Lean tools and resistance toward change. However, this study was generic and did not consider the aspects of training and within the sectoral scope of manufacturing. Furthermore, it was limited to Lean practice and not to LSS. This paper is novel as it not only evaluated the awareness level of LSS among the Indian manufacturing firms, but also proposed a training module as a means to improve the awareness.\n5. Managerial implications: 5.1 Practical implications\n6. Conclusions: This study presents an examination about the awareness of LSS, a contemporary quality excellence practice in Indian manufacturing sector using an online survey. Results from the survey responses elucidate that Indian manufacturing firms are aware of the LSS program. However, this awareness is limited. Thus, the response to RQ1 is affirmative. To enhance the current level of awareness, a LSS training module is presented. This addresses RQ2. Though the sample representation of training module presented here by itself can make a significant contribution to firms as a first step in their LSS deployments, this needs to be followed periodically with the other associated trainings that any LSS deployment would use (like Green Belt and Black Belt trainings). The authors suggest that the six-sigma toolkit should integrate with the Lean tools while designing these trainings as associated synergies could be leveraged. The present approach provides directions to a tailor-made LSS awareness training as per a specific organizational need and objective. We believe this would bring a wider functional experience to staff in their current roles and enabling the managers to update their professional knowledge and capabilities in quality excellence specific to LSS. Additionally, it would help them evaluate the alternative approaches in a better way, which could benefit their company and create an impact that makes the employees adopt LSS practices in their day-to-day activities. Future research could reinforce, complement or contradict our findings through other empirical methods that could include cases studies or interviews or surveys with larger samples.\n", "role": "user" }, { "content": "Purpose: Nowadays, firms are keen on improving the quality culture in the organizations. The proven success of Lean and Six Sigma has given rise to the synergetic Lean Six Sigma (LSS) approach that has been catching fire in the past one decade. However, there exists a gap between the understanding and implementation of LSS in the organizations, especially in the emerging economies. Taking this as a valuable opportunity, the purpose of this paper is to present a development of LSS training module.\nDesign/methodology/approach: This study starts with a literature review of LSS to reinforce the understanding of the research subject in scope of manufacturing sector. Then, an online questionnaire was designed and used to collect responses from 181 companies located in the Indian sub-continent. Subsequently, the results obtained from the survey were analyzed using COARSE approach.\nFindings: This study reveals two key findings and associated contributions. First, it was found that the overall awareness of LSS within the responded manufacturing firms is about 70.4 percent. Second, there is no single standard training module that exists in any of the sampled firms to cater to their quality programs. Hence, in order to improve the LSS awareness which could subsequently help managers as a resource for creating an efficient workplace, this paper presents a structured LSS training framework.\nResearch limitations/implications: Although this paper presents the importance of LSS and associated awareness level among the responded firms, more empirical evidence is required to generalize the model findings. Second, this study is scoped to firms that work out of the Indian sub-continent, and this provides a future opportunity to expand the scope of this research toward a global study for a comparison between emerging and developed economies. Third, this study is limited to manufacturing firms and hence paves an opportunity to research on a similar theme in services context as well.\nPractical implications: Before embarking on an LSS journey, an organization can use the LSS training module proposed in this study to assess the employee awareness on LSS. Furthermore, organizations that already have a mature LSS practice can incorporate the LSS training module for periodic evaluation of the employees for effective change management.\nOriginality/value: The training module presented in this paper is the original contribution by the authors. This is no association to any single identifiable organization or associated funding. The direct practical implication of its application in real time is the value that managers could derive from the proposed LSS training framework.\n", "role": "assistant" } ]
Inter-organizational social capital as an antecedent of a firm's knowledge identification capability and external knowledge acquisition
[ "Spain", "Absorptive capacity", "External knowledge acquisition", "Biotechnological and pharmaceutical industries", "Inter-organizational social capital", "Knowledge identification capability" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: External knowledge acquisition is a process by which firms integrate knowledge from their environment into their existing knowledge bases (Fey and Birkinshaw, 2005). Knowledge acquisition is also an essential capability for firms developing activities in dynamic environments where competitive advantages depend on renewal, innovation and quick responses to changes in technology, competition and preferences and the needs of customers (Grant, 1996). Most of these companies need to continuously integrate external knowledge to their business processes, as it is complex and dilatory to develop internally all the required knowledge for successful competition (Liao and Marsillac, 2015). Research on knowledge acquisition is thus key to obtaining an understanding of the way that firms are able to identify and develop strategies to gain access to valuable external knowledge without neglecting the importance of complementing it with in-house developed knowledge (Grimpe and Kaiser, 2010; Berchicci, 2013).\n2. Theory and hypotheses: 2.1 External knowledge acquisition and inter-organizational structural social capital\n3. Methods: 3.1 Population, data collection and sample\n4. Statistical analysis and results: The hypotheses were tested by means of a structural equation model (SEM) using the partial least squares (PLS) technique. PLS is a statistical tool of multivariate analysis used for modelling latent constructs by considering non-normality conditions for data and small sample sizes (Hair et al., 2013). The PLS path method is typically applied in two stages. First, the measurement model is analyzed; second, the structural model is calculated and the hypotheses are tested.\n5. Discussion and conclusion: 5.1 Results discussion\n", "role": "user" }, { "content": "Purpose: This paper concentrates on the antecedents of external knowledge acquisition of companies based on their inter-organizational relationships. Specifically, it considers social capital (i.e., the result of a firm's inter-organizational relationships) as an essential precursor of knowledge identification capabilities and deliberated knowledge acquisition strategies. This study aims to propose that cognitive and relational dimensions of a firm's inter-organizational social capital are mediating factors of the relationship between structural social capital and knowledge identification capabilities and the relationship between structural social capital and the deliberated acquisition of external knowledge, respectively. The relationship between knowledge identification capability and external knowledge acquisition is also analyzed.\nDesign/methodology/approach: This is a cross-sectional quantitative study with a sample of 87 firms from Spanish biotechnology and pharmaceutical industries. From an extensive literature review, we developed three hypotheses that were tested using the partial least squares technique and structural equations model.\nFindings: The results only support a mediating effect of cognitive social capital in the relationship between structural social capital and knowledge identification capability and a partial mediation effect of relational social capital in the relationship between structural social capital and knowledge acquisition. In addition, the findings show that firms with more advanced abilities to identify and assess the value of external knowledge will be likely to develop optimal deliberated strategies to acquire effectively such knowledge from its network partners.\nResearch limitations/implications: The limitations of this study are small sample size and the cross-sectional nature of the study. The study also focuses on only two specific and innovative industries.\nPractical implications: Managers should understand that \"good\" management of inter-organizational social capital allows the firm to develop dynamic capabilities for the identification and acquisition of valuable knowledge. The results of the study show that managers should concentrate on building knowledge identification capabilities and should also be aware of the possibilities that social capital can provide to a firm to formulate and implement effective strategies for external knowledge acquisition.\nOriginality/value: To date, there are relatively few studies focussing on knowledge identification capability and its relationships with the dimensions of a company's social capital as enablers of external knowledge acquisition. For managers, the identification of valuable knowledge by using inter-organizational relationships and networks is an essential issue, especially in innovative industries characterized by continuous change. Theoretically, this research highlights that social capital contributes to the development of dynamic capabilities, allowing the firm to sense and seize business opportunities based on external knowledge acquisition to achieve competitive advantages.\n", "role": "assistant" } ]
Detection of malicious traffic on back-bone links via packet header analysis
[ "Internet", "Firewalls", "Control systems", "Data security", "Computer networks" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In this study, internet back-bone traffic has been investigated with respect to potential security problems and many security-related anomalies in packet headers have been found. Internet traffic contains many erroneous packets; some are the result of software and hardware errors, other the result of intentional and malicious activities. We have searched for anomalies in contemporary, highly aggregated internet back-bone traffic. The results show that header problems can be found within almost every aspect being analyzed. In this study, 27.9 billion frames have been collected and protocol headers on network and transport layers have been analyzed in order to point out all behaviour that could potentially result in a security problem for connected hosts. As a result, a systematic listing of all possible packet header anomalies is provided, including their frequencies as seen \"in the wild\" on an internet back-bone link. In addition, the possible security implications of each anomaly observed are discussed.The study of back-bone traffic gives a complementary view to studies of traffic with low level of aggregation, such as traffic in local networks. Back-bone data provide the opportunity to gain a broader picture of different types of malicious traffic present on the modern internet. Besides detection of various types of malicious traffic, specific attack patterns that never show up when studying traffic reaching a smaller network can be observed (e.g. distributed denial of service (DDoS) attacks). Furthermore, rare attacks are more likely to be detected within large amounts of diverse and aggregated traffic and might therefore also reveal previously unreported attack types. Some traffic may seem legitimate when studying only one host but may turn out to be malicious when studying a larger portion of the internet.This study sets out to update and extend older studies (Bykova et al., 2001; John and Tafvelin, 2007; Mahoney and Chan, 2001) which have reported about packets not following modern internet standards (IP, TCP, UDP, ICMP). Additionally, detailed figures about invalid use of fragmentation are included, which is an important security issue previously not covered in the same extent, with the exception of a study about general observations of fragmented traffic (Shannon et al., 2002). While the previous studies only focus on some specific aspects, the present study presents a wider, systematic overview of the most commonly found misbehaviours on the modern internet and how, and in what extent, common protocols are misused by attackers in their search for vulnerabilities in internet-connected systems.As a result, frequencies of occurrence for different kinds of malicious traffic are presented, such as invalid IP packets headers, incorrect use of IP fragmentation, IP address problems, ICMP, UDP and TCP misuse. In the search for anomalies, possible vulnerabilities are listed in table form with references to the packet headers, an approach which should make it easy to find potential problems and make it possible to evaluate the completeness of the study with respect to what header fields are being analyzed. Besides identification of header anomalies deviating from accepted internet standards, particular well-known attacks and their common names (such as Land, Jolt, sPing, Teardrop, Boink, etc.) are pointed out in the tables and the analysis.Many attacks like the ones described above are old and well-known and would therefore be expected to be very rare on the internet and seemingly unlikely to be found in our data. However, the results of the study show that many of these attacks are still present on the internet. One reason may be that the recent arrival of new operating systems and mobile devices with small and newly written IP stacks (e.g. mobile phones and PDAs) may have made these attacks meaningful again. A recent example of this was the introduction of Windows Vista, where it turned out that the beta versions were vulnerable to a large number of such old well-known vulnerabilities (Newsham and Hoagland, 2006).This study shows that it is possible to detect many commonly known attacks from an analysis of network and transport layer packet headers. We therefore believe that our approach to identify potential problems and the scale of our datasets allow us to classify, detect and report a substantial portion of malicious, incorrect and unwanted traffic present on the internet today. The results of this study should not only be interesting for practitioners and researchers but should also be valuable input for work with intrusion detection systems, firewalls and support the design of all kinds of networked applications that must withstand network attacks.1.1. Limitations\n2. Classification of anomalies: In order to provide a systematic overview of possible header misbehaviours and to ease the reading of this paper, the headers of the protocols being analyzed are shown in Figures 1-4 (IP, TCP, UDP and ICMP headers). Header fields highlighted in grey are considered to contain potentially unusual or harmful values and are therefore included in this study. Roman numbers in the figures indicate different types of anomalies within the header fields and Table I provides descriptions of each anomaly. This type of classification makes it possible to identify all packets violating internet standards in any way. The motivations for investigating many of the anomalies can be found in the right column of the tables.\n3. Data description and methodology: The dataset used in this study (John and Tafvelin, 2006) was collected from September to November 2006 on an OC192 back-bone link of the Swedish University Network (SUNET). The packet header traces have been collected on a highly aggregated back-bone link at 277 randomly selected times during 80 days, in order to provide a good statistical representation of all internet traffic during the time-period at this location. At each randomly selected time, two traces of 10 minutes duration were stored. When recording the packet level traces on the 2x10GB links, payload beyond transport layer was removed and IP addresses were anonymized due to privacy concerns using the prefix preserving CryptoPAN (Xu et al., 2001). After further pre-processing of the traces as described in (John and Tafvelin, 2006) and (John and Tafvelin, 2007), the traces were moved to a central storage. An analysis program was run on the raw traces to extract malformed packet headers and invalid series of fragments. The reduced data were then stored in a database together with statistical summaries for each particular observation as listed in section 2. For packets of special interest, corresponding flows have been extracted from the raw traces and analyzed in detail using available packet visualization software.The complete dataset consists of 554 traces including 27.9 billion frames. 99.98 percent of the traffic was IPv4 carrying 19.5 TB of data in 636 million flows. During the single ten-minute intervals, depending on time of day, between 13,000 and 37,000 unique IP addresses were observed belonging to the region of western Sweden connecting to 300,000-1,000,000 unique addresses on the main internet. A breakdown of transport protocols in the IP traffic is summarized in Table II. Not only numbers and fractions of packets are shown but also IP fragments and fragment series are listed (first three lines).The analysis of fragmented traffic requires correlation of fragmented IP packets to create fragment series. Following Shannon (Shannon et al., 2002), a fragment series has been defined as a list of fragments observed on the network derived from a single original IP packet. Consequently, fragments have been grouped into fragment series based on the IP ID, protocol and the source and destination IP fields. Furthermore a timeout value of one second was chosen to further separate fragment series. As opposed to the study by Shannon et al. (2002) which was carried out on OC12 links, the timeout had to be chosen smaller in order to compensate for the higher throughput of the links measured (OC192). In rare cases, wraparounds of the IP ID space have been observed within a few seconds, which made such a small timeout necessary. Considering the transmission rates of modern computers, a timeout of one second seems to be sufficient to capture all fragments belonging to a certain series, which was proven to be true by a number of empirical tests on the dataset. Furthermore, fragment series observed in the first or the last second of a measurement interval have not been tested for completeness, in order to avoid bias due to border effects.Earlier studies have shown that only 0.06 percent of the traffic was fragmented on the measured network (John and Tafvelin, 2007). The increased fraction of IP fragments in the dataset used for this study (0.9 percent) is explained by a special bulk data transfer event from a space observatory to a data centre in Europe. During seven time intervals, 213 million fragments in 9.4 million fragment series have been transferred on the outgoing link using a customized fast bulk transfer protocol based on UDP. Figures for the remaining fragmented traffic without the mentioned bulk transfer are summarized in the final two rows of Table I. Disregarding this special event, only about 0.15 percent of the IPv4 traffic was fragmented.\n4. Observed misbehaviours and anomalies: In the following subsections, occurrences of the anomalies classified in Section 2 as seen in our dataset, are presented in tables. The index columns use the same Roman numbers as introduced earlier. For IP level anomalies a transport protocol breakdown is provided as well. Each table is followed by remarks and discussions about the anomalies being observed, including an interpretation and discussion of their probable causes.4.1 IP header anomalies\n5. Summary and conclusions: In this paper, we first provide a systematic classification of header fields not following internet specifications with a potential to cause security problems. This systematic classification serves as a starting point for identification of such header inconsistencies within our large dataset consisting of packet header data collected on a contemporary, highly aggregated internet back-bone link with diverse traffic composition. Occurrences of each header anomaly as observed \"in the wild\" are then presented followed by detailed discussions about possible causes and an interpretation of the observations with respect to security relevance.As a general observation, it is surprising to see that that many old, well known attacks can still be found. On the upside, some former popular attacks, such as Ping-of-death and the IP source route exploit have not been observed at all. Generally, a constant \"noise\" of malformed or inconsistent packet headers was observed, similar and consistent with the observations of constant scanning activities in another recent connection-level study (John et al., 2008). This type of background noise is in some cases likely to be caused by rare hardware and software errors, but most must be attributed to the possibilities even inexperienced hackers have today to generate more or less random packet headers with existing networking tools.Also, a number of exceptional events of malicious activity have been observed. An ICMP DoS attack with otherwise unsuspicious echo reply messages has been identified due to IP address analysis regarding reserved IP spaces. A sequence of fragmented datagrams has been sent in high intensity from a single host during short time intervals. The detailed analysis of the fragment series revealed a directed Frag attack, using incomplete fragment series with the intention to exhaust resources at the receivers. Furthermore, an analysis of IP ID values of zero appeared to be a successful approach to detect different fragmentation anomalies, and observations in the reserved bits field of the TCP header revealed a series of SYN/ACK attacks. Port number values of zero proved to be effective in detecting port scanning campaigns, both for TCP and for UDP. Finally, a DoS attack applying ICMP redirect messages has been observed.There are many interesting future research possibilities to improve the results and insights of this study, such as a complementary flow-level investigation of scanning traffic or a similar packet inspection including at least some application-level data, to name but a few. However, the results of this study show that it is possible to detect a substantial part of malicious activities just from inspection of header data. The observations also show that inspection of IP addresses spaces, IP ID values, port number values, the entire flags section in the TCP header (reserved bits and signalling flags) and ICMP messages are the most effective mechanisms to find malicious traffic and therefore form a basic set of rules which should be included into all modern firewall and IDS systems.The results presented here are based on a rare dataset of aggregated back-bone traffic and are intended to guide and support network administrators and application developers in their constant task of tuning their systems in order to mitigate the wide range of incoming malicious attack traffic. Furthermore, we believe that this study helps researchers and practitioners to gain a better understanding of the characteristics of today's internet traffic in order to remain proactive.\n", "role": "user" }, { "content": "Purpose: - This study seeks to investigate modern internet back-bone traffic in order to study occurrences of malicious activities and potential security problems within internet packet headers.\nDesign/methodology/approach: - Contemporary and highly aggregated back-bone data have been analyzed regarding consistency of network and transport layer headers (i.e. IP, TCP, UDP and ICMP). Possible security implications of each anomaly observed are discussed.\nFindings: - A systematic listing of packet header anomalies, together with their frequencies as seen \"in the wild\", is provided. Inconsistencies in protocol headers have been found within almost every aspect analyzed, including incorrect or incomplete series of IP fragments, IP address anomalies and other kinds of header fields not following internet standards. Internet traffic was shown to contain many erroneous packets; some are the result of software and hardware errors, others the result of intentional and malicious activities.\nPractical implications: - The study not only presents occurrences of header anomalies as observed in today's internet traffic, but also provides detailed discussions about possible causes for the inconsistencies and their security implications for networked devices.\nOriginality/value: - The results are relevant for researchers as well as practitioners, and form a valuable input for intrusion detection systems, firewalls and the design of all kinds of networked applications exposed to network attacks.\n", "role": "assistant" } ]
Industry analysts - how to conceptualise the distinctive new forms of IT market expertise?
[ "Biography of artefacts", "Conceptual models of expertise", "Industry analysts", "Institutionalizing expertise", "Sociology of professions" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper examines the emergence of a distinctive new form of expert labour - industry analysts who provide research for enterprise technology buyers and producers. Over the last three decades they appear to have gained considerable influence over the Information Technology (IT) market (Bernard and Gallupe, 2013; Pollock and Williams, 2016, in press). Industry analyst research is read and acted on by Chief Information Officers planning organisational IT strategies, IT procurement teams choosing between multi-million pound workplace IT solutions, and company managers and employees wanting to learn more about the potential benefits of digital technologies for their business (Firth and Swanson, 2005; Burks, 2006). Industry analyst output is also extensively read by technology vendors and investor analysts (Snapp, 2013). They have been described (Ikeler, 2007, p. 234) as the \"single most influential validators\" of IT products and services. Foremost amongst these is Gartner Inc., which has emerged ahead of competitors such as Computer Intelligence, Dataquest, Forrester, Input, International Data Corporation (IDC), Ovum and Yankee Group. Its 2013 annual revenue of US$1.784 billion, meant that Gartner was six times larger than its nearest rival Forrester and accounts for about 39 per cent of the industry analyst market (Noble, 2014). We have studied how Gideon Gartner established this firm and in so doing constituted a novel and distinctive model of expert work. This paper considers how we may best conceptualise this expertise.\nContext of our study: Our initial exploration of expertise (Williams et al., 1998) noted the burgeoning number and range of specialist fields of knowledge and associated experts. This pattern has continued, unabated, into the twenty-first century. Growth in the number of experts, and their perceived salience has been accompanied by the proliferation of new kinds of knowledge, including the industry analysts we examine here. The increasing complexity of computer-based systems and the search for improved organisational performance are just some of the factors seen to drive the growth of management consultants and other forms of business expertise. Moreover, \"lean firm\" prescriptions since the 1980s have stressed the economic advantages, indeed competitive imperative, of outsourcing many functions, including specialist labour, that might previously have been employed by the organisation. The outsourcing process has extended from areas seen as outwith the organisation's \"core competences\" to include more strategic issues. As a result, today most firms must buy-in substantial parts of their technical and change management expertise.\nThe Sociology of Professions: \"Professionalisation\" has been held out as one way to resolve this problem of resorting to expert advice. Professionalisation describes the processes whereby a specialist occupation seeks to achieve autonomy/control over the conduct of work and its status/rewards by exercising monopoly over the legitimate application of a body of specialist knowledge, \"organizable as common resources for a body of individuals\" (Abbott, 1988, p. 324), to solve client problems. A body of research over the last three decades from what has been termed the Sociology of Professions has explored the wider uptake of this model, derived from the older professions of medicine and law and the sciences. But do these particular conceptions apply to business knowledge - and the new forms of expertise, such as industry analysts - that are emerging today?\nThe second generation in the Sociology of Professions: From the 1970s a radical critique emerged of this view - which became characterised as the \"functionalist\" account - and the privileged status it accorded to professional expertise (Brante, 1988). Thus Freidson (1970), one of the earliest contributions to what we might call the \"second wave\" in the analysis of professions, criticised (in the case of medicine) the prevalent functionalist theories of professions as \"scientifically-based, practically-efficacious, and socially altruistic\" (Coburn, 1992, p. 497). Larson (1977) in particular revisits Weber's idea that professional groups exercise social closure to advance their interests. Her concept of \"professional project\" highlighted the strategies for sectional occupational advancement accompanying professionalism. Notwithstanding the diversity of currents in the Sociology of Professions, most of these, in reaction against the functionalist account, share an emphasis on how expert communities exercise social closure - to gain authority over an area of expert knowledge and to exclude others - in order to increase rewards and, crucially, to achieve autonomy/control over the exercise of their work (Gross and Kieser, 2006). In similar terms, Freidson (1994) argues that professionalisation seeks to retain control over the definition, organisation, execution and evaluation of its activities. This encompasses control over who has the right to produce certain services (and how such rights are acquired) and how those services are produced (Muzio et al., 2011).\nNew approaches: a third wave of Sociology of Professions and expertise: The Sociology of Professions took as its starting point the classic historical professions - medicine, religion and law, etc. - seen as a template for analysing the wider spread of professionalisation. This approach has, arguably, unhelpfully skewed analysis of expertise and professional work. It has tended to draw a gulf between professions and other specialist occupations (and their strategies to attain expert status) even though these share many common characteristics (Evetts, 2006). Various writers, noting this, have argued that classical professionalisation should be seen as just one strategy of occupational power (Fincham, 2006; Muzio et al., 2011).\nMethodology: This paper draws upon extensive ethnographic research conducted over a period of over a decade. We obtained early insights into the importance of Gartner Inc. in a previous, extended investigation into the development, procurement, implementation and maintenance of enterprise systems (Pollock and Williams, 2009). Building upon this earlier research we were able to apply snowball methods to build a wide sample of contacts with industry analysts and other players allowing us to view industry analysts from a number of different perspectives. These respondents proved remarkably willing to discuss their work and to facilitate further exchanges. In this way we were able to gain detailed insights into how industry analysts carried out their work. This encompassed ways in which industry analysts acquired information in detailed interaction with vendors and adopters, produced assessments of the market and presented these outputs to clients.\nHow industry analysts characterised their expertise: How industry analysts distinguish themselves from management consultants\nStrategic choices underpinning the distinctive form of industry analysis expertise: We have explored the strategic choices pursued in particular by Gideon Gartner that underpinned the emergence of this new category of expert. Though the methods of knowledge production employed by industry analysts embody some elements that resemble conventional technical specialists, they also differ in some striking ways. For example, our research has drawn attention to the increasingly elaborate and formal methodologies used by Gartner for the production of key knowledge outputs like its Magic Quadrants, together with increasingly systematic internal regulation within Gartner of the production of these public outputs, which go through an internal \"peer review\" process (Pollock and Williams, 2009). Though these may resemble the knowledge creation and governance systems of medical and scientific professionals, the analysts themselves describe their knowledge production as pseudo-academic rather than scientific (interview, Thomas). Thus Gartner outputs are subject to internal quality control within the firm (e.g. through standardised methods and internal review) rather than the peer review by the external professional community that underpins the verification of scientific knowledge. We see these features as closely linked to the need for industry analysts to create an audience for whom they are expert (Turner, 2001). In other words, we see industry analysts exemplifying a different kind of expert to the established groups that Sociology of Professions has tended to focus upon which have already achieved wide acceptance of the efficacy of the knowledge base and methods. This exigency pervades much of their activities.\nConclusions: It would appear that the recently emerged role of IT industry analyst offering advisory services drew elements from a number of domains of practice, including conventional quantitative and market research, academic scientific research as well as others including journalism, financial analysis and marketing.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to explore conceptual issues arising in an empirical study of the emergence of a distinctive new form of expertise - of industry analysts and in particular the leading firm Gartner Group that exercises enormous influence over the Information Technology (IT) market.\nDesign/methodology/approach: The paper critically reviews existing analytical frameworks and especially work from the Sociology of Professions. This has largely focused upon groups which have already succeeded in gaining wide acceptance of the effectiveness of their methods and knowledge. For emerging expert groups a key challenge is to create an audience for whom they are expert (Turner, 2001). The study contributes to a \"third wave\" of studies that shift the focus of enquiry from the operation of professional institutions to the conduct of expert work - and how knowledge is produced, validated and consumed. The paper draws upon an extended ethnographic study of Gartner Inc., and other industry analysts to characterise some key features of their expertise. Data sources include over 100 hours of participant observation of industry analysts and their interactions with vendors and technology adopters at IT industry conferences; interviews with over 20 industry analysts from Gartner (including a telephone interview with its founder Gideon Gartner) and other analyst organisations; a substantial body of interviews with technology vendors and clients (particularly in relation to the Customer Relationship Management technology sector); together with a review of Gartner documentation and reports.\nFindings: The paper compares the empirical findings of industry analysts with accounts from current literature on management consultants and other groups such as journalists and financial analysts. Industry analysts, like consultants, have not sought to follow a classical professional model. Thus the brand reputation of big (industry analyst or consultancy) firms provides an alternative warrant of the quality of expertise to professional institutions. However, Gartner analysts identify differences as well as similarities between their work and management consultants. Gartner's ability to rank the offerings of IT vendors requires them to adopt formal methodologies and internal review procedures to produce defensible knowledge and demonstrate their independence. Industry analysts need to establish cognitive authority over rapidly changing technological fields. This imparts some \"public good\" elements to their knowledge.\nOriginality/value: The paper suggests ways forward for analysing new forms of knowledge intermediary in business and accounting, applying perspectives from the \"third wave\" of studies, and involving detailed study of the \"epistemic systems\" through which such knowledge is produced, consumed and validated (Knorr Cetina, 2010).\n", "role": "assistant" } ]
Employee turnover: features and perspectives
[ "Turnover" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: We want to retain our best people. The aims of this viewpoint are to provide the manager or supervisor with some research-grounded information regarding why people tend to remain in an organization, and possible actions a manager can take to influence talented employees to stay. Managers normally have considerable leverage in the leave or stay rationale.\nWhy it Matters; What it Costs: Turnover matters as much of it is represented by undesirable consequences for the organization, the manager, and/or co-workers. Not all turnover is negative or dysfunctional as the departing employee may not have been a consistent, high-level performer; a vacancy may provide opportunities for promotion for remaining employees or acquisition of new, high-level performers.\nInformation that Shapes This Viewpoint: When it comes to employee turnover one size does not fit all, that is, there are many factors that can influence an employee to think about and then, perhaps, leave an organization. The same reasoning holds for the reverse -- there are specific factors that influence an employee to stay, even when pay or other critical employment features are not entirely positive.\nSome Suggestions for the Manager: These suggestions are intended for an individual manager to think about and practice - right now. In some instances manager input may have little influence in retaining talent as many policy matters in organizations are complex and far reaching (e.g. salary structure, reward system, promotion policy).\n", "role": "user" }, { "content": "Purpose: This viewpoint provides the individual manager or supervisor with foundational information in order to better grasp the critical aspects of employee interest in leaving or staying on the job, and some suggestions for influencing talented employees to stay with the organization.\nDesign/methodology/approach: Design involves two distinct features. First, a comprehensive review of empirical research on the topics of turnover and turnover intention helps to identify and then summarize the most agreed-upon characteristics that define employee turnover. Second, relying upon published reports and empirical research the intention is to identify those interventions a manager or supervisor may undertake to counter turnover among valued employees.\nFindings: The reviews undertaken led to several motivators or stimuli that encouraged an employee to leave or stay with an organization. It was discovered that turnover is much more complex than matters relating to job satisfaction, compensation, or an undesirable boss. The various motivators for leaving or staying present several opportunities for managerial intervention aimed at retaining talented employees.\nOriginality/value: The value of this viewpoint is the concise presentation of the critical aspects of a work environment that influence an employee to stay or leave an organization. These aspects are expressed such that a manager or supervisor is guided about actions to take to enhance the retention of desirable employees.\n", "role": "assistant" } ]
Communities of practice as stimulating forces for collective learning
[ "Learning", "Knowledge sharing", "Group work" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: In today's world, organizations have to learn constantly, in order to respond flexibly to changes in the environment and to stay competitive. For becoming or sustaining a learning organization, management has the task to make the best possible use of the knowledge of the workers in the organization, as it is the human capital that holds the most valuable potential for organizational learning (De Laat and Simons, 2002; Simons and Ruijters, 2001).There is a growing consensus that the best way to improve organizational learning is not to (simply) focus on capturing, codifying and documenting knowledge of individuals, but rather to concentrate on ways through which knowledge can be shared, discussed and innovated. With the shift of attention to work-related learning, informal ways of learning have also increased in value in the opinion of educational researchers as well as HRD-practitioners (Bolhuis and Simons, 2003; Eraut, 2000). The problem is that those implicit or informal learning processes are difficult to detect, because people do not consciously recall and perceive this learning, and it is difficult to evaluate the outcomes. Implicit learning often results in tacit knowledge, which is context-specific, personal and difficult to communicate. Tacit knowledge nevertheless can be very powerful in learning and innovation processes (Nonaka and Takeuchi, 1995). In order to see what is learned so that one can share and evaluate it, it is important that tacit knowledge is made explicit. By doing so, knowledge becomes available and negotiable for others, a kind of social experience.Several scholars have addressed the importance of groups where workers can share and develop knowledge together in or about their professional practice. The theory on communities of practice (CoPs) is a known example of this. Lave and Wenger (1991, p. 98) defined a CoP as \"... a set of relations among persons, activity, and world, over time and in relation with other tangential and overlapping CoPs\". Later Wenger extended this idea and defined CoPs as groups of people who share a concern, a set of problems, or a passion about a topic, and who deepen their knowledge and expertise in this area by interacting on an ongoing basis (Wenger, 1998).In the last few years, the term CoP has been extended to encompass new meanings that were not part of Lave and Wenger's original idea (Hildreth et al., 2000). As Bood and Coenders (2004) argue, one can speak of a paradox. Lave and Wenger (1991) describe CoPs as groups that develop and act through self-organization - CoPs are groups that emerge naturally. A lot of organizations, however, create CoPs as a knowledge management tool to support or stimulate learning in the organization. Indeed, CoPs are commonly used and appear in several types and varieties to enhance knowledge exchange and organizational learning. And the term CoP is applied to a wide range of groups, from project teams to functional departments. The question could be raised to what extent these (intentionally created) groups actually are CoPs.There is however no diagnostic or management instrument available, yet, that enables making judgements about the actual functioning of groups as a CoP. In this article, a first attempt will be made to build a framework that helps judging groups for several aspects of their functioning. We will report research that aims at providing insight in the way CoPs function by framing CoPs as described in literature and compare this to case study insights of groups of workers as they act in practice. The following research question is leading: what are the characteristics of communities of practice, and when are they stimulating for collective learning?The learning taking place within groups will be evaluated by examining what characteristics of CoPs these groups have, how they learn collectively and what mechanisms stimulate or hinder learning. Although it is recognized that all learning involves the level of the individual at some point, the research presented in this article focuses on the collective learning processes in these groups, since collective learning is seen as a strong mechanism for organizational learning (Dixon, 1999; Nonaka and Takeuchi, 1995). After presenting the research, the framework as used in the study will be discussed for its potential usefulness for managers to facilitate groups in their organization towards CoPs structures that stimulate organizational learning.\nTheoretical framework: Characteristics of CoPs\nResearch model: Based on these theoretical notions, a research model can be constructed for a study of the learning function of a CoP. This model contains of three elements:1. group characteristics (communities of practice);2. externalization process (collective learning); and3. collective learning outcomes.As explained in the theoretical framework, the variable group characteristics is in this research specified by the theory of CoPs. In this model the process of externalization, as described by Nonaka and Takeuchi, defines collective learning. The ideal collective learning outcomes are assimilated, as they are an important factor of collective learning and the relation to organizational learning (see Figure 1).The validity of this model regarding the review and judgements of the learning function of CoPs in practice will be examined by analysing case study data on three groups in organizations.The following research questions are leading:1. Do the three groups function as CoPs?2. To what extent do processes of externalization take place?3. To what extent does this lead to collective learning outcomes?Research method\nThe three cases: The teaching department\nInstruments: We draw upon existing research material that is collected in a previous research project (Gielen and Hoeve, 2002). The existing data provided much information on the characteristics of the group and the social processes present, but less information on collective learning processes. In a follow up research, Mittendorff (2004) studied collective learning processes that are taken place in the three cases.The data is collected between October 2001 and March 2002 and additional data in the Spring of 2003. In the first round, data were collected at three levels:1. On individual level through depth-interviews, concerning learning activities, problem solving, working and learning effects.2. On group level through group conversations, concerning learning processes, professional identity, norms and values. Also, data were collected through observation of group meetings.3. On organizational level through document research and conversations with HRM-managers of the participating organizations. The focus lied on the knowledge management in the organization.In the second data collection round, individual semi-structured interviews were held with a core member of the group for each case, focusing on group characteristics, externalization processes and collective outcomes.\nAnalysis: The collected data were analyzed in two steps. First the existing data was analyzed with the framework described in the research model. The first author read the transcripts of the interviews and observations from the first data collection round, and assigned codes to fragments corresponding to the \"sub-characteristics\" from the research model. For example: \"We talk in jargon all the time. A stranger would not understand much of it.\" was coded as \"shared practice\" or \"It's difficult to discuss a subject that everyone finds interesting. Some people are strongly focused on the real traditional agriculture, although we decided that knowledge circle focuses on countryside renewal.\" as \"no shared domain\". The coder wrote memos during this process as a means of immediately recording all thoughts and difficulties relating to the coding process to be used in the next procedure of consensual validation.The coded fragments were summarized and displayed in matrices to create an overview. The displays were discussed critically with the third author (the researcher that collected and analyzed the data of the first data collection round). Differences in the two analyzes were discussed and where differences were identified, the relevant parts of the transcripts and memos were read again and discussed extensively in order to reach consensus.This first step resulted in a description of how the cases function according to the three elements in our conceptual research model. Based on these case reports it was decided for which variables additional data need to be collected.In the second step the additional data was coded in accordance with the \"sub-characteristics\" from the research model. These coded fragments were used to update the displays. Again these displays were discussed critically with the third author. This discussion resulted in a more detailed description of how the cases function according to the three elements in our conceptual research model. These within-case reports were aimed at providing a clear view of the three cases separately. The cross-case analysis involved a comparative analysis, in order to deepen our understanding and explanation of the differences in results between the three cases.\nResults: The teaching department\nConclusions and discussion: This article aimed at understanding communities of practice (CoPs) as stimulating forces for collective learning. Three working groups were analyzed with three research questions in mind, namely how the three groups function as a CoP, to what extent processes of externalization take place, and to what extent these processes lead to collective learning. By reflecting on the results we now try to answer our research questions and discuss these findings.Do the three groups function as a CoP?\n", "role": "user" }, { "content": "Purpose: - The purpose of this research is to get a clear view on how can we judge groups in relation to the characteristics of a community of practice (CoP), and the presence of collective learning in these groups.\nDesign/methodology/approach: - A review of literature on collective learning and CoPs led to the development of a conceptual model, which was tested through case study research against empirical data from three groups in organizations.\nFindings: - The groups differed concerning group characteristics, but also concerning the collective learning processes and learning outcomes present. The group that can be characterized as a CoP learns a lot, but the (learning) processes in the group are not always in favour of the organizational learning process.\nResearch limitations/implications: - The conceptual framework was helpful to evaluate the characteristics of CoPs in relation to collective learning. These findings suggest that it will be interesting to expand the model, for example with consideration to the way CoPs experience the need to change.\nPractical implications: - The developed framework might help managers to judge if groups in an organization have characteristics of a CoP, if they are in balance and what might be needed to develop towards an ideal CoP with a great learning potential.\nOriginality/value: - A first attempt is made to build a framework for judging CoPs for several aspects of their functioning. The research also shows that CoPs are not always stimulating forces for organizational learning.\n", "role": "assistant" } ]
Replacing the ideology of information by exploring domains of knowledge: A case study of the periodization of Philippine history and its application to information studies
[ "Philippines", "Asia", "History", "Wikipedia", "Domain analysis", "Wikis" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: In an article written in 1995 Phil Agre argues that we need to view the concept of information as an \"object of certain professional ideologies, most particularly librarianship and computing [...]\" (p. 225). It serves \"to position librarianship as a neutral profession\" (p. 226) in so much as it draws attention away from the internal debates and ideologies of any of its user communities. He notes that if one considers the world of academia, for example, it is not information that is focus of attention of its seasoned practitioners, but literatures. Up to now, particular user communities have generally remained content \"to treat the work of exploring literatures as a series of discrete problems to be reformulated in the language of information\", but Agre suggests that this might change as digital technologies increasingly mediate what he refers to as \"the institutional circuitry\" of user communities and make possible multiple approaches to the delivery of what to date we call information. In such a situation \"it may become possible - and perhaps even unavoidable - for librarians to abandon the idea of information and replace it with the specialized ideology that governs the circuitry of a particular institution\". In doing so librarians or information workers will have to deal with the consequences of the abandonment of the neutrality provided by the ideology of information: \"If librarians attempted to organize works in the ways their patrons orient to them [...] certain difficulties would follow. It would be necessary to make explicit some frequently contested matters, such as who founded the literature, which research groups are dominant [...] Librarians would find themselves effectively positioned as participants in the disciplines' conflicts but without the disciplinary standing needed to make their views stick\" (p. 226).\nThe domain of Philippine historiography: The rest of this paper examines in a preliminary and partial way the domain of Philippine historiography as viewed by Filipino historians. Domains are made up of a combination of ontological, epistemological, and sociological theories and concepts (Hjorland and Hartel, 2003), but here the focus is on the ontological. Hjorland and Hartel (2003) define ontological theories as descriptions or explanations of reality \"and how it is structured\" (p. 239). It deals with the objects or topics or elements of disciplinary inquiry. Periodization, or how time is divided and labelled, is central to the ontology of historiography. Hence developing an understanding of the richness of the periodization schemes used by Filipino historians is the first goal of this paper. The second is to apply this understanding to the development of online information resources in a way that illustrates how certain professional historians see their field and by implication demonstrates the plurality and complexity of historical knowledge production. In this aim, the paper seeks to advocate a project based on the results of domain analytic research for the \"sceptical librarian\" interested in moving beyond the ideology of information.\nPutting an understanding of a disciplinary domain to use in LIS-related activities: Domain analysis is important as an academic contribution to the construction of LIS as a distinct discipline, as well as being an inherently stimulating approach to the study of fields of knowledge, but it also needs to serve practical purposes as well. Luckily it is not difficult for it to do so. Hjorland describes 11 approaches to domain analysis, among which are activities such as producing guides to the literature and subject thesauri which are very practical indeed (Hjorland, 2002a). Domain analysis can also be useful in designing information systems and among these systems Wikipedia is not in any way the least important to our contemporary information infrastructure.\nConclusion: At the beginning of this paper I noted the view of Phil Agre that at some point in the future the possibilities of digital technology may require librarians to rethink their reliance on the ideology of information and that although he was right to point out the difficulties such a move would entail, Wilson's concept of the sceptical librarian and Hjorland's programme of domain analysis provides means to move forward. In the case study that followed I provided a small contribution to the analysis of a particular domain; namely, some of the ontological structures that underlie the periodization of Philippine history, and an application of that analysis to the world of Internet content provision and online learning tools for an age in which the ideology of information should have less relevance.\n", "role": "user" }, { "content": "Purpose: This paper examines the structure of Philippine historiography as viewed by Filipino historians. The purpose of this paper is to develop an understanding of the knowledge domain of Philippine history and in particular how its practitioners organize their field of study in terms of periodization. At the end of the paper an application of this analysis is proposed, the development of an online encyclopaedia of Philippine history.\nDesign/methodology/approach: Interviews were arranged with willing historians at two of the premier institutions of higher learning in the Philippines: the Ateneo de Manila and University of the Philippines. The historians were asked three general questions: what in their opinion, are the key defining events in Philippine history? What are the key historians for each of those events? And what are the key debates regarding these events? For the purpose of this paper it is the results of the first question that are in focus as it deals with one of the fundamental tools of historical analysis, periodization.\nFindings: Philippine history was found to be periodized in a variety of ways, from the traditional to other approaches that stress either Filipino rather than colonial agency or the uneven trajectories of historical development that depend on region, class, or language group. A final approach viewed Filipino history as a network of relations spanning space and time. Wikis designed around the results of domain analysis make it possible to provide information on topics of importance to a discipline as well as reveal something of its deeper structure. Combined with traditional concerns, such as use of appropriate sources, this would serve to help develop a deeper awareness of the nature of knowledge production.\nOriginality/value: This paper represents both a contribution to the study of knowledge domains, as well as an application of that study to the work of information professionals. Putting the spotlight on Philippine historians and history also helps the LIS discipline to move away from its traditional North American and European focus. Studies of knowledge producing bodies in the rest of the world are important and overdue.\n", "role": "assistant" } ]
Individuals, teams and organizations score with gamification: Tool can help to motivate employees and boost performance
[ "Motivation", "Employee engagement", "Gamification" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Employee engagement is dubbed a recipe for individual and organizational success and, therefore, organizations across the globe are investing in it.\nDefining gamification: Gamification is more than just a game. It is the use of game elements and design in non-game contexts. It helps in igniting and sustaining connections with employees. When crafted intelligently and innovatively, it can help in infusing interest and incentivizing learning. Gamification can help in employer branding by shaping an organization's corporate identity and reputation.\nExperiences of gamification as a driver for employee engagement: The thrill of game is what makes gamification work. Mollick and Rothbard (2013) provided an influential logic to this. It was rooted in the idea that work is not always fun, games are fun, so turning work into a game will make work fun.\n", "role": "user" }, { "content": "Purpose: Explains how gamification can help to motivate employees and boost individual and organizational performance.\nDesign/methodology/approach: Defines gamification and provides some examples of how it can work in practice.\nFindings: Charts how gamification helps to engage employees by infusing vigor, making them dedicated and enabling them to be absorbed in their work.\nPractical implications: Explains that gamification can help in infusing a feel of ownership of performance and results.\nSocial implications: Outlines some of the social forces that can help to make gamification a success.\nOriginality/value: Reveals that through gamification - designed sensitively, crafted creatively and implemented positively - organizations can target burn-out and help to foster an engaging work environment.\n", "role": "assistant" } ]
Rethinking value proposition tools for living labs
[ "Value", "Value proposition", "Living labs" ]
[ { "content": "Summarize the following paper into structured abstract.\n\n1. Introduction: As living labs have become instruments for user involvement in various development and research activities, the academic interest in living labs has increased. Interest has mainly focussed on the methodologies used in living lab settings during the co-creation and testing phases (e.g. Almirall et al., 2012), whereas tools developed as part of the commercialisation of the innovation have received less academic attention. This is against a backdrop where living labs are seen as a means for strengthening European competitiveness and growth (Peltomaki, 2008). Alongside the increase in usage of living labs in user involvement and co-creation, active theoretical discussions have taken place around value and value co-creation, especially the new theoretical perspectives opened up by Vargo and Lusch (2004) in their seminal work on service-dominant (S-D) logic. This paper further details how theoretical discussions around S-D logic connect to living labs, by focussing on the one hand on the value concept and on the other on managerial tools used to build value propositions in conceptual analysis.\n2. Living labs as innovation platforms: Although researchers have had lively discussions about living labs for a decade, there is no consistent definition for the concept (see e.g. Schuurman et al., 2012a; Leminen, 2013 for reviews of the definitions). However, scholars have agreed on the main characteristics of living labs: users as co-creators in innovation processes, experimentation in a real-life context, and multi-stakeholder involvement.\n3. The nature of value: A conceptualisation of value is at the heart of discussions on the S-D logic, being present in several of the grounding premises. Elaborations on the meaning of value and the locus of its creation have emphasised either use or exchange value interchangeably, and respectively the consumer's/user's or producer's role in its creation (Vargo et al., 2008; Vargo and Lusch, 2011; Ng and Smith, 2012; Honkanen, 2014). Key perspectives to emerge have centred on value-in-exchange, value-in-use and value-in-context. Value-in-context has been analysed as value in a social context and value in the context of networks.\n4. The value proposition concept: Building on the deepened understanding of value, especially value-in-context, we shall next further elaborate the discussion on building value propositions. We shall first give a short overview of definitions on value propositions, followed by value propositions with value-in-context perspective, concentrating in particular on value propositions in service ecosystems and building on the work of Frow et al. (2014).\n5. Three tools for building value propositions - research design: In this section, we introduce three tools for building value propositions: the Value Proposition Builder(tm) (VPB) by Barnes et al. (2009), the Value Proposition Canvas (VPC) by Osterwalder (2012) and the People Value Canvas (PVC) by Wildevuur and colleagues (2013). The inclusion criteria for these tools were popularity of usage of the tools, as well as good linkages to the topics of this manuscript: value, value propositions and living labs. The tool by Barnes et al. (2009) was chosen because it builds on the seminal pioneering work of Lanning (1998) and thus brings into the analysis aspects from the early discussions; moreover, it is widely available through Wikipedia, where it is described as a tool for building value propositions, and thus can be seen as well known. The VPC has rapidly gained high attention as it builds on the well-known framework of the Business Model Canvas (Osterwalder and Pigneur, 2010). Osterwalder (2012) suggests working with the canvas in an iterative manner, revising the value proposition after experimenting and testing, and thus the tool is aligned with the process-like nature of product and service development in living labs. The PVC represents the user-centric innovation approach, and the tool itself was developed in the context of living labs.\n6. Discussion: The findings from the conceptual analysis of the value proposition tools are summarised in Table I. Next, we shall elaborate on each of the findings for each theme in the analysis.\n7. Summary and conclusions: To summarise, we wish to emphasise that value and value creation are understood as interactional and contextual and are shaped by social forces and norms, we well as being influenced by a wider actor network or ecosystem. All social and economic actors (including users) are seen as resource integrators, implying that consumption is about integrating resources acquired from different sources. Hence, acknowledging the importance of context and the key role of the service ecosystem is vital when developing managerial tools for building value propositions in living labs. Furthermore, we wish to highlight that the defining elements of living labs are aligned with the premises of the service-dominant logic.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to provide a conceptual analysis on value proposition tools to be used in future empirical research and in building managerial insight. The conceptual analysis focusses on a living lab framework and recent theoretical developments around the concept of value that are reflected in the context of three managerial tools for creating value propositions.\nDesign/methodology/approach: Using abductive reasoning, the descriptions of the tools were analysed as cultural texts, as language-in-use in a social context.\nFindings: In the context of the living labs approach, the Value Proposition Builder(tm) (VPB) seems to conflict with the ideas and premises of user-centric innovation processes. In the Value Proposition Canvas (VPC), the co-creation aspect is rather vague, as the enterprise and its offerings are presented as creators of value for the customer. Thus, this tool somewhat contradicts the living lab approach. The People Value Canvas (PVC) is aligned with the service-dominant logic and the premises of living labs. However, all three tools largely neglect a deeper acknowledgement of the role of the wider context, the service ecosystem, and the role of networked actors as resource integrators. Moreover, none of the tools explicitly point out the role of enterprises as intermediaries in constructing invitations for value co-creation.\nOriginality/value: The paper contributes to the SDL and living labs literature by conceptual analysis on different value proposition tools; the VPB(tm), the PVC, and the VPC which are relevant for academics as well as practitioners creating new understanding and insights on the connectedness of the living labs framework and SDL as well as their relationship to managerial tools. By identifying the absent elements of S-D logic from managerial value proposition tools, the paper contributes to current discussions by giving attention from scholars towards investigating managerial tools and by providing a new conceptual analysis for future empirical research. The critical analysis of the managerial tools contributes to managerial practice by emphasising the need to consciously evaluate the benefits and failures of tools for developing their organisations.\n", "role": "assistant" } ]
Client perceptions of barriers to partnering
[ "Change management", "Partnership", "Procurement", "Perception", "Construction industry", "Sweden" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: In many countries, the construction industry has, over a long period, attracted criticism for its relationships, with conflicts and disputes, poor collaboration, lack of customer focus and end-user involvement cited as significant amongst its shortcomings (Latham, 1994; Egan, 1998; Ericsson, 2002; Chan et al., 2003). These input factors lead directly to inefficiencies in outcomes such as time and cost overruns, and low productivity, quality and customer satisfaction. Practitioners, researchers and society at large have, therefore, called for a change in attitudes, behaviour and procedures in order to increase the chances of project success and an improved end product (Love et al., 2000; Dubois and Gadde, 2002). Generally, increased collaboration between project actors is argued to be a suitable antidote for many of the industry's problems. Since traditional procurement procedures and contracts offer little incentive for cooperation to emerge, these formal processes are potential root causes for the opportunism and lack of cooperation that characterise construction projects (Cheung et al., 2003). Perhaps as a consequence of this criticism, the interest in more collaborative relationships, mostly referred to as partnering, has increased in recent years as an alternative approach to traditional procurement and project governance. Partnering aims to increase cooperation and integration between the actors by building trust and commitment whilst decreasing disputes. Studies show that it can be beneficial for all parties involved, bringing about advantages in the areas of quality, safety performance, sustainability, dispute resolution, human resource management, innovation, as well as time and cost reductions (Barlow et al., 1997; Egan, 1998; Chan et al., 2003, 2006).Partnering is not, however, so easy to implement successfully. It implies a \"paradigm shift\" (Larson, 1995) and a fundamental change of behaviour and attitudes for all actors involved (Kululanga et al., 1999; Rhodin, 2002). Such a change can be difficult and several studies have shown that implementing partnering can be problematic (Bresnen and Marshall, 2000a; Ng et al., 2002; Bresnen et al., 2005). As clients are in the position of leader and head facilitator of the partnering arrangement, it is important that they implement partnering in the correct way and for the right reasons (Ng et al., 2002). Increasing awareness of partnering implementation should allow clients to appreciate the kinds of barriers faced and, thus, how procurement procedures might be modified to overcome them. The purpose of the research reported here is to identify critical barriers to partnering perceived by construction clients and then to investigate measures that are taken to overcome them during implementation. Empirical data have been collected in Sweden through a survey of construction clients.\nLiterature review: barriers to partnering: The purpose of the literature review was to identify and describe potential barriers to partnering and also to formulate hypotheses about how to overcome these barriers by suitable procurement procedures and project management. Through the literature review three main types of barriers were identified: cultural, organisational and industrial (Post and Altman, 1994; Childerhouse et al., 2003), along with six hypotheses reflecting the means to overcome them.Cultural barriers\nSurvey study: barriers to partnering: The purpose of the survey study was to investigate if the potential barriers identified in the literature review were considered critical by construction clients and if the grouping of the barrier items into three main types were valid. Additionally, correlation between the perceived barriers and the clients' procurement procedures were measured in order to test the hypotheses formulated in the literature review. The survey was, first piloted by five respondents, resulting in only minor changes. The total sample investigated was 104 construction client organisations, all of which are members of a national construction client association (Byggherrarna) that represents the majority of professional construction clients in Sweden. The members are regional, national or international industrial and property companies, municipalities and regional authorities, and government agencies, which procure construction work involving civil engineering, industrial facilities and housing.Registered contact persons in these organisations were approached by email or telephone in order to ask if they or another more suitable person in their organisation were willing to participate in the study on account of their organisation. Four people declined to participate at this stage, due to lack of time. A paper version of the survey was then sent out by mail to the 100 people who had agreed to participate. These people were mostly procurement managers, project managers or directors of the construction and facilities departments in their organisations. Hence, they were responsible for procurement decisions. After two reminders, a total of 87 responses were received, representing a response rate of 84 per cent. A test for non-response bias was conducted by comparing major variables for early and late respondents. This is because several non-responses were explained by the intended respondents' lack of time. No significant differences were found between these two groups in this test.In this paper, the empirical data only are discussed in the context of the importance of barriers and different aspects of the organisations' procurement procedures. The respondents were requested to rate potential barriers in a list of suggested barriers, derived from the literature review, using a seven-point Likert scale where 1= no barrier and 7= critical barrier. Furthermore, they were asked how often they used different procurement procedures (e.g. to what extent do you use the collaborative tools listed below during the duration of construction projects? 1= very seldom and 7= very often). In this way, correlations between perceived barriers and procurement procedures could be investigated.Factor analysis and construct reliability\nConclusions: The survey study supports the literature review in dividing the barriers to cooperation into three different constructs: cultural, organisational and industrial barriers. The only barrier item that did not support the grouping suggested in the literature review was labour unions, which loaded on the cultural factor rather than the industrial factor. This result suggests that labour unions are important actors since they influence the (conservative) culture in the construction industry. This is in line with Craft's (1991) finding that labour unions are generally very conservative and reluctant to change. The survey results show that cultural barriers (i.e. conservative culture, adversarial attitudes, short-term perspective, lack of sub-supplier involvement in specifications and labour unions) constitute critical barriers to change towards increasing cooperation. Also, organisational barriers (i.e. traditional organisation of the construction process, new competence requirements and traditional procurement procedures) are considered critical. Industrial barriers (standard contracts and laws and regulations) are not, however, considered critical.The regression analyses did not find any strong and significant correlations between the perceived barriers to cooperation and clients' procurement procedures, as hypothesised in the literature review. All hypotheses were rejected. It seems, therefore, that clients' procurement procedures are not affected by their perceptions regarding critical barriers to cooperation. When hypotheses of moderate strength are rejected, it is something interesting that should be investigated further (Weick, 1989). This study also shows that two-thirds of the construction clients perceive increased cooperation among the actors in construction projects to be more important than competition in order to facilitate project success. Hence, they should be interested in overcoming the identified barriers through purposeful procurement procedures. Even so, neither their wish to increase competition nor cooperation in the construction process was found to affect their procurement procedures. These results are disappointing, but helpful. The inference is that while (Swedish) construction clients wish to increase cooperation in the construction process, this desire does not affect their project management and procurement procedures, which are still focused on competition. This result is also interesting from the perspective of Weick's work and worthy of further investigation.Earlier research has shown that, when implementing partnering in the supply chain, suitable procurement strategies are very important (Akintoye et al., 2000), since they influence the degree of cooperation and supply chain integration and, ultimately, the overall success of the project (Briscoe et al., 2004). The empirical results in this paper are therefore troublesome, since clients' desire to increase cooperation is difficult to achieve without changing their procurement procedures. The negative results in this study may arise for different reasons. Akintoye et al. (2000) and Saad et al. (2002) reached similarly \"negative\" findings in empirical studies regarding the implementation of partnering in the supply chain. It was found that the actors perceived it important and beneficial, but they lacked an understanding of the concept and the prerequisites associated with its successful implementation. This may also be the case in this study; clients may be unaware of how their procurement procedures affect their chances of creating a cooperative climate. This unawareness may be derived from the clients' lack of practical experience and theoretical education regarding partnering. At the time of the survey (Spring 2006) the use of partnering was fairly limited in Sweden. Hence, there were no strong industry norms advocating partnering. Such lack of expectations for using partnering may be an important reason for the absence of partnering (Phua, 2006). In order to increase understanding of partnering implementation and to affect industry norms (i.e. promote the message of partnering), the respondents' association, Byggherrarna, has launched several partnering courses for construction actors in general and their member organisations in particular. This initiative is a potentially useful way for increasing the likelihood of wider partnering implementation in the Swedish context in the absence of governmental policies and guidelines. Moreover, in countries in which industry norms of partnering exist (e.g. through governmental policies) there may also be a need to increase understanding of how to interpret policies and implement partnering.Another reason for the contradictory results may be that clients are simply unwilling to implement a change that involves partnering (Bresnen et al., 2005, Fernie and Thorpe, 2007). The incentives for the individual project/procurement manager to start using new and less familiar procurement procedures may not be strong enough. Lack of top management commitment and conservative culture, for example, may deter individual decision makers from trying new approaches even though they are potentially more suitable than traditional approaches. These and other potential explanations would be worth investigating in future research. Qualitative research in close collaboration with construction clients should be possible in order to investigate how clients can overcome critical barriers to partnering through purposeful procurement and project management procedures. Similar research in other countries would also be of great interest in order to find out if industry norms of partnering affect clients' procurement and project management procedures, as Phua's results indicate.\n", "role": "user" }, { "content": "Purpose: - The purpose of this paper is to identify critical barriers to partnering, as perceived by construction clients, and the specific measures that are taken to overcome them during implementation.\nDesign/methodology/approach: - Empirical data were collected through a survey study of 87 professional construction clients in Sweden.\nFindings: - Clients regard the most critical barriers as those attributable to cultural and organisational aspects. The analysis also shows that clients' perceptions of these barriers do not, in fact, affect their procurement procedures. Two-thirds of clients in the survey wish to increase cooperation with actors in the belief that it will favour project success. Their intention does not have any bearing on their procurement and project management procedures, which are still aligned to competitive bidding. Two potential reasons for this inconsistency are discussed: clients may be unaware of how their procurement procedures affect cooperation, and/or the individual decision maker may not have strong enough incentives to start using new and less familiar procurement procedures even though they are potentially more suitable than traditional procedures.\nResearch limitations/implications: - The quantitative data are limited to clients' perceptions of barriers to partnering; a contractor perspective is not included in the survey.\nPractical implications: - The research results can serve as an alert for construction clients that their procurement procedures need to be adapted if they want to achieve the move towards increased cooperation that they say they do.\nOriginality/value: - This paper offers a unique analysis of the correlations between desired outcome in the form of increased cooperation, and actual behaviour in the form of procurement procedures.\n", "role": "assistant" } ]
Using a storytelling intervention in schools to explore death, dying, and loss
[ "Public health", "Storytelling", "Children", "Death", "Dying", "Bereavement", "End of life care" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Death is often a source of fear and anxiety and is generally avoided as a topic of conversation (Kellehear, 2014). In the UK, 1 in 29 children (approximately one child in every class) will be bereaved of a parent or sibling, while 1 in 16 will be bereaved of a close friend (Penny and Stubbs, 2015; Winston's Wish, 2020). These children are at increased risk of poorer mental and physical health, lower educational attainment, substance misuse, criminal activity and early mortality (Guldin et al., 2015; Kaplow et al., 2010; Stikkelbroek et al., 2016). Bereavement in childhood may therefore disrupt the trajectory expected in healthy development. Compassionate Communities, a public health approach to palliative and end of life care, aims to reduce the impact of these negative physical, social, psychological and spiritual outcomes by improving community participation and local support networks (Wegleitner et al., 2015).\nMethod: Context\nResults: Entries\nImplementation outcomes: Acceptability and appropriateness\nDiscussion: The use of storytelling as a structured intervention to explore death, dying and loss in schools demonstrated feasibility. Further exploration of ways to encourage wider participation is needed. There was a risk that exploring ideas surrounding the end of life would not stimulate interest, or potentially cause upset. However, the participatory nature of the storytelling exercise enabled active engagement and creative exploration. The intervention also appeared to prompt difficult but beneficial family conversations about recent bereavements, as well as conversations about end of life care preferences and wishes. In one report, involvement with the project appears to have had a positive influence on how a family responded to and supported a dying family member. By enabling these discussions, the intervention may have contributed to facilitating greater psychological preparation for and resilience to death, dying and loss among families (Metzl and Malissa, 2008), lessening the risk of negative developmental outcomes (Kaplow et al., 2010; Stikkelbroek et al., 2016), and complex bereavement (Wegleitner et al., 2015).\nConclusion: While it is indicated elsewhere that teachers, pupils and parents frequently feel ill-equipped to engage in conversations around death and dying; using a creative storytelling intervention to approach the topic in schools appears to be an acceptable, appropriate and feasible approach. Moreover, feedback from parents and pupils suggested the intervention had prompted important and beneficial discussions about end of life care to take place in families, which may have otherwise been avoided. The positive response to the intervention indicates that schools are a valuable community asset that could be empowered to support children and families affected by death, dying or loss.\n", "role": "user" }, { "content": "Purpose: Experiencing bereavement in childhood can cause profound changes to developmental trajectories. This paper aims to evaluate the feasibility of implementing a public health intervention in schools to encourage pupils aged 12-15 years to independently explore ideas of death, dying, loss and end of life care in a structured and creative format.\nDesign/methodology/approach: A co-produced storytelling intervention was implemented in an independent school in Norwich, UK. Pupils wrote up to 1,000 words in response to the title, \"I Wish We'd Spoken Earlier\". Their participation was voluntary and extra-curricular. Stakeholder feedback was used in addition to the submissions as a measure of acceptability, appropriateness, adoption and feasibility.\nFindings: In total, 24 entries were submitted. Pupils demonstrated their ability to engage thoughtfully and creatively with the subject matter. Feasibility for the storytelling intervention was demonstrated. Importantly, the intervention also prompted family conversations around preferences and wishes for end of life care.\nResearch limitations/implications: To determine whether the intervention has psychological and social benefits will require further study.\nPractical implications: Educational settings can be considered as anchor institutions to support a public health approach to end of life care.\nOriginality/value: The positive response from all stakeholders in delivering and supporting the intervention indicates that schools are a community asset that could be further empowered to support children and families affected by death, dying and loss.\n", "role": "assistant" } ]
Information literacy literature in the social sciences and humanities: a bibliometric study
[ "Information literacy", "Quantitative analysis", "Bibliometric study", "Information literacy instructions", "SCImago Journal Rank (SJR)", "Source Normalized Impact per Paper (SNIP)" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Information literacy has a number of characteristics and reveals people's desire to achieve educational, social, occupational and economic goals (Crawford and Irving, 2009; Lloyd, 2005; Lloyd and Williamson, 2008). Zurkowski (1974), inventor of the term information literacy, observed that an information-literate person knows how to apply information resources to routine work. Information literacy focuses on critical thinking and helps to search, evaluate and use information adequately. The SCONUL (2011) working group on information literacy marked out seven pillars of the literacy landscape, and noted that an information-literate person identifies information, defines the scope, plans, gathers, evaluates, manages and presents the information in a consolidated form. Students in the Social Sciences and Humanities encounter problems in identifying information needs for assignments, evaluating resources and producing quality work that comes up to the expectations of their teachers. The concept of information literacy needs to be understood properly so that its advantages can be recognized (Head, 2008; Balanskat et al., 2006; Aharony and Bronstein, 2014). A large number of publications on information literacy have come out in the past 30 years (Rader, 2002).\nLiterature review: Since the inception of the concept, research on information literacy has grown exponentially. A large number of papers have been published about information literacy competencies, imparting instructions in information literacy programs and allied areas. Nevertheless, only a few studies have conducted a quantitative analysis of literature in the field. A bibliometric study was conducted using the Web of Science (WoS) data set, focusing on global research publications on information literacy in the field of library and information science, by Kumari et al. (2015). The study revealed that the American continent contributed 690 (52.43 per cent) articles, followed by the European continent with 338 (25.68 per cent). Gonzalez-Valiente (2014) conducted a bibliometric study on information literacy research in Cuba during 2000-2013. The most prolific authors on the subject were found to be affiliated with health and higher education institutions. Journals in the area of library and information science published the most research on information literacy in Cuba. Nazim and Ahmad (2007) conducted a study on information literacy research trends, using the LISA Plus database. Journal articles indexed in LISA were selected for the study. The USA was found to be the leading country, with 51.23 per cent of the total articles published in 58 journals; during 1980-1990, only nine papers were published on information literacy, whereas the year 2000 onward saw a proliferation of such publications, and 372 papers were published during the period 2000-2005. Interestingly, 63.16 per cent of the papers were published by single authors, while more than three collaborating authors account for only 2.71 per cent of the total publications. Williams and Coles (2007) suggested that lack of sufficient awareness of information literacy may be a factor in limiting the use of research information, along with the challenges of lack of time and accessibility of information sources.\nObjectives of the study and methodology: This study confines itself to the following objectives:\nLimitations of the study: The study has been conducted using papers indexed in the Scopus citations database. However, Scopus does not cover all the journals published in the area of the study. Journals published at local levels are not indexed in Scopus. Citations of documents are indexed in Scopus only after 1996. Scopus is an international-oriented database; it has 67 per cent coverage of publications from the USA, Canada, the UK, The Netherlands (Elsevier), Germany (Springer) and Switzerland (Pipp, 2006). Its coverage of journals published in Asia and Africa is lower. Another limitation of the study is that it does not cover theses, dissertations and book chapters.\nResults: The data set was analyzed from different angles. The author has attempted to group the data under different headings. Details of the results are presented below.\nFindings and conclusion: Undoubtedly, the importance of information literacy is widely acknowledged by the academic fraternity in all sectors. The results clearly exhibit that research on information literacy has accelerated at a considerable pace in recent times. The 1,990 publications covered in the study originated from 79 countries. The average number of publications per year on information literacy in the Social Sciences and Humanities during the period was ~165.8. The data set of the study was analyzed for several different aspects: the highest growth rate (106.7 per cent) was in the year 2005; the 1,990 documents under study have been cited 10,025 times up to December 31, 2013; the documents are published in 160 journals; the average citation per publication is ~ 5.0; 1,385 papers have been cited at least once; the highest number of citations (101) was in 2010.\n", "role": "user" }, { "content": "Purpose: This paper aims to map information literacy literature in social sciences and humanities published during the period of 2001-2012.\nDesign/methodology/approach: The data for the study are obtained from Scopus, accessible at www.scopus.com. Study used the Transformative Activity Index (TAI) and relative citation impact (RCI) to know the impact of most productive countries and prolific institutions. The SCImago Journal and Country Rank accessible at www.scimagojr.com/ was used to determine the SCImago Journal Rank and source normalized impact per paper.\nFindings: The study found that 1990 documents originating from 79 countries were published in this study area. These papers are published in 160 journals with an average ~12.51 papers per journal. These papers have been cited 10,025 times with ~5.0. average citations per publication. Study also found that information literacy literature is published in 16 languages and the majority of the papers are in English, 1,879 (94.4 per cent). The highest growth of publications (106.7 per cent) was found in 2005. The USA contributed the highest number, 1,035 (52 per cent) papers. Moreover, of the 15 most productive countries, three recorded TAIs >100, and 12 countries recorded TAIs <100. In all, 160 institutions worldwide have contributed in information literacy research. Study also found that maximum literature published on the subject by a single author is 828 (41.6 per cent). Universidad de Granada, Granada, Spain has produced the highest number of papers (24, or 1.2 per cent) and received 61 (0.6 per cent) citations, while University of Strathclyde has the highest RCI (~2.7) for its publications. Pintos, Maria from Universidad de Granada has published the maximum number of papers (18) that have been cited 78 times.\nSocial implications: The study endeavors to showcase information literacy research outcomes in social sciences and humanities. It involves quantitative analysis of the literature in this domain using bibliographic elements such as keywords, authors, affiliation, publication and citations.\nOriginality/value: No study has been conducted so far to map the information literacy literature in social sciences and humanities. Study will be useful in understanding the progress on information literacy in the area of social sciences and humanities. The study is significant for social scientists to foster further research in this emerging area.\n", "role": "assistant" } ]
Knowledge management practices: An exploratory study at the Malaysian higher education institutions
[ "Knowledge management", "Knowledge sharing", "Tacit knowledge management" ]
[ { "content": "Summarize the following paper into structured abstract.\n\nIntroduction: Knowledge management (KM) plays a dynamic role in the higher education institutions (HEIs) success, particularly through effective planning, organising, monitoring and coordinating the KM assets related to intellectual capital. Thus, KM could enrich knowledge sharing and overall performance (Hossain et al., 2013). HEIs are considered to be knowledge-intensive organisations (Howell and Annansingh, 2013; Ramachandran et al., 2013) and knowledge-creating institutions, and they are also considered to be in the knowledge business (Rowley, 2000). They create new knowledge through research, disseminate knowledge through teaching and learning, research and development, communication, dissemination of science, and create jobs through spin-offs (Alexandropoulou et al., 2009; Fullwood et al., 2013). In a knowledge economy, KM has been recognised as one of the determining factors for innovation and competitive advantage (Darroch, 2005; Dasgupta and Gupta, 2009).\nLiterature review: Looking back at the work of Davenport and Prusak (1998), knowledge is defined as coming from one's experiences, values, contextual information and expert insight, which then provides the basis for evaluating and incorporating new experiences and information. KM can be seen as the way that people create, transfer and adopt knowledge resources at different levels of interrelationships or networks. The work by Phelps et al. (2012) identified those network levels to be interpersonal, intra-organisational and intra-organisational. Within a construction organisation, KM is a set of tools and practices developed and implemented to gather, store, share, protect and make project-related experience and training available for use.\nKnowledge management in higher education institutions: HEIs have always been involved in KM practices. The three missions of universities, research, education and service to society are closely linked with knowledge creation, knowledge dissemination and knowledge transfer, respectively (Rowley, 2000; Alexandropoulou et al., 2009; Fullwood et al., 2013; Ramachandran et al., 2013). However, to remain competitive in the knowledge economy, HEIs need to manage their knowledge processes within the context of a deliberate KM strategy. A key prerequisite to successful KM is an awareness of the factors that promote knowledge creation, sharing and transfer in HEIs. Enablers to KM have a positive effect and the prospect of its being beneficial. Thus, there are various reasons as to why KM practices are significant to HEIs and as well the way the institutions approach KM practices that can have huge improvements in the sharing of explicit as well as tacit knowledge. Research works of KM in the HEIs aided in the comprehension of the various benefits that are brought by the application of KM in the processes of educational institutions like research, curriculum development, alumni, student affairs and services, administrative services and strategic planning (Kidwell et al., 2000). Thus, the creation of a KM system within the institutions is vital to capture, identify, transform, evaluate, disseminate and consolidate the institutional knowledge (Sharimilah Devi et al., 2008).\nMethodology: A search on the website of the Malaysian Ministry of Higher Education (www.mohe.gov.my) reveals that there are about 37 private universities, university colleges, foreign university branch campuses and about 544 private colleges currently in operation. There are also private HEIs from overseas institutions and/or affiliated with overseas institutions, for example, the UK, Australia, India and China. A semi-structured questionnaire that seeks to generate in-depth insights was deemed to be appropriate (Pearlson et al., 2016). The questionnaire was divided into two sections. Section A comprised of participant's qualification and important factors to tacit knowledge progression, and Section B questions were related to KM practices. An interview schedule was designed and piloted through meetings with three senior academicians who were experts in KM research field. The experts provided information and guidance to facilitate the study during the interview. Then, the second round of questionnaire was created and 1,100 questionnaires were distributed to polytechnics, community colleges, private universities, university colleges, foreign university branch campuses, private colleges and private HEIs affiliated with overseas institutions using personal administered method assisted by volunteers. The authors collected a total of 80 questionnaires from Kuala Lumpur (7.2 per cent response rate), 91 from Selangor (8.3 per cent response rate), 27 from Penang (2.5 per cent response rate), 38 from Malacca (3.5 per cent response rate) and 36 Johor (3.3 per cent response rate). The result was a sample of 273 individuals, that is, 24.8 per cent of the entire survey respondents. These five states were selected on the basis of the private HEIs' ranking, location and survey sample for this research. The participants were informed about the survey's objective, confidentiality, and at any time, they could decline to answer any question or withdraw from the survey. A survey type sample size calculation was utilised, meaning that a sample error formula, rather than the power analysis formulae that are usually utilised in experimental research, was used. The decision in selecting the random sample was to have a confidence level of 95% and a 10 per cent (0.10) sampling error. On the contrary, purposive sampling was used for qualitative data collection. All usable questionnaire responses were analysed using SPSS analysis software and questionnaire tool. Qualitative data analysis was achieved by identifying patterns and themes in the collected study data. The sample was checked for response and non-response bias.\nResults: The summary of the major findings was organised according to the subject matter raised by the research questions of the study. The results reflect the survey questionnaire and observation. Table II illustrates the respondents' qualification: 10 (3.7 per cent) respondents have a PhD/Doctorate degree, 145 (53.1 per cent) respondents have a Master degree, 106 (38.8 per cent) respondents have a Bachelor degree, 5 (1.8 per cent) respondents have a Diploma, 2 (7 per cent) respondents have a professional certificate and 2 (4 per cent) respondents have other qualification. The data show that the respondents were aware of the KM practices that fit the purpose of this study.\nConclusion: A key aspect to create awareness of KM practices is the identification of what KM means to employees at the HEIs, a review of their knowledge assets and the design of a strategy that can facilitate the creation, sharing and transfer of knowledge, to their institutions' competitive advantage. Furthermore, the HEIs in Malaysia require visionary leadership, which can introduce these policies and create the right climate for knowledge creation, sharing and transfer as a basis for enhanced research and innovation in the country. As part of this initiative, consideration should be given to investment in a dedicated KM office, interactive web portals and knowledge repositories, collaborative and multi-disciplinary and inter-institutional research projects, rewards and incentives, access to data and databases and increased collaboration with private sector. More widely, this study suggests that HEIs need to create broader KM practices, strategies and cultures. Further research on KM processes and policies in HEIs can contribute to a more robust and insightful knowledge base in this area. Additionally, KM practices need to be tapped from institutional skills and the already existing intellectual capital. A supportive institutional climate can, therefore, bring systemic transformation to the entire institution.\n", "role": "user" }, { "content": "Purpose: The purpose of this paper is to investigate KM practices that may be in place in the higher education institutions (HEIs) and whether the KM practices are made known to the employees for improving the teaching and learning environment provided at the Malaysian higher education institutions.\nDesign/methodology/approach: Data were collected using a personal administrated method made available to private higher education institutions academic members in five states with 30 or more employees. A total of 1,100 survey questionnaires were handed out, out of which 273 were collected and were usable (24.8 per cent response rate). The sample was checked for response and non-response bias. Results were tested using SPSS application and questionnaire tools.\nFindings: It was essential to establish the knowledge management (KM) capacity in key areas such as the ability to recognise experts within the institution, leadership's innovation, knowledge sharing and knowledge acquiring work culture, and technology usage. KM tools and techniques would help the institutions to meet their competitive goals; therefore, it is vital for HEIs to create KM awareness among the employees.\nResearch limitations/implications: Similar to most studies, it is anticipated that the participants' awareness of KM practices at their HEIs is very high. The samples were collected to evaluate the general view of KM awareness and how participants perceived KM practices. The total samples received for this study were expected; however, they were sufficient to study the impact.\nPractical implications: This paper provides support for the importance of KM practices and employees awareness at HEIs to enhance innovation and performance teaching and learning environment.\nOriginality/value: This paper is one of the first papers to find empirical support for the role of KM practices at HEIs. Further, the positioning of KM practices as a competitive tool can be considered as an influential factor to competitive advantage.\n", "role": "assistant" } ]