title
stringlengths 6
301
| keywords
listlengths 0
23
| messages
listlengths 2
2
|
|---|---|---|
Farmers' perspectives: Impact of climate change on African indigenous vegetable production in Kenya
|
[
"Kenya",
"Farmers",
"Perceptions",
"Climate change",
"African indigenous vegetables",
"Agro-climatic zones"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Sub-Saharan Africa (SSA) has been identified as one of the regions that are most vulnerable to the impacts of climate change (CC) (Bryan et al., 2013; IPCC, 2014). CC is seen as one of the major factors limiting Africa's efforts to achieve food security because of the continent's dependency on rain-fed agriculture and the low capacity of smallholders to adapt to CC (IPCC, 2014; Phirri et al., 2016). Furthermore, widespread poverty, inequitable land distribution and the declining size of farmland make the region more vulnerable (IPCC, 2014). It is projected that SSA will witness increases in temperature, changes in rainfall intensity and distribution and a rise in incidences of extreme weather events (e.g. droughts and floods), pests, weeds and disease epidemics (FAO, 2015; Connolly-Boutin and Smit, 2015). In Kenya, for example, since the early 1960s, the mean temperature has been increasing and rainfall patterns have become irregular and unpredictable, with the country experiencing an increased number of extreme events such as droughts and floods (GoK, 2013a; Thompson et al., 2015). As the full effects of these changes are yet to be felt, there is evidence in the literature that these changes are already having an impact on agricultural production. For instance, combined infestations of pests and diseases in plants have been reported to result in losses of over 50 per cent of the yield of major crops. Plant diseases are also estimated to cause up to a 20 per cent reduction in the yield of principal food and cash crops worldwide (Gautam et al., 2013; West et al., 2015), while weeds have been found to contribute the greatest potential yield losses of up to 34 per cent compared with those caused by pests and diseases (Oerke, 2005). According to Thornton et al. (2014), changes in climate variability and the frequency of extreme events may have a substantial impact on the prevalence and distribution of pests, weeds and crop diseases. Additionally, increases in minimum temperatures have negative impacts on rice yields by up to 10 per cent for each increase in temperature by 1degC in the dry season. Furthermore, increases in maximum temperatures lead to reduced yields of maize by 1.7 per cent for each day above 30degC under drought conditions (Thornton et al., 2013). Research shows that the impact of CC on agricultural production is likely to intensify in future because of an expected further increase in temperature. This increase is predicted to cause a yield decline of 14 per cent (rice), 22 per cent (wheat) and 5 per cent (maize) in SSA. The consequence of this is increased poverty and vulnerability of the people who primarily depend on rain-fed agriculture for their livelihoods (IPCC, 2014; Kabubo-Mariara and Kabara, 2015; Adhikari et al., 2015).\n2. Research methodology: 2.1 Study areas\n3. Results and discussion: 3.1 Farmers' perceptions of climate change\n4. Conclusions and recommendations: Most of the research that has been carried out on AIVs has ignored and downplayed the range of potential effects on production caused by CC variables. AIV farmers are aware that their area is getting warmer and drier. Increased temperatures, an increased frequency of droughts and dry spells, late onset, early cessation and decreased rainfall, are among the major climate hazards reported. Climate data revealed no trends in the mean annual rainfall in the selected ACZs in Kenya, but the mean annual rainfall in the semi-humid zone is increasing significantly. The main constraining factors identified in AIV production were increased temperatures, erratic rainfall patterns and inadequate rainfall caused by frequent dry spells and droughts. Farmers were aware of the moderate effect of climate on AIVs. The current volatility in rainfall patterns has made it difficult for farmers to plan their cropping calendar to suit the changes. Therefore, CC adaptation programmes need to provide accurate projected weather patterns to farmers. This study also revealed that farmers' perceptions were significantly associated with their ACZs, while there are highly significant differences in the mean rainfall of the three ACZs. Severity of pests and disease are more prevalent in the dry season than in the wet season. Too much rain and frequent dry spell conditions caused by CC alter the infestation of pests and manifestation of diseases as well as growth of some new weeds. This has resulted in lower yields for AIVs. Changes in the manifestation of pests, diseases and weeds in some AIVs could imply that farmers may be pushed into applying pesticides and herbicides on their vegetables. There is therefore a need to make farmers aware of the need to use pesticides and herbicides sustainably, including integrated pest management practices, to reduce environmental damage. The changes in the severity of pests and diseases in spider plant and African nightshade calls for a development of more resistant cultivars of these AIVs. This could be designed within programmes aimed at ensuring that farmers benefit from opportunities offered by engaging in AIV production. To achieve this, it is important to ensure that farmers gain access to good quality and affordable resistant seed varieties. Farmers' perceptions and observed trends should be communicated to policy makers at local, national and international levels. This will provide a pathway for effective planning of context-specific adaptation options and climate communication programmes. However, further research is recommended to ascertain the quantity of yield losses because of pests, disease and CC.\n",
"role": "user"
},
{
"content": "Purpose: Understanding farmers' perceptions of how the climate is changing is vital to anticipating its impacts. Farmers are known to take appropriate steps to adapt only when they perceive change to be taking place. This study aims to analyse how African indigenous vegetable (AIV) farmers perceive climate change in three different agro-climatic zones (ACZs) in Kenya, identify the main differences in historical seasonal and annual rainfall and temperature trends between the zones, discuss differences in farmers' perceptions and historical trends and analyse the impact of these perceived changes and trends on yields, weeds, pests and disease infestation of AIVs.\nDesign/methodology/approach: Data collection was undertaken in focus group discussions (FGD) (N = 211) and during interviews with individual farmers (N = 269). The Mann-Kendall test and regression were applied for trend analysis of time series data (1980-2014). Analysis of variance and least significant difference were used to test for differences in mean rainfall data, while a chi-square test examined the association between farmer perceptions and ACZs. Coefficient of variation expressed as a percentage was used to show variability in mean annual and seasonal rainfall between the zones.\nFindings: Farmers perceived that higher temperatures, decreased rainfall, late onset and early retreat of rain, erratic rainfall patterns and frequent dry spells were increasing the incidences of droughts and floods. The chi-square results showed a significant relationship between some of these perceptions and ACZs. Meteorological data provided some evidence to support farmers' perceptions of changing rainfall. No trend was detected in mean annual rainfall, but a significant increase was recorded in the semi-humid zone. A decreasing maximum temperature was noted in the semi-humid zone, but otherwise, an overall increase was detected. There were highly significant differences in mean annual rainfall between the zones. Farmers perceived reduced yields and changes in pest infestation and diseases in some AIVs to be prevalent in the dry season. This study's findings provide a basis for local and timely institutional changes, which could certainly help in reducing the adverse effects of climate change.\nOriginality/value: This is an original research paper and the historical trends, farmers' perceptions and effects of climate change on AIV production documented in this paper may also be representative of other ACZs in Kenya.\n",
"role": "assistant"
}
] |
A critical review of research on information behavior in assigned learning tasks
|
[
"Information retrieval",
"Learning",
"Individual behaviour"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: The world wide web (later the web) is an easily accessible and ubiquitous information environment, which has reshaped people's information behavior by enabling instant information searching in various contexts (Bilal, 2004; Bilal and Kirby, 2001; Fidel et al., 1999; Savolainen, 1999). The web has become an important information resource in teaching and learning at all levels of formal education (Bilal and Kirby, 2001; Chen, 2003). It is an open-ended information environment, which complements and extends traditional learning materials. Instant access to the web's information resources has radically enlarged the information space available to the learner (Rogers and Swan, 2004; Walton and Archer, 2004).The seemingly easy access to a wealth of information resources is a challenge as well as an opportunity. Easier access to information does not automatically mean better learning outcomes. The effect can be counter-productive. Studies on learners' information behavior suggest that pupils and students often search documents for the \"right\" answers copying and pasting pieces of text instead of constructing and articulating their personal understanding. Novice learners without a clear conception of what they are doing face difficulties in assessing the quality and validity of information searched (Alexandersson and Limberg, 2003; Hultgren and Limberg, 2003; Limberg, 1999; McGregor and Streitenberger, 2004).The problems mentioned above can and should be examined from the viewpoints of information seeking and retrieval (IS&R) research and of learning research. In this paper, we focus on the IS&R literature and adopt the task-based approach (Ingwersen and Jarvelin, 2005; Vakkari, 2003). The task-based approach seems a convenient choice, since learning assignments can be regarded and studied as information intensive work tasks (Ingwersen and Jarvelin, 2005). The strength of this approach is that it can develop shared conceptual frameworks with research on learner-centered instructional methods. Tasks are a growing research area, yet the focus of the studies seldom extends to the actual use of information or the learner's ways of understanding information content (Alexandersson and Limberg, 2003; McGregor and Williamson, 2005; Spink and Cole, 2006; Vakkari, 1997). On the other hand, past research on learning cannot give us a clear picture of the relationships between information behavior and learning. For example, in the models of problem-based learning, information seeking may be presented as a black box in the learning cycle but the box is seldom opened for a detailed analysis (Poikela and Poikela, 2005).Our interest in learning tasks originates from the goal of developing information literacy instruction. Effective seeking and use of information are included in information literacy standards such as ACRL[1], ANZIIL[2], and SCONUL[3] as necessary premises for intellectual access to information and learning to learn. This motivates us highly to focus on the characteristics of learning tasks, which could foster the learner's motivation for information literate behavior (Edwards and Bruce, 2004).This paper contributes to the inadequately explored area of IS&R research and learning research. Research on information behavior in learning tasks is scattered and difficult to represent coheren tly. To clarify related issues we propose:* the task-based IS&R research approach to learning tasks; and* a research framework to organize and interpret research on information seeking and use in learning tasks.The paper is organized as follows. First, we discuss the task-based IS&R approach to learning tasks. Next, we review empirical IS&R research focusing on information behavior in learning tasks. Finally, we summarize our findings and present a model visualizing the research framework for learning-related information behavior.\n2 Basic concepts, models and approaches: The aim of this section is to elaborate the basic concepts of learning task, information behavior and task-based IS&R research. We conclude by presenting a process model characterizing the core concepts and information related processes involved in the performance of a learning task.2.1 Learning task\n3 Empirical IS&R studies on learning tasks: In this literature review we cover studies focusing on the characteristics of and the relationship pertaining between the learning task and its information related subtasks. However, we recognize that studies focusing on the influence of a learner's personality on information behavior are too important to set aside. Researchers taking that approach have identified patterns in information behavior that correspond with findings obtained in task-based studies. Thus, we have discussed learner characteristics that have been shown to relate to learning task characteristics with respect to information behavior.The characteristics discussed in the review have been empirically studied as dependent or independent variables, but, where empirical research was not available and the issue at hand seemed important, (pre-)theoretical, non-empirically tested characteristics were discussed as well. Rather than attempting an exhaustive literature review covering all studies published on the subject, and meticulously restricting ourselves to empirical results, we aimed to outline a research framework for the learner's information behavior in learning tasks. The characteristics are organized into classes or research dimensions according to the classification scheme introduced in the Integrated IS&R Research Framework by Ingwersen and Jarvelin (2005, pp. 313-6; see also Pharo, 2002):* Actor dimensions-learner dimensions (LD).* Organizational task dimensions-learning task dimensions (LTD).* Access and interaction dimensions (AID).The document dimensions and algorithmic dimensions identified by Ingwersen and Jarvelin were ignored at this point. This helped to keep the review manageable and to enable the core aspects of the problem area to be studied in sufficient detail.3.1 Learner dimensions\n4 Summary and discussion: The process model (Figure 1) and the Integrated IS&R Research Framework were utilized to organize the literature and to discuss characteristics of and relationships between the learner, the learning task, and the access and interaction research dimensions. The results are summarized in Figure 2.The figure presents a task-based perspective on the IS&R research related to information behavior in learning tasks. The solid lines represent relationships studied explicitly in the reviewed literature. The broken lines represent our interpretation of the literature and our proposals for the relationships to consider in further task-based research. We discuss the main findings as follows.The patterns of information seeking resemble the conceptions of information seeking even though the researchers in question adopted different perspectives (Limberg, 2005) One might be tempted to link the patterns to the stages of the learning task, but empirical research does not support such a conclusion (Heinstrom, 2002, p. 239). Individual characteristics of information seeking patterns may be explained by other variables than the overall pattern they have been linked to. For example, anxiety is a common experience in the early stages of the task, but it is also a characteristic of the fast surfer; thoroughness may characterize information seeking at the early stages of the task, but it may also be a habitual way of seeking information or a reflection of the epistemological beliefs of a discipline (Heinstrom, 2002, pp. 211, 228, 231). These findings suggest that researchers focusing on tasks should control or at least acknowledge the characteristics of the individual and the environment.The learning process itself has received little attention. However, IS&R researchers are aware of the challenge of designing learning tasks capable of motivating typical schoolchildren, who will not commit themselves to assigned learning tasks. Excessive openness of the learning assignment is likely to impede learners' efforts to formulate a focus for their learning tasks. The formulation of a focus can be thought as a change in information behavior attributable to fundamental transformations in the way of knowing about a topic. If a focus cannot be formulated, and there is an assignment to be carried out, the learner is likely to adopt a survival strategy and remain unmotivated. The failure to formulate a focus can - given its effects - be linked to extrinsic motivation and product orientation, which are both typical especially for schoolchildren. The surface study approach, which is in part an effect of extrinsic study motivation, could be linked to product orientation as well. Thus, information behavior associated to the product orientation may be typical for the learner rather than just an effect of a temporary lack of motivation associated to the pre-focus stages of the learning task.How the orientation towards the product or the process of learning could relate to learning outcomes, described in terms of the patterns of knowledge construction, is not explicitly discussed in the literature. However, it is perfectly conceivable that, for example, the process orientation and the integrative approach to knowledge construction are linked. Whatever the factors involved here are, the issue is important, because the literature demonstrates that the learner holding a certain conception of information seeking and use, or adopting a certain study approach, achieves corresponding learning outcomes. We conclude that many of the discussed phenomena might be closely linked or conceptually overlapping. We recommend that instead of proposing more models and terms researchers should pay more attention into integrating their research frameworks.The reviewed literature contained little empirical research on information use - in contrast to information seeking and searching - in the documentation process, on the characteristics of resulting learner-produced documents, and how these relate to learning outcomes. The steps involved when a mixed bag of notes on search results and documents retrieved is developed into a more or less coherent paper have not been investigated. The reviewed studies did not present information seeking and writing as convergent and simultaneous processes but rather as consecutive phases, where learners went into school libraries to make notes and then returned to their classrooms to write their research papers. The role of writing in learning and writing as a form of reflection have been studied in learning research, for example, as the use of portfolios (Lonka et al., 2001). Kavli and Mikki (2006) report on the development of web-based learning system aimed at integrating information seeking and writing. The book Writing as a Learning Tool (Nelson, 2001) includes several papers on the use of writing in learning.These main findings need to be contrasted with the goals of information literacy education in schools and universities. We believe that strengthening good information behavior practices, as a sub-goal of learning tasks, is a key to information literacy (Tuominen et al., 2005). The teacher needs to have well-established instructional tools to design and direct learning tasks to achieve this sub-goal. It is not enough that the teacher is able to recognize the symptoms of surface learning, fast surfing and orientation towards documentary output rather than learning outcomes; the teacher should be able to persuade the learner to take constructive approaches in information behavior. The ideal situation where the learner is able to concentrate on a focused aspect of the subject content should be endorsed sufficiently in real life learning situations. The task-based approach offers a promising avenue to synthesize research on learning assignments, teacher interventions, learning process, and information behavior. This paper has provided only a preliminary picture of the factors influencing the learner's information behavior in the learning task. Ideas adopted from learning research would help to develop IS&R research in this area, especially with respect to teacher interventions and the learning and documentation process.\n",
"role": "user"
},
{
"content": "Purpose: - This paper aims to give a critical review of the empirical information seeking and retrieval (IS&R) literature focusing on learning related information behavior. It also aims to propose the task-based approach to link research on learning and information behavior in learning tasks and to organize and interpret the literature.\nDesign/methodology/approach: - This paper takes the form of a critical literature review.\nFindings: - Researchers have identified patterns of information behavior and variations in the conceptions of seeking and using information for learning tasks. The results are fragmented and a framework should be developed to guide further research. Learners' information behavior in learning tasks may be attributable to various explanatory factors, which have not been taken into account in any single study. The concept of focus formulation appears to link information seeking and learning. The role of the writing process in learners' information behavior and the characteristics of the resulting documents have received least attention in the research field.\nResearch limitations/implications: - The review emphasizes task-based IS&R literature. Learning research may shed more light on specific questions.\nPractical implications: - The paper may help teachers in designing learning tasks.\nOriginality/value: - The paper provides a synthesis of recent studies on information behavior in learning tasks and identifies new paths for further research.\n",
"role": "assistant"
}
] |
Cross-sectoral differences in the drivers of innovation: Evidence from the Irish Community Innovation Survey
|
[
"Innovation",
"Community Innovation Survey",
"Sector"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: This paper analyses sectoral differences in the propensity to innovate and the extent to which the mechanisms through which firms innovate vary across sectors. This is accomplished through the use of the Irish Community Innovation Survey (CIS) 2004-2006. Four broad sectoral classifications are identified; high-technology manufacturing (HTM); all other manufacturing (AOM); wholesale, transport, storage and communication (WTSC); and financial intermediaries (FI). Peneder (2010) alludes to tension between firm- and sector-level studies of innovation activity. He notes that the former point to heterogeneous behaviour among individual firms while the latter show significant differences between sectors and observed consistencies in sectoral data. This tension, he argues, has important implications for innovation policy in that \"industry characteristics matter and cannot be ignored [and] their accurate understanding helps to design policy programs and tailor them more effectively to the needs of targeted firms\" (Peneder, 2010, p. 324).\n2. Literature review: Howells (2002) and Lissoni (2001) argue that innovation is of vital importance, not only for business success, but also for economic growth and social well-being. This paper, in analysing whether the determinants of innovation vary across sectors, aims to provide insight into whether firms differ across sectors in the mechanisms through which they generate innovation. This may contribute to better policy formation and targeting.\n3. Methodology: Analyses by Doran and O'Leary (2011) and Hall et al. (2009) identify different propensities to innovate for firms in different sectors. However, they do not assess whether the drivers of innovation vary across sectors. This paper moves beyond the traditional method of controlling for sectoral factors using dummy variables.\n4. Data: The data set utilised by this paper is the Irish CIS 2004-2006. This survey was conducted jointly by Forfas (Ireland's National Policy Advisory Body) and the Central Statistics Office in Ireland. A total of 4,150 surveys were issued with 1,974 responses. The survey is directed to companies employing more than ten persons and engaged in a range of sectors.\n5. Empirical results: 5.1 Results of the restricted model\n6. Conclusions and implications: This paper estimates an innovation production function which analyses the effects of external interaction and internal R&D on firms' innovation performance, using data from the Irish CIS 2004-2006. While it is common to control for differing propensities to innovate across sectors through the inclusion of sectoral dummy variables in innovation production functions, this implicitly assumes that the importance of innovation inputs do not vary across sectors. In a key contribution, through the estimation of an innovation production function, for four differing types of innovation, and the subsequent testing of these functions for parameter stability across sectors, this paper provides an empirical analysis of whether the importance of innovation inputs vary across sectors. For all forms of innovation, the likelihood ratio test indicates that there is parameter instability across sectors. This suggests that there is a strong degree of heterogeneity in the drivers of innovation across sectors.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyse differences in the drivers of firm innovation performance across sectors. The literature often makes the assumption that firms in different sectors differ in their propensity to innovate but not in the drivers of innovation. The authors empirically assess whether this assumption is accurate through a series of econometric estimations and tests.\nDesign/methodology/approach: The data used are derived from the Irish Community Innovation Survey 2004-2006. A series of multivariate probit models are estimated and the resulting coefficients are tested for parameter stability across sectors using likelihood ratio tests.\nFindings: The results indicate that there is a strong degree of heterogeneity in the drivers of innovation across sectors. The determinants of process, organisational, new to firm and new to market innovation varies across sectors suggesting that the pooling of sectors in an innovation production function may lead to biased inferences.\nResearch limitations/implications: The implications of the results are that innovation policies targeted at stimulating innovation need to be tailored to particular industries. One size fits all policies would seem inappropriate given the large degree of heterogeneity observed across the drivers of innovation in different sectors.\nOriginality/value: The value of this paper is that it provides an empirical test as to whether it is suitable to group sectoral data when estimating innovation production functions. Most papers simply include sectoral dummies, implying that only the propensity to innovate differs across sectors and that the slope of the coefficient estimates are in fact consistent across sectors.\n",
"role": "assistant"
}
] |
A conceptual Lean Six Sigma framework for quality excellence in higher education institutions
|
[
"Conceptual framework",
"Higher education institutions",
"Quality excellence",
"Lean Six Sigma (LSS)",
"HEIs"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Quality products and services are essential for firms in the competitive global markets. Deterioration in the quality of output would significantly erode the customer satisfaction levels. This basic understanding of quality calls for attention in quality excellence. Excellence in quality is not merely delivering the required output to the customers, but a management strategy to enhance and align the processes in a way that excellence becomes inherent part of the ways of working. The research on quality excellence has been building over the theories of Deming, Juran and Crosby (Sunder M., 2016b). From past two decades, researchers have seen a paradigm shift of viewing quality from manufacturing to services. There has been ceaseless efforts made by researchers to comprehend that the quality perspectives across service sectors and higher education sector is not an exception.\n2. What is quality excellence?: In the 1970s, many organisations faced economic difficulties, and lost the competitive edge which they enjoyed in 1950s and 1960s. The almost inevitable result was deteriorating market shares, unrelenting losses and unemployment by the firms that ranged from General Motors and Harley Davidson to Xerox and Motorola (Koch and Fisher, 1998). This created a strong need for quality excellence in organisations for improving quality and productivity and reducing costs. Contributions by Edward Deming, Philip Crosby and Joseph Juran have laid a concrete foundation to tackle these challenges, and this was later popularised as the term quality excellence. They have devoted their lives to helping organisations improve the quality of their products and services. The evolution of quality excellence from then till date has seen various progressive steps, and numerous methodologies emerged as a result. The first amongst them was the TQM. The popularity of TQM has arisen from a widespread appreciation that a comprehensive quality assurance system is insufficient to ensure competitive quality performance (Tannock, 1991). With the help of TQM, many firms such as Texas Instruments, Xerox, IBM and Motorola were able to improve their business positions by overcoming threats from global competition and other changes in the business environment (Lozier and Teeter, 1996). Furthermore, extending TQM to quality excellence, various studies were undertaken and different recognition mechanisms were developed such as the Malcolm Baldrige Award, European Foundation for Quality Management, the Deming Prize Criteria, etc. Based on these studies, a wide range of management issues, techniques, approaches and systematic empirical investigations have been generated (Bayraktar et al., 2008).While TQM has been used in the manufacturing sector for a long time, service applications of TQM were relatively low. Specifically in education services, even though a high use of TQM was reported by administrative, support and academic departments in higher education, only a small number of institutions (17 percent in 1995 and 15 percent in 1998) employed a complete TQM model (Vazzana et al., 2000). Following the era of TQM was the Kaizen movement or continuous improvement (CI). The concept of CI comes from the Japanese term, Kaizen, that was initially developed and spread by Masaaki Imai who is known as the Father of CI. According to Chang (2005), the CI cycle consists of establishing customer requirements, meeting their requirements, measuring success and continuing to check customers' requirements to find areas in which improvements can be made. Manos (2007) defined CI as a series of subtle and gradual improvements that are made over time. Lean was another recognised quality excellence philosophy which promoted CI culture in organisations. A few researchers defined Lean as a way to focus on customer, whereas others defined Lean as a way of reducing waste and identifying value. However, there was no consistent clarity on the definition of Lean. The uncertainty stems as Lean evolved over a long period of time and because of its mistaken equivalence with other structured quality methodologies (Hopp and Spearman, 2004; Womack et al., 1990; Spear and Bowen, 1999).\n3. How is HEI different from manufacturing industry?: When it comes to the enquiry of quality excellence through LSS for higher education services, it is essential to understand the uniqueness of HEI. This is because the higher education services sector is very different from that the manufacturing sector where most of the quality methodologies have emerged from. According to Oldfield and Baron (2000), higher education can be seen as a pure service, suggesting that it possesses all the unique characteristics of a service. Gruber et al. (2010) emphasise that the higher education industry offers services that are predominantly intangible, perishable and heterogeneous. It is evident that higher education is not something which could be intensely defined or measured as a metric of knowledge. It is heterogeneous as it offers a wide range of courses or knowledge offerings. Higher education as a service also satisfies the perishability criterion since it is difficult to store, and evaporates over time if not refreshed periodically. However, there are ways to handle this characteristic, for instance, the emergence of e-learning and video technology. An important difference between manufacturing and higher education services is difficulty in systematically measuring key metrics such as customer satisfaction and employee involvement (Bayraktar et al., 2008). From the customer definition viewpoint, the higher education sector is significantly different from manufacturing and rest of the services. This is because the students who are considered as the primary customers of the HEI cannot pay money and buy knowledge. Unlike products and other services, education cannot be quantified as a matter of magnitude with money equivalents. Beaver (1994) considers students as primary customers for the higher education industry and raises concern on using student grade distribution to assess quality in analogy with statistical control methods used in industry. He also feels that students are more than customers purchasing a product since students' learning has various contributing factors beyond the classroom, such as social and family background. Apart from students, staff members, alumni, parents (paying fees), governments (as they work closely with the university sector and provide them with funding, etc.) can be considered as customers/stakeholders making the customer base broader for the HEIs.\n4. LSS in HEIs: When it comes to LSS in HEIs, the applications need not be rationed to the academic processes. University administration and infrastructure are few other areas which could be looked upon for improvement opportunities. However, there is very little academic literature available on LSS in HEIs. This may be due to the reason that many HEIs are just in the nascent stages of deploying LSS, and yet to reap its complete benefits. Raifsnider and Kurt (2004) highlighted the importance of LSS in HEIs by a case-study application to document management services to simplify and streamline document-intensive business processes, such as enrolment. Antony et al. (2012) presented the challenges and barriers to be encountered during the introduction of LSS in the HEIs. They also highlighted the sustainability of LSS in the higher education context by prescribing a few useful tools and techniques for process improvement problems. Though the knowledge of LSS toolkit is important for the implementation of LSS, identifying the standards based on stakeholder requirements becomes the key. Demonstrating the need for the overall stakeholder's satisfaction for implementing and sustaining LSS, Chen et al. (2006) argued that alongside the students' satisfaction the projects in an HEI environment should also focus on the satisfaction of teachers and other administrative staff as well. In their research study of educational institutions in Taiwan, the authors claims that there is a need to establish a set of performance management tools to create educational objectives and standards, and to increase the competitiveness of university education in a globalised environment.\n5. The LSS framework for quality excellence in HEIs: Considering the uniqueness of project-based learning for students at HEIs, it is fairly easier to deploy a structured quality excellence framework, provided the above readiness factors are in place. From the above discussion in Sections 1-4, the literature clearly defines the opportunity for developing such an LSS model for imbibing quality excellence for HEIs. The below mentioned LSS framework is proposed in order to complement this need. The model is developed keeping in the mind the fundamental purpose of LSS which is to make HEIs more competitive than the competitors while facing challenges such as globalisation of education, pressure of funding and challenges of technology. For this purpose, the model is proposed in a fairly generic way. However, the HEIs are advised to consider the cultural parameters, priorities and leadership while deploying the suggested model, as these could be unique and quite different for each HEI, and cannot be generalised across the higher education sector. Further, appropriate intuition should be applied in order to understand the challenges from the HEIs. The challenges facing LSS implementation in HEIs have similarities with many other sectors that have embarked on the LSS journey. These could include lack of motivation, misalignment of senior management, inserting the process into the academic side of the house, deploying a project-based approach, preconceived views of stakeholders, selection of solutions in the context of HEIs and sustenance of improvements (Antony, 2015).\n6. Leadership agenda for sustaining the LSS deployment in an HEI setting: Many people strongly suggest LSS must be launched at the executive level because it proves to be more successful when it is led by top executives in an organisation. GE, Honeywell, Bank of America, Commonwealth Health Corporation, Sony, Whirlpool, Motorola and Bombardier are several examples of companies in which successful implementation of LSS are closely connected to a thorough commitment by top management. LSS deployments are much easier when top management has a passion for the initiative.\n7. Conclusion and agenda for future research: Though Lean has been a great quality excellence methodology for success in HEIs, the shortcomings of Lean are tackled effectively by LSS. LSS as a hybrid methodology has proved to be a game changer in the services sector than Lean implemented in isolation. The synergies that Lean, when used along with Six Sigma methods, will bring are not only an effective toolkit for improving the quality of services, but also acts as a management strategy for quality excellence.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present the potent application of Lean Six Sigma (LSS) in higher education services and to suggest a conceptual framework of deploying LSS in the higher education institutions (HEIs).\nDesign/methodology/approach: This paper aims to outline the importance of the quality excellence criterion in general through different constructs from the literature including TQM, Lean, Six Sigma and LSS. The paper further expands on the applicability of LSS in the HEI setting, with an understanding of HEI as comparatively different from the manufacturing sector from where LSS originated. The published literature on LSS and authors' experience in the field of LSS (as practitioners) were used as sources for devising a conceptual framework for LSS deployment in HEIs.\nFindings: The key finding from the study is the confirmation of LSS applicability in HEIs. The study also confirms that LSS is not anyone's job but a key part of leadership agenda. The six-staged conceptual framework prescribed for the HEIs as part of the paper is the key contribution of this study. The model describes that LSS readiness is the foremost step in the LSS deployment journey in HEIs. The strategic perspectives of aligning the organisational vision for quality excellence need to be achieved by establishing a need for LSS through leadership. Then, developing an LSS deployment strategy becomes the next step. Educating the appropriate stakeholders (students) and team formation becomes the next important steps in the LSS deployment. Once the above steps are implemented in right way, identifying and implementing LSS projects becomes critical. The execution and closure of the LSS projects leads to quality excellence in HEIs.\nResearch limitations/implications: Since the framework prescribed here is a conceptual framework, it deserved a testing in the real-life context. This leaves an opportunity for future researchers to test and expand on the implications of the model. Though the LSS framework is customised for HEIs, the fundamental factors essential for any change management initiative holds correct for LSS as well, and hence these are not elaborated as part of this paper.\nOriginality/value: This paper would serve as an excellent resource for both academicians and LSS practitioners in HEIs, for deployment. The framework presented here is the original work contributed by the authors, as the first of its kind in the academic literature relevant to LSS in HEIs.\n",
"role": "assistant"
}
] |
Patient speaking for patients: What constitutes genuine patient input into pharmaceutical policymaking?
|
[
"User involvement",
"Health care",
"Governance structures",
"Patient safety",
"Patient perspectives",
"Patient experience"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: In the last 20 years there has been a major global expansion of health institutions and agencies seeking to involve patients in eliciting their input on health matters. One of the most active areas of patient involvement in health care revolves around consumers of pharmaceuticals. Often terms like \"patient engagement\" or \"patient-oriented\" describe endeavors designed to involve patients in crucial aspects of drug decision-making including drug safety assessments, clinical research, drug evaluation and drug coverage. Danish researchers studying this area succinctly captured the importance of the lay public in developing pharmaceutical policy, writing that \"because the lay public and medicine users are not one entity, the challenge facing policymakers is to identify and incorporate the many diverse voices\"(Traulsen and Almarsdottir, 2005).\nWhy is patient engagement important to clinical governance?: The role of patient engagement is vital for clinical governance for a number of reasons: when citizens, patients, or health care consumers are involved in health system decision-making their input can lead to more ethical, sustainable and equitable care. As the World Health Organization has concluded: \"Considerable evidence suggests that patient engagement can improve their experience and satisfaction and also can be effective clinically and economically\"(Coulter et al., 2008).\nThe case of flibanserin: two strikes, then a home run!: The history of drugs for female sexual dysfunction and the FDA's approval of flibanserin in 2015, captures many of the key features of the role of the pharmaceutical industry in shaping the nature of disease in order to market products for that disease (US Food and Drug Administration, 2008). This concept of \"selling sickness\" has been well known for over a decade and the intimate involvement of the pharmaceutical industry in trying to develop a \"female Viagra\" from the early 1990's has been well documented (Moynihan, 2005). Drug makers saw the incredible popularity of the market for drugs for erectile dysfunction and naturally were keen to exploit that market. The key problem was that many argued that female sexual dysfunction was not well-defined and some argued did not exist as a clinical entity (Moynihan and Mintzes, 2010). It also was fairly clear among scholars and sexologists not connected to the drug industry that female disinterest in sex was largely linked to relationship and social issues and was quite clearly not a clinical disease that required drug treatment (Tiefer and Kaschak, 2001).\nWhat's next?: Researchers who have studied patient engagement conclude that it \"comes at a cost and can become tokenistic\" and that \"research dedicated to identifying the best methods to achieve engagement is lacking and clearly needed\" (Domecq et al., 2014).\nResearch: priority-setting partnerships: One of the problems with pharmaceutical research, often which coalesces around diseases, is that the research funders, usually pharmaceutical companies, never forget their key priorities which are profits. Since patients are the ones who swallow drugs, they are not so concerned with those profits but are most vested in the outcome of treatments. Unfortunately the commercial orientation of the pharmaceutical industry ultimately answers to shareholders, and this prompts drug manufacturers to use all modes of influence along the entire decision chain: researchers, clinicians, patients and payers. The many ways in which drug companies fund and \"partner with\" patient groups has been widely documented, and other than promoting the drug company's version of treatment, the patient groups they have as \"partners\" are there to provide valuable public testimonials not only for marketing but for convincing regulatory officials of the value of a treatment (Moynihan and Cassels, 2005). This is what happened with flibanserin.\nTreatment experiences: Patients like Me: Patients like Me is a USA for-profit company that allows people to sign up free of charge in order to \"compare treatments, symptoms and experiences\" with others (Patients Like Me, 2016). Patients like Me is a private company that openly asks people who have particular diseases or conditions to share their experiences using medical devices or pharmaceuticals and this information is sold the makers of drugs and other products. While this is a form of patient engagement, its goals are open, and information given is expected to inform manufacturers of the \"real life\" experience of their products. While some critics might say this enables manufacturers to more easily spot public relations issues as they arise, others would argue this patient involvement is valuable and could result in safer, more appropriate use of medical technologies.\nDrug coverage decision-making with consumer input: the case of stakeholder input into BC pharmacare: In British Columbia, due to changes to the way the ministry consults with \"stakeholders\" any invitees who wish to give input to the ministry on drug approval matters must either be a caregiver for or a person suffering from the disease the drug is to treat, or be part of an \"official\" list of patient groups registered with the Ministry (Schuchman, 2011). In one stakeholder meeting with over 300 invitees to discuss \"ongoing priorities within the Ministry of Health\" \"about three quarters of the invitees were representatives of drug companies, with the remainder who were associated with patient groups, chain drug stores, universities, health authorities and lobbying companies.\"\nConclusion: We can only hypothesize on the reasons why public involvement in policymaking is dominated by special interest groups yet policymakers today need to be able to \"identify, incorporate and prioritize the many diverse needs.\"\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to recognize the vital reasons for including public and patient voices in health policy decision-making, but illustrates the challenge it creates for decision-makers who must consider whether those voices represent patient interests or corporate interests.\nDesign/methodology/approach: This paper takes the form of a narrative review.\nFindings: The history of flibanserin, a controversial new drug to treat a debatable condition, illustrates how a public relations campaign could circumvent the well-established process to weigh evidence of potential harm vs benefit by one of the most robust drug regulators in the world.\nPractical implications: It is both vital to recognize a fundamental problem that exists when corporate interests deceptively assume the mantle of \"the patient voice\" and then act to reduce that influence while supporting and building capacity in genuinely independent, consumer-focused activities.\nOriginality/value: This paper suggests that organizations interested in consumer protection and the safe and cost-effective use of health resources create policies and procedures that can foster genuine consumer involvement while recognizing the danger to patient safety and consumer interests when consumer involvement is hijacked by vested interests.\n",
"role": "assistant"
}
] |
A QFD-based expert system for industrial truck selection in manufacturing organizations
|
[
"Decision making",
"Computer applications",
"Quality function deployment",
"Automated materials handling"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In any manufacturing organization, raw materials, work-in-progress, finished goods, scraps, packing materials, etc. need to be transported/transferred from one stage to another according to the requirements of the production line, which can often be tedious, expensive and troublesome, and also, there is a possibility of materials being lost or damaged while in transit. These materials are of various shapes, sizes and weight, which demand for a systematic and scientific method for their movement, protection, storage and control throughout the manufacturing process. Efficient handling of materials boosts employee morale, improves customer's service, reduces inventory, shortens delivery time, speeds up throughputs/productivity and trims down overall production cost, which are imperative for sustainability and growth of a manufacturing organization in today's global competitive environment. Moreover, material handling (MH) can account for 30-75 percent of the total production cost, and it can be primarily responsible for reducing the operating cost by 15-30 percent (Kulak, 2005). It is often regarded that MH only adds to the cost of a product and does not impart value to it. Although, MH does not provide a product with form utility, but the time and place utility furnished through it can enhance the real value of the product, i.e. the value of a product increases with proper handling of materials. Therefore, it is important that transportation of materials from one processing stage to another takes place with the aid of suitable material handling equipment (MHE) instead of manual handling. An appropriate MHE facilitates reduction in material damage in order to improve quality, reduces overall manufacturing time while designing efficient material movement, improves material flow control, creates safe and hazard-free work environment for employees, utilizes time and equipment in a better way and is also crucial for longer service life and smooth running of the organization. A suitably selected MHE reflects on the efficiency and speed of the manufacturing operations, which ultimately result in shortened order completion cycles. Hence, the investment in MHE is strategic in nature and is always based on long-term requirements, taking product volumes and varieties into consideration. Taking into account all the above-mentioned points, it is evident that MHE selection plays a key role in enhancing the overall production effectiveness and profitability of a manufacturing organization.\n2. Literature review: In the literature, there are several studies focussed on selection of apposite MHEs for various manufacturing systems. Malmborg et al. (1987) developed a prototype expert system considering 17 equipment attributes and 47 devices for industrial truck type selection. Malmborg et al. (1989) designed a microcomputer-based expert system \"EXIT\" in Turbo PROLOG for solving industrial truck selection problems. Park (1996) developed an expert system called \"ICMESE\" for selection and evaluation of MHEs suitable for movement and storage of materials in a manufacturing facility. Chan et al. (2001) presented an analytic hierarchy process (AHP)-based expert system to automate the design of an MHE selection system and provide artificial intelligence in the decision-making process. Yaman (2001) adopted a knowledge-based approach for combining MHE selection and pre-design of an MH system for a given facility layout. Kulak et al. (2004) applied fuzzy information axiom for development of a decision support system (DSS), \"MHAD\" for solving MHE selection problems. Lashkari et al. (2004) presented an integrated approach for operation allocation and MH system selection in a cellular manufacturing cell. Cho and Egbelu (2005) designed a web-based system \"DESIGNER,\" which could model and automate the MH system design process, including selection of the best suited MHE for a specific handling task. Kulak (2005) developed a DSS named as \"FUMAHES\" for selection of MHE in a real case based on fuzzy information axiom principles. Chakraborty and Banik (2006) applied AHP method for selecting the optimal MHE under a specific handling environment and organizational requirements. Vijayaram (2006) reviewed the significance of expert systems while selecting appropriate handling equipments in engineering industries. Kirmanli and Ercelebi (2009) designed an expert system within a Kappa PC shell, for hydraulic excavator and truck selection for surface mining applications. Mirhosseyni and Webb (2009) presented a hybrid approach integrating fuzzy knowledge-based system and genetic algorithm for selection and assignment of the most appropriate MHE. Onut et al. (2009) combined fuzzy analytic network process and fuzzy technique for order preference by similarity to ideal solution (TOPSIS) to select the most appropriate MHE while also considering its cost efficiency. Tuzkaya et al. (2010) implemented fuzzy AHP and preference ranking organization method for enrichment evaluation for evaluating various MHEs in a manufacturing organization. Baykasoglu et al. (2013) combined fuzzy TOPSIS and fuzzy decision-making trail and evaluation laboratory method for solving truck selection problem of a land transportation company. Karande and Chakraborty (2013) applied weighted utility additive method to solve an MH equipment selection problem. Sawant and Mohite (2013) developed a DSS \"AGVSEL\" for selection of automated guided vehicles based on TOPSIS, block TOPSIS and modified TOPSIS methods. Mousavi et al. (2014) adopted a fuzzy grey group compromise ranking method to deal with MHE evaluation and selection problems having uncertain information. Yazdani-Chamzini (2014) employed two fuzzy multi-criteria decision making (MCDM) methods, i.e. fuzzy AHP and fuzzy TOPSIS to select the most suitable MHE under an uncertain environment. Hadi-Vencheh and Mohamadghasemi (2015) proposed a new hybrid fuzzy MCDM model for solving MHE selection problems. Khandekar and Chakraborty (2015) applied fuzzy axiomatic design principles for selecting the most appropriate MHE for a given handling task. Prasad et al. (2015) designed and developed a software prototype based on QFD technique for selecting two most suitable bulk-type MHEs, i.e. excavator and wheel loader for specific applications at a construction site. Patel et al. (2016) developed an integrated AHP-grey relational analysis model for MH selection in a galvanizing steel industry.\n3. QFD methodology: Customers are vital for any manufacturing organization; understanding and meeting their ever changing demands is the key to success while not recognizing them is a recipe for failure. QFD methodology is an extensively implemented tool for capturing the stated and unstated needs of the customers in these days. This quality improvement tool was developed in late 1960s in Japan by Akao (1990) and was first implemented at Mitsubishi Heavy Industries Kobe Shipyard in 1972. It establishes market's needs utilizing the voice of customers and transforms those needs into design, production and manufacturing process characteristics. This results in a systematic engineering process, which prioritizes and links the product development process with customers' requirements to assure quality in a product as defined by the customers, leading to the highest customer satisfaction. The main advantage of this methodology exists in its ability to translate subjective quality criteria into objective ones that can be quantified and measured. The voice of customers' can be extracted through various ways, such as direct discussion or interviews, surveys, customers' feedback, warranty data, field reports, etc. This understanding of the voice of customers' is then summarized in a quality planning matrix, commonly known as house of quality (HoQ) matrix, as shown in Figure 1.\n4. A QFD-based expert system for industrial truck selection: The manufacturing organizations often operate in a dynamic environment driven by changes in market conditions, customer demands, product design and processing technology (ElMaraghy et al., 2014). This QFD-based expert system is designed to interrelate the needs of customers with respect to an industrial truck along with its desired technical specifications, derive priority weights for the considered technical requirements and select the most appropriate industrial truck for performing a specified handling task within the manufacturing organization from an assortment of available feasible alternatives. Figure 2 presents the framework for design and development of this expert system.\n5. Illustrative examples: In order to demonstrate the applicability and usefulness of the developed QFD-based expert system, it is applied to solve two real time industrial truck selection problems.\n6. Conclusions: There are an assortment of manual, semi-automated and automated MHEs, and technologies available to assist in the movement, protection, storage and control of materials and products throughout the manufacturing process. Selection of the most appropriate industrial truck from a pool of alternatives available in the market based on various conflicting criteria is a cumbersome and time consuming task. Thus, a QFD-based expert system is designed and developed to reduce the gap between the identified best industrial truck and real time operational requirements in a manufacturing unit. It helps the decision makers in shortlisting the evaluation criteria considering the voice of customers and interrelate technical requirements with customers' requirements to incorporate the customers' demands into the finally selected model of various industrial trucks. The application of this expert system on a real time job floor would aid in systematizing and restoring information with a view to accomplish the desired goals of a manufacturing unit with better exactness within the restricted available budget and time while exterminating the enormous calculations involved. It would work progressively, devoid of getting emotional, tensed or exhausted. Dissimilar to human beings, it does not get anxious, fatigued or dreaded, and works steadily during crisis circumstances. It can be utilized to save scarce expertise in the enterprises, to accumulate knowledge of persons who are retiring, and to maintain corporate adeptness with the intention that they can be extensively disseminated to other factories, offices or plants of the organization. It would also help knowledge to last long for an indefinite period. It can be employed to encompass knowledge of numerous experts. It is also competent enough to work under group decision-making environment. The database containing the technical specifications of various industrial trucks can be upgraded from time to time to make it more exhaustive, versatile and dynamic, while incorporating the remaining unconsidered industrial trucks and those that are yet to come in near future. It is designed for a specific type of MHE selection problem, i.e. industrial truck, but in future, there is enough scope for demonstrating its applicability for selection of other categories of MHEs. Development of a web-based comprehensive MHE selection system for worldwide access may also be included as another future scope.\n",
"role": "user"
},
{
"content": "Purpose: Availability of accurate quantity of materials, at correct place and at right time is extremely critical for increasing production effectiveness of any manufacturing organization. This can be achieved through employing an appropriate material handling equipment (MHE) capable of performing the desired operation. Therefore, choosing a right MHE from the available options is a key concern for the success, growth and competitiveness of a manufacturing organization. The purpose of this paper is to describe the design and development of an expert system based on quality function deployment (QFD) methodology in Visual Basic 6.0 for selecting the most appropriate industrial truck which is a commonly practiced MHE in any manufacturing organization.\nDesign/methodology/approach: A QFD-based approach is adopted to incorporate customers' needs into the evaluation criteria on the basis of which industrial truck selection is carried out. The applicability of the developed expert system in solving industrial truck selection problems is demonstrated using two illustrative examples.\nFindings: While applying this QFD-based model, CPCD 80x manufactured by Heli is recognized as the most suitable forklift truck for transporting unitized loads within a manufacturing unit with some spatial constraints, and for loading/unloading packages/boxes/cartons and place them at the desired locations in a manufacturing unit, ETV 216 manufactured by Jungheinrich evolves out as the most suitable reach truck.\nOriginality/value: Till date, numerous research articles have been published suggesting the applications of various mathematical models, multi-criteria decision-making methods and knowledge-based systems for solving MHE selection problems, and it is intriguing to note that none of the previously adopted methods has proposed a systematic procedure for selection of the evaluation criteria and interrelated the needs of customers with the technical specifications of MHEs while identifying the best alternative for performing a specified operation. These issues can be addressed through application of this developed QFD-based expert system, which can translate customers' needs into organizational functions that are implementable in the decision-making/selection procedure.\n",
"role": "assistant"
}
] |
Retail strategies for rose wines in Italy: a hedonic price analysis
|
[
"Survey research",
"Hedonic price",
"Branding",
"Rosé wine",
"Retail price",
"Wine attributes"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Wine is a complex good. Wine consumption has been thoroughly investigated over the past few decades, with emphasis on understanding what factors affect the consumer's purchase process. The strong relationship between wine and terroir combined with the high variability in grape varieties, origin, enology techniques and producer's know-how and reputation leads to a highly differentiated market in which bottle price ranges from a few euros to thousands. From a consumer perspective, the wine market shows strong segmentation (Cuellar and Claps, 2013) according to price and quality perception (Carew et al., 2017; Costanigro et al., 2009), budget constraints, habits and the consumer's wine knowledge and experience.\n2. Literature review: 2.1 Hedonic price studies\n3. Research methodology: The hedonic price methodology, that is, examining the relationship between a bottle's price and its attributes, continues to maintain research' interest worldwide (Oczkowski, 2015). A review of the literature on hedonic wine models is found in Estrella Orrego et al. (2012) and, more recently, in Oczkowski and Doucouliagos (2015).\n4. Research results: The hedonic model results allow us to estimate the per cent price impact of covariates having a significant price effect. For continuous variables and log transformation, this percentage is the price elasticity estimated by regression coefficients, while for dummy variables, this impact (also called semielasticity) has been estimated using the formula proposed by Kennedy (1981). For the economy of this paper, we report the full results of the direct effect models and a summary for the cross-effects models.\n5. Conclusions, implications, limitations and future research: The rose wine market in Italy is small in consumption but important in production because of exports. Rose wine consumption has a long tradition in Italy; however, it has decreased in volume because of a strong increase in the Prosecco consumption.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze the market of rose wines in Italy, to outline retail strategies and to investigate to what extent the price is affected by branding these wines.\nDesign/methodology/approach: A survey has been carried out on retailers by collecting data about wines as intrinsic attributes (grape variety, blending, origin, alcohol content, etc.) and extrinsic attributes (brand, price, packaging, etc.) and about outlet and retail environment. The hedonic analysis required a rearrangement of data survey, while a Box-Cox transformation allowed to control the strong heteroskedasticity detected of the data.\nFindings: Results provide strategies for still, semi-sparkling and sparkling rose market segments. Still rose wines are strongly differentiated, while the price is affected by the appellation, grape variety, blending, brand and outlet features. Two main strategies are suggested: the first focuses on appellations endorsing consumer's brand loyalty; the second is driven by retailers while involving weaker brands. Different pictures emerged for semi-sparkling and sparkling wines, as producers and retailers tend to follow consumer's preferences for fresh and easy drinking wines as well as to extend the product assortment.\nResearch limitations/implications: Results for sparkling rose wines cannot be generalized. The high fragmentation hinders the hedonic model performance in capturing the price effects of brands, appellations, grape variety and wine blend.\nPractical implications: The hedonic analysis provides suggestions for rose wine producers that should reinforce their brand through associations among intrinsic attributes, such as appellation, and extrinsic ones, such as price, while satisfying retailer requirements.\nOriginality/value: The paper contributes to the knowledge base about the Italian rose wine market, which is mostly export-oriented. Model results help to understand why the domestic consumption is stagnant with respect to other countries such as France or the USA.\n",
"role": "assistant"
}
] |
Faith (re)engages with business: cultural, legal, and managerial dimensions
|
[
"USA",
"Diversity",
"Religious tolerance",
"Legal rights and responsibilities",
"Religious discrimination"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Allow the human spirit to follow its tendency, and it will regulate in a uniform way political society and the holy city; it will seek, if I dare say so, to harmonize earth with heaven.\nIntroduction: As the noted French political writer Tocqueville suggests above, society and religion are capable of being united in a manner which maximizes the benefits of both. This article seeks to show that Tocqueville's statement, drawn from his examination of life in the USA almost 200 years ago, can be equally true today. Rather than consider the full panoply of potential relationships between American society and religion, however, the focus of this article deals with the question of why and how religion, also characterized as faith and spirituality, can - and perhaps even should - co-exist today within the realm of business.\nThe USA is a deeply religious nation: \"The twenty-first century [...] is dawning as a century of religion. Virtually everywhere, apart from Western Europe, people are turning to religion for comfort, guidance, solace, and identity\" (Huntington, 2004, p. 15). Unquestionably, the USA is a faith-based nation, with matters of religion impacting debates involving complicated political issues at the national level, economic decision-making within families, and important moral issues facing individuals, among many other areas of American life. Before examining the role religion has played and continues to play in business within the USA, this section provides a cursory analysis of the importance of faith in the broader ethos of the American people.\nReligious tolerance is a dominate characteristic of American culture: The popularity of religion within the USA is a direct consequence of the high degree of religious tolerance that has existed and currently exists in the country. Early seventeenth-century settlers to the new land of America left established communities in European nations largely for religious reasons and brought with them a religious zeal (Huntington, 2004). For example, referencing the historic voyage of Governor John Winthrop in 1630 from England to what would later become the state of Massachusetts, Eck (2001) eloquently writes:\nAmerican law embraces religious liberty: Along with deep cultural underpinnings of religious tolerance, the legal framework of the USA encourages - and at times mandates - acceptance into American society religious belief and accompanying methods of practicing one's faith. Religious tolerance from a legal perspective is illustrated below first by examining broad constitutional provisions pertaining to religious freedoms, followed by a cursory exploration of three acts of Congress passed within just over the past one-half century that either directly or indirectly support the virtue of religious tolerance.\nReligious tolerance as a precursor to vibrant religiosity: The virtue of religious tolerance, a hallmark American culture and reinforced by legal theories and actions, is critical to the ability of the US population to strongly embrace the ever-changing nature of a spiritually based experience. In examining the essence of collective expressions of faith in the USA, Finke and Stark (2005) found that beginning in \"at least 1776, the upstart sects have grown as the mainline denominations have declined. And this trend continues unabated, as new upstarts continue to push to the fore\" (p. 235). Continuing, Finke and Stark explain the lifecycles associated with belief systems:\nReligion and business: The early history of religion's influence on business\nConclusion: The people of the USA today are deeply influenced by religion, as they have been for centuries. Cultural and legal forces, drawing strongly on the virtue of religious tolerance, buttress the high degree of religiosity that continues to be a characteristic of the nation. As a result, US society is well suited for the continuing evolution of the faith-based experience, particularly regarding the ongoing popularity of religions centering on spiritual rather than institutional features and the greater diversity of faiths one might experience in the country.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to explore the changing nature of religion (also described in terms of faith or spirituality) as experienced in the USA, particularly regarding the interrelationship between business and religion within the contexts of culture, law and management. With a solid understanding of these subjects, business leaders, judges and public policy officials will be able to more effectively deal with issues arising from a more religious workplace.\nDesign/methodology/approach: This paper begins by examining the religious nature of the USA, with attention paid to the embrace of religious tolerance from a cultural perspective and religious liberty from a legal perspective. The piece then looks at the characteristics associated with religious vibrancy. Then, the paper delves into the impact of religion on business, both yesterday and today. Legal and managerial literature, studies and perspectives are used to determine how business can respond and perhaps even embrace a more religious business environment.\nFindings: After describing and critiquing cultural, legal and managerial dimensions associated with an increasingly religious business environment at worker and owner levels, this paper suggests there exists currently an adequate legal structure to serve the needs of religious workers and owners, if properly managed. More research is needed to find solutions to situations where competing interests conflict, but this paper provides a foundation upon which further study can be conducted.\nOriginality/value: This paper draws from cultural, legal and management sources to provide an understanding of the current religious environment facing business, legal and public policy leaders.\n",
"role": "assistant"
}
] |
"A whole new future" - identity construction among disadvantaged young adults
|
[
"Careers",
"Adolescents",
"Career development",
"Career guidance",
"Work identity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: A major part of who we are is defined by what we do. Work is a central component of our identity, and its presence or absence can have serious consequences for our psychological health and well-being (Jahoda, 1982). In fact, having a positive work-related identity (i.e. a favorable self-definition that is tied to participation in work activities; Dutton et al., 2010) is related to various positive outcomes such as the ability to adapt to new work settings, job satisfaction, confidence in making career decisions, and employment (e.g. Beyer and Hannah, 2002; Skorikov and Vondracek, 2007; Wrzesniewski et al., 2003).\nDisadvantaged young adults: Disadvantaged young adults face numerous barriers in entering employment and in sustaining a fulfilling career. They often lack work experience, qualifications and/or education and, therefore, confidence (Diemer and Blustein, 2007; Sirin et al., 2004). This group of young adults is generally categorized as \"high risk non-college bound youth\" (Worthington and Juntunen, 1997), \"urban youth\" (Chaves et al., 2004), or \"young people Not in Education, Employment or Training (NEET)\" (Bynner and Parsons, 2002). The profiles of these individuals tend to look the same: predominantly young people from low socioeconomic backgrounds with career trajectories that are characterized by various short-term unskilled jobs and spells of unemployment. Additionally, they have no or low income, low educational achievement, and strained familial relations (Bynner and Parsons, 2002).\nConstructing a positive work-related identity: Although individuals are strongly motivated to maintain their current identities, identities can and do change (Petriglieri, 2011). Literature on identity construction posits that a certain event (such as a threat or a career transition) helps individuals to see the need for identity change and encourages the creation of \"possible selves\": ideas about who one wants to become, likes to become, or fears becoming (e.g. Ibarra, 1999; Markus and Nurius, 1986). During transitions, people often develop multiple provisional constructions of possible new identities (i.e. provisional selves), which remain provisional until they have been rehearsed and refined with experience. These subsequent possible selves are end states that drive identity construction and serve as a future, goal-oriented component of the self-concept. The essence of identity construction lies in bridging the gap between one's current and one's desired identity, which is achieved by strategies such as identifying and imitating role models, experimenting with new behaviors, and evaluating progress toward the desired identity (Ibarra, 1999, 2003, 2005).\nThe current study: How do disadvantaged young adults construct a positive work-related identity? The current study takes a qualitative approach to answer this question. We deliberately chose to use a qualitative approach for three reasons. First, a qualitative approach is particularly useful for exploratory investigations or when existing research does not result in clear theoretical expectations (Rynes and Gephart, 2004). Second, identities are often articulated as narrative stories and are strongly shaped by one's motives (Ibarra and Barbulescu, 2010; Ibarra, 1999; Savickas, 2012). As such, a qualitative approach can provide adequate information for exploring and understanding how motives and social context can shape the process of constructing a positive work-related identity. Finally, when it comes to designing reemployment interventions, studying individual narratives of the experience of a successful reemployment program may provide better insights than solely examining the effectiveness of these programs.\nMethods: Context\nResults: Two main findings became apparent in the course of our analyses. First, the data revealed that participants were not necessarily driven by specific motives or a desired \"possible self.\" Instead, a new identity seemed to emerge during the program rather than actively pursued. Second, the data revealed several ways in which participants' work-related identity emerged. These pathways corresponded to previous research on identity construction (e.g. Ibarra, 1999; Pratt et al., 2006; Van Maanen, 1998). However, our data indicated that the identity construction of disadvantaged young adults additionally required that they decreased the salience of their former unemployed identity in order to consolidate their emerging work-related identity.\nDiscussion: The transition from disadvantaged young adult to employed young adult with positive career perspectives is a particularly difficult one. This study explored how disadvantaged young adults constructed a positive work-related identity during their participation in a reemployment program, and what enabled or constrained a successful transition to employment. The results revealed that the construction of a positive work-related identity was not necessarily preceded by clear motives or a \"possible self\" that functioned as a goal-directed component (Ibarra, 1999; Markus and Nurius, 1986). Rather, participants' work-related identity emerged through four pathways after the program had started.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore how disadvantaged young adults construct a positive work-related identity in their transition from unemployment to employment, and what enables or constrains a successful transition.\nDesign/methodology/approach: The authors conducted semi-structured interviews with 29 apprentices of a reemployment program (Jamie Oliver's Fifteen). The qualitative data were complemented by data on participants' reemployment status one year after the program ended.\nFindings: Identity construction was not preceded by clear motives or \"possible selves.\" Rather, serendipitous events led to participation in the reemployment program, after which provisional selves seemed to emerge through different pathways. The data also suggested that disadvantaged young adults had to discard their old selves to consolidate their new identity.\nResearch limitations/implications: A successful transition from unemployment to employment may require that old selves must be discarded before new selves can fully emerge. Given that our qualitative design limits the generalizability of the findings, the authors propose a process model that deserves further empirical examination.\nPractical implications: A clear employment goal is not always required for the success of a reemployment intervention: interventions should rather focus on accommodating the emergence and consolidation of provisional selves. Yet, such programs can be simultaneously effective and unhelpful: especially group identification should be monitored.\nOriginality/value: Most research assumes that people are driven by specific goals when making a transition. The current study shows otherwise: the factors that enable or constrain a successful transition are not to be found in people's goals, but rather in the process of identity construction itself.\n",
"role": "assistant"
}
] |
Parental engagement in school-based health promotion and education
|
[
"Professionals",
"Parents",
"Child and adolescent health",
"School health promotion"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nBackground: School-based health education programmes make an important contribution to health and education outcomes for children and young people (Lucarelli et al., 2014; Langford et al., 2015; Barry et al., 2017). Recent research identifies positive benefits across a range of health and education measures, including improvements to nutrition, levels of physical activity, social well-being and bullying, learning and academic achievement (Lucarelli et al., 2014; Langford et al., 2015; Barry et al., 2017). The significant role schools play in health promotion and education is underscored by the World Health Organisation's Health Promoting Schools (HPS) framework, which recognises the value of both the taught curriculum and broader school environment to the achievement of positive health (World Health Organisation, 1998).\nThe study: This exploratory qualitative interview study was conducted in a socially and economically deprived inner-city area in England. As evidenced, deprivation not only determines health and education outcomes (Marmot et al., 2010), but has been found to influence engagement in health education and related practices (Kim, 2009) for both children and adults. The urban area selected was below regional and national indicators on all health and social care indicators, including life expectancy, level of development at age five, and the number of children living in poverty (Public Health England (PHE), 2016). In total, 71.5 per cent of the population identified as white British, compared to the England average of 89.3 per cent. The area includes a proportionately higher number of looked after children than regional or national averages (PHE, 2016) and 41.8 per cent of primary age pupils in the study locality were eligible for the pupil premium (Department for Education (DFE), 2016). The pupil premium is government funded and provides free school meals for eligible pupils, as well as extra funding to schools on a per capita basis.\nFindings: Ten professionals participated in interviews all of whom were female. Table I provides the professional backgrounds of participants. Interviews ranged from 16 to 39 min with a mean interview time of 30 min. Three main themes were identified from interviews: topical health issues for school aged children; successes and enablers to school health education and promotion; and challenges and barriers to parental engagement.\nBarriers and challenges to parental engagement: Despite providing many positive examples of successful health promotion and engagement with parents, all respondents described challenges. Changes that had taken place to the commissioning of health services in schools and availability of school nurses, financial austerity and the dominant target setting approach (e.g. focus on obesity measures such as recording children's body mass index) was viewed as having a deleterious effect on the ability to build and maintain partnership arrangements with parents:We've been squeezed with the government cut backs. The nurse used to be in every Friday and she used to have various children that she would weigh. She would talk to various kids for one reason or another and they were brilliant at coming in and giving a talk(Participant 6).What we're finding when we go into schools and we say, \"Well who's your school nurse?\", \"Well I haven't a clue, I don't know who the school nurse is, they've not been in\". Or Heads get very cross because they'll say, We've not seen them for ages, we have no idea what they're doing. They might drop in, but we don't know they're coming [...](Participant 9).\nDiscussion: Findings from this study provide insights into the perspectives of health and education professionals on parental engagement school health education and promotion and highlight the main priority areas professionals identify for supporting children's health. The results indicate similar issues and concerns across professional groupings, particularly in relation to key health challenges for children and families. Many of the identified issues reflect dominant public health discourse and the wider social determinants of health for school-aged children, including concerns about healthy eating and physical activity (Marmot et al., 2010; PHE, 2015).\nConclusions: This study underscores the importance of engaging parents in school health education to promote better child health outcomes. Professional perspectives on parental engagement signal both challenges and opportunities when working with parents and families, particularly in contexts of multiple disadvantage and poverty. Sustained financial and strategic leadership are needed to foster partnerships with parents and to demonstrate a commitment to the improvement of child health practices and outcomes.\n",
"role": "user"
},
{
"content": "Purpose: Children's health and life chances are affected by many factors, with parents and schools holding influential roles. Yet relatively little is known about parental engagement in school-based health education and specifically, from the perspectives of health and education professionals. The purpose of this paper is to examine professionals' perspectives on parental engagement in school-based health education.\nDesign/methodology/approach: An exploratory qualitative study was conducted with ten health, education and local authority professionals from a socio-economically deprived area in England. Semi-structured interviews explored the role of professionals within the school health curricula, roles that parents played in school health, and barriers and enablers to parental engagement in school health education.\nFindings: Reported barriers to engagement related to assumptions about parents' own health behaviours, impacts of funding and inspection regimes, and protected time for health within the school curriculum. Enablers included designated parental support workers based in the school, positive role modelling by other parents, consultation and engagement with parents and a whole school approach to embedding health within the wider curriculum.\nPractical implications: Findings from this study suggest the importance of building meaningful partnerships with parents to complement school health education and improve child health outcomes.\nOriginality/value: This paper addresses an important gap in the research on parental engagement in school-based health education from the perspectives of health and education professionals. Effective partnerships with parents are crucial to the success of school health education.\n",
"role": "assistant"
}
] |
Factors influencing research data communication on Zika virus: a grounded theory
|
[
"Research data",
"Scientific communication",
"Open data",
"Data sharing",
"Open science",
"Academic culture",
"Data communication"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Deriving from Garvey's (1979) scholarly communication notion, which includes the ideas of production, dissemination, and use of information, the concept of research data communication in this paper refers to the processes of production, sharing, and use of data for the development of a given field of research. Initiatives to systematize such processes were documented even in the 1950s (Ruttenberg and Rishbeth, 1994), and in the 1970s, this topic was already found in information science literature (Sondergaard et al., 2003). In the 1990s, the discussion broadened from the Human Genome Project (HGP), which invited researchers from all around the world to share their data so as to contribute to the sequencing of the human genome (Zatz, 2000).\nMethodology: This research adopted a systematized review and a grounded theory as its methodology. The systematized review of the literature was elaborated according to Grant and Booth (2009) proposal. Thirteen selected articles were selected from the Library and Information Science Abstracts (LISA) database (Table I). The search was conducted in mid-2016, through the search argument ((ti (\"research data\") OR ti (\"scientific data\")) AND sharing). The criterion used for the search privileged the documents that represent the subject of the sharing of the research data as the main subject and that reflected this subject in the title of the paper. The analysis systematized from a form[1], allowed to identify a preliminary categories system about the theme. The Gephi software was used in the ordering of categories.\nFactors that influence research data communication, according to information science literature: Scientific literature in the field of information science took a role of preliminary source of information on the factors, which influence communication of research data. The systematized review of the literature allowed the identification of 17 factors that influence the processes of communication of research data. The result was obtained by performing three steps. The first consisted in the free labeling of the elements. In the second step, all elements were translated into similar terms called preliminary categories. In the third step, the Gephi software was used for to connect the categories and the authors (Figure 2).\nGrounded theory on factors influencing research data communication on Zika virus: Research data communication has been pointed to as a phenomenon capable of improving the flow of communication and optimizing the processes of science, as Gezelter (2009), Lopez (2015) and Hey and Trefethen (2008) indicate. Despite its benefits, open communication of data is still not the reality of research practice and, when it does take place, it is expressed with singularities. In the context of international guidelines for open communication of publications and data concerning the Zika virus theme, identifying the factors influencing this process becomes appropriate. From the methodology described, seven factors and their respective subfactors were identified that preponderantly influence research data communication on the Zika virus. Figure 3 illustrates the grounded theory resulting from this research, in which factors, subfactors, and their relations are presented. The relations are indicated by arrows, verbs, and letters from A to N, in a scheme guiding the explanation of the following theory.\nConclusions: A selective encoding strategy elaborated the theoretical model representing factors influencing research data communication, and leading to the proposition of a substantive theory. The theory, hence, indicates relations among the categories and subcategories in the process of explaining the studied phenomenon.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to propose a theoretical model to illustrate factors influencing research data communication on diseases caused by the Zika virus.\nDesign/methodology/approach: The grounded theory methodology was adopted. For data collection, interviews were conducted with 13 Brazilian researchers involved with the Zika virus theme. Data analysis was performed using the open, axial and selective coding processes, according to the principles of the grounded theory.\nFindings: Based on data collection and analysis, seven theoretical categories representing intervening factors in Zika virus research data communication were identified. The emerging theory showed the centrality of researchers' expectations for acknowledgment and reward, influenced by the aspects of research funding and academic culture. Three factors involved in the macro processes of research data communication: data use, data production and data sharing. In the use and production processes, factors related to collaboration figured most strongly. In the production and sharing processes, the factors regarding data processing and the use of research data repositories were more pronounced. Finally, data sharing, and the possibility of reusing data are directly affected by the social context of Zika virus disease as an emerging disease.\nOriginality/value: The study presents a theory developed systematically to explain the phenomenon of communication of research data on Zika virus. The theory presents a set of intervening factors of the process of communication of research data and discusses the factors in light of the fundamentals of information science.\n",
"role": "assistant"
}
] |
Regulatory focus and message framing's effects on intention to consume ethnic food in China
|
[
"China",
"Food neophobia",
"Regulatory focus",
"Message framing",
"Ethnic food",
"Regulatory fit"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Ethnic food consumption is gaining traction worldwide. The market for ethnic foods is growing rapidly over the years. In China, the demand for ethnic foods from grocery stores and eateries by Chinese middle-income consumers is rising year on year. The increasing demands for ethnic foods have led to the mushrooming of ethnic foods restaurants in Chinese cities, and also increasing importation of ethnic foods from other countries. Additionally, the demands for ethnic foods are not limited to domestic market. In 2017, Chinese tourists traveled overseas on 131m occasions (China Tourism Research Institute and Ctrip Tourism Group, 2018). Data from the International Association of Tour Managers show that overseas travel spending by Chinese tourists reached $261.1bn in 2016 and ranked 1st worldwide. While traveling in foreign countries, Chinese tourists were found to consume local foods (Chang et al., 2010; Li et al., 2011). According to the \"Outbound Chinese Tourism and Consumption Trends: 2017 Survey,\" which is jointly issued by Nielsen and Alipay, Chinese tourists travel overseas mainly for leisure, such as dining and shopping, with 55 percent enjoying food as their top priority. The top 3 categories of expenses by outbound Chinese tourists are shopping (25 percent), hotel accommodations (19 percent) and dining (16 percent) (Nielsen and Alipay, 2018). Ethnic food consumption is an important part of destination tourism.\n2. Literature review: 2.1 Regulatory focus and ethnic food consumption\n3. Study 1: The purpose of Study 1 was to test regulatory focus's effect on intention to eat ethnic foods and food neophobia's mediation. It is expected to see those who have higher prevention focus report lower intention to eat ethnic foods, and higher food neophobia which then leads to lower ethnic food consumption.\n4. Study 2: The purpose of Study 2 was to test message framing's moderation in relation to regulatory focus and intention to eat ethnic foods. It is expected to see that those who are prevention focus report higher intention to eat when they are in loss-framing situation than in gain-framing situation. And those who are promotion focus report higher intention to eat when they are in gain-framing situation than in loss-framing situation.\n5. General discussion: The primary purpose of this study was to examine the effect of regulatory focus on ethnic food consumption, the mediating role of food neophobia and the moderation role of message framing in relation to regulatory focus and ethnic food consumption. Results from the experiments carried out showed the following conclusions. First, regulatory focus can influence consumer's ethnic food consumption. Consumers with different types of regulatory focus show a discrepancy on the intention to eat ethnic foods. Promotion-focus consumers have a higher intention to eat ethnic foods than prevention focus. Second, food neophobia plays an important mediation role in the relationship between regulatory focus and ethnic food consumption. Compared with prevention-focus consumers, promotion-focus consumers have lower food neophobia, and lower food neophobia leads to higher ethnic food consumption intention. Third, regulatory fit can increase consumer's intention to eat ethnic foods. Regulatory fit forms when promotion-focus person reads gain-framing message, or prevention-focus person reads loss-framing message. Then, promotion-focus consumers show higher eating intention in gain-framing situation, and prevention-focus consumers show higher eating intention in loss-framing situation. Regulatory fit does not change a product's objective value, but it can increase the consumers' subjective evaluation of ethnic foods. The reason being regulatory fit makes the consumers transfer the value from regulatory fit to the ethnic foods.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the effect of regulatory focus on Chinese consumers' intention to consume ethnic food, the mediating role of food neophobia and the moderating role of message framing with regard to regulatory focus and ethnic food consumption.\nDesign/methodology/approach: Lab experiments method was used in this paper, two studies were designed to test the relationship between regulatory focus, food neophobia, message framing and intention to eat ethnic food. Study 1 was to test the influence of regulatory focus on intention to eat ethnic food, and the mediation role of food neophobia. Study 2 was to test the moderation role of message framing.\nFindings: Results indicated that consumers with promotion focus have higher intention to eat ethnic food than consumers with prevention focus. Prevention-focus consumers have higher food neophobia, which leads to lower intention to eat ethnic food. Food neophobia plays the mediating role in the relationship between regulatory focus and intention to eat ethnic food. Regulatory fit can increase consumers' intention to eat ethnic food. Promotion-focus consumers show higher eating intention in gain-framing situation, while prevention-focus consumers show higher eating intention in loss-framing situation.\nResearch limitations/implications: The study was undertaken in China. Further studies should include respondents living in countries other than China.\nPractical implications: This research provides a venue for marketers of destination tourism, especially for ethnic food marketers to introduce and advertise ethnic foods to tourists. Regulatory fit is important for destination tourism. To improve consumers' eating intention, this research suggests that ethnic food marketers should pay attention to regulatory focus of consumers from different regions and cultural background, and design corresponding message framing for consumers with different regulatory focus to form regulatory fit.\nOriginality/value: First, this study has proposed and tested regulatory focus' effect on intention to consumer ethnic food. Food neophobia is used to explain the mechanism of relation between regulatory focus and intention to eat ethnic food. Also, message framing is introduced to define the boundary of relation between regulatory focus and intention to eat ethnic food.\n",
"role": "assistant"
}
] |
Analysis of reverse logistics implementation practices by South Australian construction organisations
|
[
"South Australia",
"Supply chain management",
"Construction industry",
"Reverse logistics",
"Practices",
"Mixed methodology"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The construction industry is renowned globally as the largest producer and contributor of waste to landfill (Wang et al., 2010). Against this backdrop, the strategies proposed to address this issue have been mostly geared towards improving supply chain management (SCM) by minimising waste and adding value by effective stewardship of information and refining logistics (Cheng et al., 2010; Shin et al., 2011). With reference to the practices proposed for refining SCM, implementing reverse logistics (RL) is regarded as a remedial measure that moderates the detrimental impacts of construction projects on the natural environment (Schultmann and Sunke, 2007b) and enables organisations to be more efficient and effective through the attainment of economic benefits and sustainable competiveness (Agrawal et al., 2015). Similarly, recent studies such as Okongwu et al. (2015) have provided empirical evidence for SCM practices impact on financial and non-financial performance. Conversely in other sectors such as automotive industries in China, Yu et al. (2014) established the relationship between three dimensions of integrated green SCM and multiple dimensions of operational performance. Nevertheless, RL has yet to become common place in the construction sector (Schultmann and Sunke, 2007a; Kibert, 2012; Nunes et al., 2009). Despite anecdotal evidence suggesting local people in Australia have been using materials and components salvaged from old buildings, the uptake of RL is limited in the Australian construction industry. There is consequently a need to explore the ways of promoting RL in Australia. A review of the literature revealed an absence of studies examining the desirable practices of RL in Australia.\n2. Literature review: 2.1. Theoretical framework\n3. Research methods: 3.1. Research approach\n4. Findings: For the sake of brevity and to avoid duplication, the results of the questionnaire survey are presented together with the qualitative findings from the interviews. It is contended that combining the results of quantitative and qualitative provides a more in-depth understanding and enhances confidence in the representativeness of the research findings as argued by Creswell and Clark (2007).\n5. Discussion: The study looked at practices that could instil RL in South Australia. Practices at project, organisational and industry levels were ranked by respondents with further evidence provided by interviewees. Reducing waste generation was considered to be very important at project level while understanding of benefits of deconstruction and challenges associated with deconstruction were considered very important at organisation level. None of the practices were ranked very important for the industry level. Regulatory and financial incentives coupled with quality compliance at the industry level were rejected by respondents as least important with a mean value less than 2.9. The message is very clear that strategic practices which reinforce responsible behaviour were valued more important than mere financial or regulatory incentives. All other practices were ranked having medium importance ranging between 3 and 4 of mean value. There were six project level, two organisation level and four industry-level practices ranked medium. These 12 coupled with the three very important practices could be used to map out a strategy to install RL as a vibrant sub-sector in South Australia. As shown in Figure 2 these practices could impact on each other and form a monolithic force that could lift current RL activities to a higher level establishing it as a formidable sub-sector. The positive relationships among practices are denoted by a + sign in the diagram with thicknesses of the lines indicating their strength.\n6. Conclusions: Through a questionnaire survey and a series of interviews, this study has established some very important practices that could help promote RL in South Australia's construction industry. Through a comprehensive literature review, 18 current practices were identified as having the potential to promote RL in construction. Based on the seminal work of Carter and Ellram (1998), these practices were divided into three levels, specifically, industry, organisation and project. Out of 539 questionnaires sent to professionals working in construction organisations that represent the input, regulatory, output and competitive task environments of RL, 49 usable questionnaires were returned with a response rate of approximately 9 per cent. Six interviews with experts working in the RL sub-sector were considered a supplementary data source which is very in-depth and exploratory in nature. The combined quantitative and qualitative data were very useful for identifying practices that have the potential to promote RL in South Australia.\n",
"role": "user"
},
{
"content": "Purpose: A large number of benefits have been reported when reverse logistics (RL) is fully implemented in the construction industry. However, RL is yet to become common place in the construction sector, particularly in Australia. The particular sub-sector in which RL operates is small and weak and the remainder of the sector must embrace and accommodate it comfortably. Research is lacking on how to promoting RL in the construction industry. Very little has been done to identify the current practices that have the potential to promote RL industry-wide. The purpose of this paper is to identify the practices that work well in the sector, a strategy could be mapped out to promote RL to all stakeholders.\nDesign/methodology/approach: In order to fill the above gap, the present study used a mixed method approach to gather and evaluate current practices and their potential to promote RL in South Australia's construction industry. Practices that were identified using a comprehensive literature review were evaluated with a questionnaire survey and series of interviews involving construction professionals.\nFindings: The findings are that practices facilitating deconstruction is the most important, followed by practices facilitating the use of salvaged materials in new construction to promote RL in South Australia. Awareness of deconstruction benefits, challenges and procedures at the organisation level and facilities and services at industry level were associated with RL implementation. Availability of salvaged materials in the market was found to influence its use in new construction and as a consequence its demand. Designing for reverse logistics is another practice that could facilitate deconstruction and the onus of its promotion lies mainly with the designers.\nResearch limitations/implications: This research was confined to one state in Australia. As such the generalisation to other states and other countries should be treated cautiously.\nPractical implications: The findings of this study can help inform the industry and its stakeholders on areas that they need to concentrate more on to make the South Australian construction industry a fully RL integrated one. To that end the authors propose some recommendations arising from the findings reported here.\nOriginality/value: This study makes a contribution to the body of knowledge on reserve logistics within a previously unexplored South Australian context. In addition, the study provides valuable insights into the contribution of RL practices to the construction industry.\n",
"role": "assistant"
}
] |
A method for generating strategy maps using ANP
|
[
"Balanced scorecard",
"Strategy map",
"Analytic network process",
"Strategic objectives"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The balanced scorecard (BSC) developed by Kaplan and Norton (1992, 1996) is a performance measurement system which has been very popular in the literature as well as in practice. According to the authors it has become a strategic management tool (Kaplan and Norton, 2001a,b).\n2. The method: 2.1 ANP representation of a BSC\n3. An illustration: 3.1 Description\n4. The strategy map: The strategy map is derived directly from the ANP model of Figure 6, in which the strategic objectives are the nodes and the relationships are the arcs of the network. The strategy map of the firm is depicted in Figure 7. It should be noted that the arcs were changed to the opposite directions.\n5. Managerial and practical implications: Kaplan and Norton (1996, 2004) have declared that the strategy map is important, because it is a way of describing the strategy of the company. However, the way of constructing it has not been widely dealt in the literature. Normally, the cause-effect relationships between strategic objectives are generated in a subjective way. Even this way of working is widely accepted in practice, some studied have shown that the declared relationships are not necessarily valid. In order to overcome this situation, the proposed method provides a quantitative tool (ANP) to establish the relationships between strategic objectives. The advantage of using ANP is that it allows the inclusion of dependence and feedback of strategic objectives and perspectives in the map.\n6. Conclusions: This paper presented a quantitative method to establish the causal relationships between strategic objectives of a strategy map of a BSC. It uses the ANP to get those relationships. The method starts with an initial ANP model represented all possible causal relationship between objectives. Then it uses a criterion to select those relationships that are \"important\". In practice, the construction of a strategy map has been done usually using the experience and judgment of the managers of companies. This paper supports that process by using a multi-criteria system, which has a strong conceptual and mathematical base.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present a method for identifying causal relationships in a strategy map.\nDesign/methodology/approach: A strategy map is a visual representation of the strategy of a company, which includes the strategic objectives of a company and the cause-effect relationships between them. Its network structure facilitates its representation as an analytic network process (ANP) model. The proposed method starts with a network with all possible relationships and then it deletes those relationships, which are not important. To do this, it uses the techniques of the ANP approach.\nFindings: It was found that ANP is a good tool for modelling a strategy map and for identifying the important relationships of a strategy map. A study case of a manufacturing is shown to illustrate how the proposed method can be used in practice.\nPractical implications: Normally, the cause-effect relationships between strategic objectives are generated in a subjective way. Even this way of working is widely accepted in practice, some studied have shown that the declared relationships are not necessarily valid. The proposed method provides a quantitative tool to establish the relationships between strategic objectives, which are obtained using a method (ANP) that has a strong conceptual base. This is an indication that the strategy map obtained represents in a better way the strategy of the company.\nOriginality/value: ANP is a methodology which is used to estimate the priority of nodes with in a network. In this work ANP is used to estimate the priority of the arcs of the network. The results of application in a company represent a good indication that the method may be implemented in other manufacturing companies.\n",
"role": "assistant"
}
] |
Intention to use flexible work arrangements: The case of workers in Korea and gender differences in motivation
|
[
"Gender differences",
"Concern about career disadvantage",
"Intention to use flexible work arrangements",
"Motivation for personal life",
"Motivation for productivity",
"Peer behaviour"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Technology advancement and a shift in industrial structure towards knowledge-based management have changed traditional way of working - working at a fixed place of work for fixed hours. Various work organisations adapting to new technologies have been emerged and face-to-face communication to get work done is not necessarily required. As more individuals are seeking work-life balance over salary, job security and promotion, new ways of working that allow employees to determine the time and place of their work have become more apparent (Twenge et al., 2010). Working conditions, such as more diversified working patterns, have been considered a key factor in choosing a job (CEDEFOP, 2010), and flexible work arrangements (FWAs) are the institutionalized programme among various types of new employment practices.\n2. Theoretical background and literature review: 2.1 Theory of planned behaviour and motivations to use FWA\n3. Hypotheses development: The research model was designed to test whether there are differences between the genders in the effect of antecedents on the intention to use FWA in Korea. The proposed research model is shown in Figure 1.\n4. Research methodology: 4.1 Variables and measurement\n5. Data analysis and results: In this study, reliability, validity and regression analysis were performed using STATA 12.\n6. Discussion: 6.1 Findings and implications\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate gender differences in motivations to use flexible work arrangements (FWAs) in Korea.\nDesign/methodology/approach: Based on a literature review on theory of planned behaviour (TPB), this study considers four motivational factors that influence the intention to use FWA: motivation for personal life, motivation for productivity, peer behaviour and concerns about career disadvantages. Survey response data drawn from 92 male and 105 female Korean workers were used to analyse differences by gender.\nFindings: As for the male respondents, all four motivational factors have a significant effect on the intention to use FWA. However, in the female respondents, the effects of concerns about career disadvantages on the intention to use FWA are not significant. The results of gender differences analysis show that significant difference was not found in the effect of motivation for personal life on the intention to use FWA while the other three motivational factors have significant differences by gender.\nResearch limitations/implications: In this research, basing its conceptual background on TPB, a novel approach is taken by introducing motivational factors as the antecedents of intention to use FWA. This is a more systematic view on individuals' behavioural mechanism relating to the intention to choose FWA. It is also meaningful in that this study looks at the intention to use FWA from a broader perspective by suggesting gender differences as critical analysis criteria given the uniqueness of Korean labour market.\nPractical implications: For an effective operation of FWA, it is important not only to launch a flexible working programme itself, but also to ensure that the users are properly understood and fairly evaluated.\nOriginality/value: Considering the motivations of utilising FWA from various angles will contribute to coming up with various measures to raise the use and effectiveness of FWA.\n",
"role": "assistant"
}
] |
Conjunctionitis: a call for clarity in construct definitions
|
[
"Research methods",
"Ambiguity",
"Clarity",
"Construct definition",
"Vagueness"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Over the past half-century, a frequently cited literature has developed regarding the construction of marketing measures (Churchill, 1979; Mowen and Voss, 2008). Because constructs are critical to theory building, clear and unambiguous construct definitions are important (Hempel, 1952). Theorists agree that achieving clarity depends critically on the wording of the definition (Bunge, 1967; Hunt, 1991; MacKenzie, 2003). This study focuses on the use of coordinating conjunctions in such definitions. Conjunctions are important because they are widely used and, as Hempel (1952, p. 15) noted, \"the total vocabulary of any scientific theory contains terms, which belong to the vocabulary of logic and mathematics such as 'not', 'and', 'or' [...] [...], etc.\" While Gilliam and Voss (2013) urge caution in the use of conjunctions, Podsakoff et al. (2016) seem to embrace the conjunctions \"and\" and \"or\" as necessary to develop the structure of the construct. Thus, understanding the effect of using \"and\" and/or \"or\" on the clarity of construct definitions is an important facet for the further development of theory and practice of construct definition and scale development.\nBackground: Hunt (1991) indicates that a good definition is one that possesses, among other things, clarity. When speaking of definitions, clarity means that the definition is differentiable from the definitions of other terms (MacKenzie, 2003). Two principle enemies of definitional clarity are ambiguity and vagueness. Ambiguity results when it is unclear the specific entity to which a term refers to (Teas and Palan, 1997). For example, the word \"noodle\" is ambiguous because it can refer to a wide variety of different types of food. Vagueness applies when a term's definition is imprecisely stated, indefinitely expressed or uncertain (Teas and Palan, 1997). The definition \"fair to all parties\" is vague because the reader has neither sense of what is meant by fairness nor who all the parties might be.\nAmbiguity and vagueness: The term coordinating conjunctions describes a set of words used in sentences to join different elements in some fashion. Seven words, known by the acronym FANBOYS, are considered to be \"coordinating conjunctions,\" namely, for, and, not, but, or, yet and so (Corcoran, 1970). Herein, the focus is on the two most frequently used conjunctions in marketing construct definitions, namely, \"and\" and \"or.\" The word \"and\" is used to join elements of the sentence while the word \"or\" is used to present one or more alternatives. Sometimes, the words \"and\" and \"or\" are used together to express more complex thoughts. Based on the construct definition literature, the framework proposed herein (Figure 1) summarizes potential interpretations of \"and\" and \"or\" on the left side of the diagram. These potential interpretations are applied to construct definitions on the right side of the diagram. Thus, Figure 1 is a guide for use in evaluating proposed definitions. The Web appendix contains a detailed description of Figure 1.\nMethods: The hypotheses require making inferences about published construct definitions. Hence, a random sample was assembled from articles published in the Journal of Marketing, Journal of Marketing Research and Journal of Consumer Research. The sampling plan used is detailed in the Web appendix. The resulting data set contains 736 construct term-definition pairs, 317 definitions (42.4 per cent) that include a total of 424 occurrences of the conjunction \"and\", 196 definitions (26.2 per cent) that contain 247 occurrences of the conjunction \"or\" and 67 definitions (9.1 per cent) that include both conjunctions. Thus, 66 per cent of the definitions contained at least one of these two conjunctions. The conjunction \"for\" appeared in just 85 definitions while the usage of the remaining FANBOYS was negligible.\nDiscussion: We suggested that construct definitions that use the word \"and\" will be judged more ambiguous than construct definitions that do not use the word \"and\" - lay judges' ratings provide support for this, as the main effect of \"and\" on ambiguity is statistically significant. Based on effect sizes, a larger number of judges evaluating a higher number of definitions would likely lead to support for this hypothesis using the academic judges. However, the finding would be qualified by the contrasts between the \"and\" and \"or\" definitions and the other three sets of definitions, which demonstrates that the principle effect in the academic data set is using \"and\" and \"or\" together. We also suggested that construct definitions that use the word \"or\" will be judged vaguer than construct definitions that do not use the word \"or\" - findings for this hypothesis are complex. The ratings supplied by the lay judges revealed a statistically significant main effect of \"or\" on vagueness. Surprisingly, this was qualified by a significant two-way interaction that showed \"and\" was also an important determinant of vagueness. On the other hand, the ratings obtained from the academic judges largely mirror the results of the ambiguity variable; that is, the data indicate that using \"and\" and \"or\" together in the same definition leads to increased vagueness. Thus, when compared to definitions that used neither \"and\" nor \"or\" definitions that used both \"and\" and \"or\" were rated both more ambiguous and vaguer by both the lay and academic judges. A comparison of the lay and academic judging studies is provided in the Web appendix.\nConclusion: This manuscript explores how meanings of expressions that contain two principal coordinating conjunctions, \"and\" and \"or,\" vary in terms of ambiguity and vagueness. Using both of these conjunctions in a construct definition seems to have a significant negative effect on definitional clarity. One key finding is that text readability as measured by ARI affected both vagueness and ambiguity for the lay judges but not the academic judges. As ARI is highly correlated with definition length, shorter definitions seem to be clearer than longer definitions for lay audiences. Definitions that avoid both \"and\" and \"or\" are also judged as clearer. Our findings should prove useful to marketing researchers and other social science researchers. We recommend academic researchers should develop a preference for creating construct definitions that are not easily misunderstood. The current results, a framework of alternative interpretations and five recommendations should help light a path to achieving that noteworthy goal.\n",
"role": "user"
},
{
"content": "Purpose: This study aims to determine the extent to which the use of coordinating conjunctions enhances or impairs definitional clarity.\nDesign/methodology/approach: In two studies, a sample of 736 construct definitions from the Journal of Marketing, Journal of Marketing Research and Journal of Consumer Research over a 30-year period was judged for ambiguity and vagueness by both academic and lay judges.\nFindings: The authors demonstrate that constructing definitions using both \"and\" and \"or\" increases the ambiguity and vagueness of the construct's meaning. The most frequently used conjunction is \"and\" which appeared in 42 per cent of the definitions. A significant percentage (26 per cent) contain the conjunction \"or.\"\nResearch limitations/implications: A framework for understanding alternative interpretations of \"and\" or \"or\" is developed. Five recommendations are proposed for evaluating the use of \"and\" and \"or\" in construct definitions. Theorists in all academic fields should not use both \"and\" and \"or\" in the same construct definition.\nPractical implications: A five-step process is proposed for evaluating the use of \"and\" and \"or\" in construct definitions. Theorists should not use both \"and\" and \"or\" in the same construct definition.\nOriginality/value: This is the first exploration of how specific wording patterns used in construct definitions in academic research affect the clarity of the definition.\n",
"role": "assistant"
}
] |
Which industrial product service system fits best? Evaluating flexible alternatives based on customers' preference drivers
|
[
"Services",
"Industrial services",
"Decision making",
"Customers"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Challenges in business-to-business markets: Companies from established industrial nations are faced with a multitude of threats, caused especially by companies from developing nations such as India or China. In the past, these threats were primarily based on the common practice of imitating products of competitors from developed, industrialised nations. These imitations exacerbate the amortisation of investments in research and development and can even render them impossible. Growing capabilities and competencies of such competitors from developing nations pose a further threat, since companies from developed industrial nations are unable to compete with the low-labour costs of the aforementioned companies. Highly dynamic markets pose the additional challenge of having to generate sustainable competitive advantages under changing conditions. Focusing on providing products does not suffice to create a viable economic basis for company success (Prahalad and Ramaswamy, 2003). Markets have experienced a shift of focus from products to market requirements and an augmentation of the importance of services. Encompassing this, significant effort is dedicated to an interwoven integration of products and services in order to generate a sustainable competitive advantage and prevent out-suppliers from penetrating the customer-supplier relationship (Manzini and Vezzoli, 2003).Against this background of changing environmental conditions, we suggest industrial product service systems (IPS2) as a possible solution. IPS2 can be described as \"as a marketable set of products and services capable of jointly fulfilling a user's need\" (Goedkoop et al., 1999). But when developing an IPS2 for a specific customer, it is important to design it tailored to the individual customer's preferences, because otherwise the customer will rather choose to invest in an alternative from a competitor, or a lower price can be charged. As customers may not have any experiences with IPS2, they are neither sure about their own preferences, nor are their preferences stable. Hence, the supplier cannot analyse the preferences by just asking the customer. Instead, the supplier has to analyse the factors or drivers, respectively, that determine the customer preferences and deduce the customer preferences from this drivers.This paper discusses those preference drivers and their influences on the customers' preferences. We also consider flexibility in an IPS2 context based on life cycle-oriented perspective. Section 2 provides a definition of IPS2, with a specific focus on the possible dimensions of IPS2 configurations, and introduces a method to determine the optimal IPS2 regarding both the customer and the supplier side. In Section 3, we explain the importance of different customer drivers and identify their respective impact on customers' decisions. Section 4 broadens this focus and includes the possibility of a dynamic adoption of an IPS2 to changing customer needs along the life cycle. The paper concludes in Section 5 and gives an outlook onto further fields of research.\n2 IPS2 characteristics: IPS2 are characterized by an integrated and mutually determining process of planning, developing, provisioning and using of goods and services (Meier et al., 2005). They constitute a problem solution for business-to-business markets, customized to individual customers' needs along the IPS2 life cycle. This customization creates advantages for IPS2 suppliers (Mont, 2001). First of all, the customized IPS2 diminishes the threat of production-cost advantages on the part of competitors from emerging markets. Second of all, the integration of customer and supplier processes diminishes suppliers' substitutability. Process integration, understood as a linkage of customers' and suppliers' value chains, creates barriers to entry for competitors. And finally, the capacity to analyze the customer process and identify the sort of product-service modules needed is a competence that cannot easily be reproduced in equal and sufficient quality by other suppliers, above all suppliers from emerging markets. Particularly these knowledge-based capacities have been ascribed with both high sustainability and relatively more difficult possibility of imitation (Prahalad and Hamel, 1990). Consequently, customizing IPS2 serves to meet the aim of establishing a customer-supplier relationship which cannot be easily broken up by out-suppliers.The potential of customizing IPS2 is primarily based on the possibility of partially substituting product and service components to meet customer requirements. This allows for various possible ways of executing customer processes, service-based or product-based. We label these technological possibilities as different mixtures of manual and automatic execution of processes. Furthermore, a second dimension has to be considered. This dimension describes, the customer decision towards make or buy of processes. Both dimensions have a continuous scale which means that a process can be executed partly manually and partly automatically or partly made and partly bought, respectively.The following simplified example shall serve to clarify uses of IPS2 and the liberties gained regarding suppliers' possible problem solutions.A company engaged in micro technology produces insulin pumps, which are to be implanted underneath the skin. Using a micro-technological turning lathe, small gears have to be produced as parts of the pumps. From a technological perspective, the question of how to exchange the spindle involved in the process arises. A first option is to let the exchange be carried out by a trained employee. In this case, the process is performed manually and by the customer. A second option would be to let an untrained employee exchange the spindle, guided by a specialist via stereoscopic goggles and information technology (IT) support. This kind of process execution is partially manual, partially automated, and it is performed partially by the customer and partially by the supplier. Third, the inclusion of technological devices carrying out the spindle exchange without human interference is an option as well. In this case, the machine status is monitored by sensors and evaluated via IT devices. Reactions to changes in conditions are carried out by means of mechatronic devices. This case describes a fully automatically performed process executed by the customer (for a different example see Sheridan, 1980, 2002; Sheridan et al., 1997).This two-dimensionality, the variability of technology on the one hand, and the decision of internal or external production on the other hand, generates additional degrees of freedom for customers and suppliers. As a consequence, a greater variety of potential problem solutions could be offered to customers.By offering an IPS2 the aim of the manufacturer is usually not to optimize the customers' value, but its own. Like the customer wants to maximize the value that the IPS2 adds to the company, the supplier wants to offer the IPS2 that leads to the highest value added to the supplier. While the customers' IPS2-value for the customer depends on the in-payments and out-payments induced by the IPS2, the suppliers' value depends on the supplier's in-payments and out-payments induced by the IPS2. The in-payments are caused by the price the customer is willing to pay for the IPS2 and its maintenance. The out-payments are induced by the development, production and maintenance. Generally, a supplier will only offer an IPS2-configuration whose cash flow are expectedly high enough to cover the capital costs, because otherwise the offer will lead to an unacceptable depletion of company value. Hence, the array of possible solutions is constraint by the requirement of a positive value of the IPS2 for the customer. Furthermore, one has to consider that not all possible combinations of make/buy or manual/automatic execution of processes are economically or technically feasible, or there may be legislative or social restrictions that constrain the scope for possible problem solutions. To sum up, problem solutions have to:* be technologically viable;* generate a positive customer value which is superior to the value of the best competing offer; and* ensure sustainable profitability for the supplier.Only IPS2 that meet all of these requirements should be developed and offered to the customer (Figure 1).Within the space of possible IPS2-solutions, the supplier has to find the solution which leads to the highest benefit/value for the supplier, taking into consideration the different willingness-to-pay and the different costs conditional to the different solutions within the solution space. This makes clear that the supplier has to take into consideration both the customers' expected willingness to pay as the suppliers' revenue as well as the expected costs in order to find the optimal IPS2-solution along the solution space, whereas willingness-to-pay and costs are interdependent. An example of such an IPS2 configuration regarding the two IPS2-dimensions is shown in Figure 2. It is important to note that in most cases an IPS2 consists of several sub processes. In this case, each sub process has its own solution space and has to be handled separately. In this example, the IPS2 consists of six processes (referred to as CP in Figure 2) with different choices of the dimensions manual/automatic process execution and make or buy.The value generation on the customer side which is decisive for his willingness-to-pay is based on customer value drivers. Depending on these drivers, customers will prefer manual/automatic process execution going alongside with a make/buy decision as a solution for a certain process. In the following, we put an emphasis on the customers' point of view and discuss the identification of the aforementioned drivers.\n3 Analyzing the effects of main drivers: 3.1 Effects of drivers on customer preferences and buying behaviour\n4 The importance of flexibility for the IPS2 configuration: As previously discussed there are various drivers whose strength determines the configuration of an IPS2. When taking into consideration today's dynamic markets, these drivers and their respective strength can be subject to change over the IPS2 life cycle. Possible changes of value drivers impact on in-payments and out-payments on the customer side. They may be induced by changing environmental conditions (such as amendments, technological developments, etc.) on the one hand, as well as by changes in customer structure (e.g. change of strategy, re-orientation in the market) on the other hand. Such changes in drivers and driver strength can result in a situation, in which an IPS2 configuration which was initially deemed best by the customer is now only considered suboptimal.Consequently, a shift in advantageousness of different solutions may take place, manifesting in changes of the IPS2 solution space. This is shown in Figure 4 by comparing solution spaces for initial IPS2, IPS2 in the middle of the lifecycle and IPS2 towards the end of the same. The three stages as well as the migration of the hashed area to the bottom left merely serve the purpose of exemplification. The occurrence of driver changes is explicitly taken into consideration along the IPS2 life cycle.As states above, changes of customer preference drivers can heavily influence the value which an IPS2 generates for an individual customer and thereby its advantageousness. Against this background suppliers are faced with three main challenges:1. they have to anticipate changes in value drivers as best as possible already during the delivery and use phase of IPS2 in the early phase of product development;2. they have to include switching options from an initial IPS2 configuration to another IPS2 configuration into the initial design; and3. they have to make sure that including such switching options do not threaten economic feasibility of the respective IPS2 for customers and suppliers alike.Suitable switching options allow customers to adjust their IPS2 to their current needs. In doing so, their current preferences for IPS2 are better met than would have been the case without adjusting the IPS2. This results in a higher value going along with a higher willingness-to-pay (for the enhancement of the value of an investment by such decisions Brealey and Myers, 2002). Besides, the higher value of the IPS2, these alterations over time reduce individual customers' risk which purchasing an initial IPS2 brings along with it.The optimal initial configuration of an IPS2 which includes switching options may deviate from the optimal initial configuration of an IPS2 which does not include switching options. If no option of switching the IPS2 configuration exists, an initial configuration would have to be able to meet customer demands at the beginning of the life cycle, but must also not become useless in case of driver changes. Hence, trade-offs in the configuration are likely to result, so that the initial configuration is likely to be suboptimal to the best initial configuration, but is still able to satisfy changing customer needs over the life cycle to a certain degree. If options of switching and adjusting the initial configuration do exist, the best initial configuration can be chosen, as it can be directly adjusted to changing customer drivers if needs be.In order for suppliers to know which kind of flexibility to offer their respective customers, they not only have to gather information to try and predict possible future changes in customer drivers, e.g. owing to a new strategic orientation of the customer. It is also of great importance to estimate the value that different kinds of flexibility have for individual customers.However, this flexibility not only affects the customers' willingness-to-pay, but also the suppliers' costs for the following reasons:1. flexibility requires holding back enough capacity to perform the switching options whenever the customer wants to change the IPS2-configuration;2. flexibility may increase the costs of development and production of the IPS2; and3. if the customer wants to reconfigure the IPS2 and the supplier has to perform changeovers of the IPS2, this again will lead to additional costs on the supplier side.All of those three aspects have to take into consideration when determining the IPS2-solution space. The supplier will only offer flexibility if the customers' additional willingness-to-pay for the flexibility is higher than the additional expenses for the supplier.Figure 5 shows possible changes in the initial IPS2 configurations of the sub processes over the life cycle.\n5 Conclusion: As shown in this paper, IPS2 could provide a solution for counteracting threats in business-to-business markets. An important characteristic of this approach is that it renders a possible offering of various solutions for a customer problem and the possibility to adopt this solution to changing customer needs. The superiority of certain problem solutions over other solutions depends on customer drivers and their respective strength. These are neither stable in their attractiveness when considering a set of possible solutions at a fixed point in time due to customer heterogeneity, nor when focusing on an individual customer over time due to market and customer development dynamics. To be able to identify customers' IPS2 preferences, we identified the drivers for these preferences and examined their effects regarding the dimensions automatic or manual process execution on the one hand and make-or-buy on the other hand. Our results are based on the assumptions we made in Section 3.1. These, however, must be considered as a starting point for further research. The drivers identified and their respective effects have to be examined empirically with regard to their completeness and especially their interdependency. We have shown that drivers can work in completely opposite directions, which causes the need to understand how these drivers interact with each other. Generating this understanding through empirical investigations with special regard to IPS2 poses a promising field for future research, with which the preferences of IPS2-customers could be captured to a far greater extent than is the case today. Determining requirements for including flexibility into the initial IPS2 configuration especially from a technological point of view (fields such as concept modelling, planning and developing could be of special interest) is another promising field for future research. Only investigating into economic as well as technological aspect and combining the results obtained can warrant a successful implementation of IPS2 flexibility as proposed in this paper.\n",
"role": "user"
},
{
"content": "Purpose: - Industrial product service systems (IPS2) are required to meet current customer needs in order to provide solutions to current customer problems. Furthermore, adaptability to changing customer decision drivers is required in order to account for customers' preference changes over time. The purpose of this paper is to present an approach where customer preference drivers for different IPS2 are identified and their directions analyzed.\nDesign/methodology/approach: - A theoretical, literature-based analysis, drawing on different fields of economic research, is made use of. A tool to anticipate customer decisions based on the identified preference drivers is devised.\nFindings: - Nine main preference drivers are identified. These are customer resources, know-how, number of employees, core competences, value-based figures, process intricacy, process standardization, process frequency and process significance.\nResearch limitations/implications: - The drivers identified on a theoretical basis should be investigated empirically. Special consideration should be given not only to the direction of their effect, but also their strength and interdependency.\nPractical implications: - The paper helps IPS2 suppliers to analyze customers' preferences and helps decide about the design and the business model for the IPS2.\nOriginality/value: - The paper is the first paper that focuses on the relevant influence factors of a customer's decision about the IPS2-design.\n",
"role": "assistant"
}
] |
How purchasing and supply management practices affect key success factors: the case of the offshore-wind supply chain
|
[
"Capabilities",
"Supply chain",
"Renewable energy",
"Contracts",
"Competition",
"Make-or-buy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The effective use of global sourcing can help companies achieve superior firm performance (Lindgreen et al., 2009, 2013). We consider how innovative purchasing and supply management practices might increase industry-level performance too, with a focus on the offshore-wind industry. Offshore-wind offers a promising renewable form of energy, such that the industry has experienced constant growth in recent years (GWEC, 2013). In efforts to improve energy-generation capabilities, firms have built more and larger turbines, a trend that has repercussions for the offshore-wind market, its technology and the complexity of the supply chain. With a supply chain-level analysis, we identify the innovative use of three purchasing and supply management practices - make-or-buy decisions, contract forms and local-to-global sourcing decisions - that can address supply chain complexities. On the basis of key success factor theory (De Vasconcellos E Sa and Hambrick, 1989), we propose a conceptual framework in which innovative purchasing and supply management practices represent the strengths that firms acquire to achieve key success factors for the industry. With the real-world industrial evidence from recent offshore-wind farm projects, including the London Array farm, we identify three key success factors in the offshore-wind industry: competition, capabilities and control.\n2. Theoretical background: Previous studies, drawing on resource-based theory, suggest that purchasing and supply management practices enhance firm-level performance (Lawson et al., 2009). Zimmermann and Foerstl (2014), by conducting an extensive analysis of previous literature, indicate a strong support for the positive relationship among purchasing and supply management practices and firm performance. We argue that innovative purchasing and supply management practices can increase both firm- and industry-level performance. We use the key success factor theory to motivate our conceptual framework and its application to the offshore-wind industry (Figure 1). De Vasconcellos E Sa and Hambrick (1989) suggest that key success factors in an industry derive from its underlying characteristics, the context of its market, and the technology available. They theorize that firms whose strengths match the key success factors of their industry achieve better performance. The authors use a broad definition of strengths, including knowledge and capability types, with purchasing classified among the latter. We argue that the growth of offshore-wind developments requires new technological advances and affects the market but also increases the complexity of the supply chain, as we detail in Section 3.2. To respond to such changes, companies operating in this industry should adapt by innovating their purchasing and supply management practices. In our conceptual framework, innovative purchasing and supply management practices constitute strengths, from key success factor theory. By reviewing previous and current projects in Section 3.3, we argue that they are consistently associated with make-or-buy decisions, contract forms and local-to-global sourcing decisions. These innovative purchasing and supply management practices match industrial-level key success factors and allow companies to achieve high performance. From our industrial evidence, described in Section 3.4, we identify competition, capabilities and control as key success factors for the offshore-wind industry.\n3. Industrial application: In this section, we describe the industry application of our conceptual framework. First, we introduce the methodology of this study. Next, we describe the offshore-wind industry and its supply chain. Then, we present the results of our analysis in relation to purchasing and supply management practices and key success factors for the offshore-wind industry.\n4. Discussion: In this section, we compare our findings with the results in previous literature. For example, our argument related to make-or-buy decisions, namely, that firms should decrease their level of vertical integration, is consistent with Corbett and Karmarkar's (2001) finding that the integration of successive oligopolists in a serial supply chain reduces profits. We also propose that manufacturers should be less protective of their intellectual property and share their knowledge with suppliers to increase capabilities. As suggested by Argyres (1996), the supply chain will bear higher costs in the short run, but once suppliers have developed enough capabilities, costs decrease. Moreover, similar to Novak and Stern (2009), we find that product complexity leads firms to build many components in-house, to increase their control over the processes.\n5. Further research: This exploratory study identifies several areas for further academic research. We outline three, associated with each key success factor for the industry. First, academics might advance the debate about competition by developing bottleneck analyses specific to an industry, to identify which supply chain phases and activities benefit most from increased competition. Researchers also might study the role of competition in the supply chain through equilibrium analyses, such as by considering how the market varies in its degree of vertical integration and presence of exclusivity agreements. Second, research should address the process by which companies develop capabilities, and offshore wind-specific capabilities in particular. Such investigations might also identify appropriate models for transferring skills across companies. Third, we call for studies that derive decision support systems that can manage the supply chain risks that affect the offshore-wind industry. Identifying appropriate risk-sharing models among companies represents an open and interesting question for contract-related research.\n",
"role": "user"
},
{
"content": "Purpose: In reference to the offshore-wind industry, this study aims to show that innovative purchasing and supply management practices can increase both firm- and industry-level performance. This paper also includes a description of the offshore-wind supply chain, which remains understudied in academic literature, despite increasing global development of offshore-wind farms.\nDesign/methodology/approach: Offshore-wind farm projects use more and larger turbines, which greatly increase the complexity of the supply chain. Innovative purchasing and supply management practices, designed to tackle this growing complexity, could help companies achieve the key success factors that define this industry. The evidence comes from real-world, offshore-wind farm projects, with the London Array farm as a principal example.\nFindings: Innovative purchasing and supply management practices include decisions to make or buy, contract forms and local-to-global sourcing. These practices affect the key success factors of the industry by increasing competition, capabilities and control.\nOriginality/value: Purchasing and supply management practices could affect industry-level performance. This paper is among the first ones to provide an analysis of the offshore-wind supply chain and its evolution.\n",
"role": "assistant"
}
] |
Three conceptual models of board role performance
|
[
"Board of Directors",
"Management roles",
"Performance management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: While boards have previously been regarded as passive (Mace, 1986), the growing power of investors and the media have to a larger extent pressured boards to play a more active role in company affairs (Coles et al., 2001; Judge and Reinhardt, 1997). The demands made on directors have grown significantly and the issues with which they have to deal with have widened by a great extent. Boards can no longer enjoy the luxury of passivity (Lorsch and Khurana, 1999). They need to be active players in shaping the companies. Directors have no choice but to perform their roles seriously (Cadbury, 2002). Board role performance is therefore an important area of research.\nDefinition and dimensions of board role performance: In this section, a general definition of board role performance is first provided. The various board roles are then discussed in more details.Board role performance is generally denoted as a board's ability to perform its roles (Cornforth, 2001; Forbes and Milliken, 1999). Based on the literature, there are four critical board roles: monitoring; service; strategy and resource provision (Daily et al., 2003; Zahra and Pearce, 1989).Monitoring\nThree conceptual models for studying board role performance: In the previous section, the definition of board role performance is provided. The four board roles from the literature - namely monitoring, service, strategy and resource provision - are detailed. In addition, three contextual factors, company size, industry type and life cycle of company are used to explain potential variations in role performance.In this section, the three conceptual models for board role performance are discussed.The structure model\nPotential contributions and conclusion: Past literature on board research centred on board structure and firm performance. Since empirical studies do not reveal a conclusive relationship between the two, attention has shifted towards board role performance. This paper examines three conceptual models for the studying of board role performance: structure, process and mediation models by developing 17 propositions for empirical research. The paper concludes by discussing the potential contributions.First, this study examines various aspects of board role performance. This is in contrast to previous research which largely investigates board role performance in general or under the dominant agency perspective which emphasises the board's monitoring role. Daily et al. (2003, p. 375) for example noted that \"In addition to the monitoring role, directors fulfil resource, service, and strategy roles\". Examining all four roles together in a single study has the potential to provide a richer perspective into board research.Second, unlike traditional governance models, the role of board process is explicitly advocated in this paper. Despite its importance, board process is still an under-investigated research issue. The first reason is that traditional governance research tends to emphasise on the structure-performance relationship. Board process is often inferred from board structure (Finkelstein and Hambrick, 1996; Pfeffer, 1983). Second, the tradition of board members being cronies of the CEO has not disappeared, and boards often are reluctant to challenge a CEO, especially a powerful and successful one. They often feel obligated to the incumbent for recruiting them to the board and they often have strong social ties to the CEO. To research behaviours of boards, one needs to conduct primary research with directors. The examination of board process may hence close an important gap in governance research (Finkelstein and Mooney, 2003; Forbes and Milliken, 1999).In addition to the limited investigation on board process, empirical evidence has also pointed to the need to study board process to better understand performance. A limitation of extant governance research is its near universal focus on a direct relationship between corporate governance mechanisms and firm performance (Daily et al., 2003). Empirical evidence on this direct relationship has not been conclusive (Becht et al., 2002; Daily et al., 2003; Johnson et al., 1996; Shleifer and Vishny, 1997). Researchers have thus questioned the assumptions underlying this linkage (Lawrence, 1997; Melone, 1994). On the other hand, the results of the limited studies which have focused on process variables are stronger (Amason and Sapienza, 1997; Ancona and Caldwell, 1988; Smith et al., 1994). In short, researchers are paying more attention to board process on board role performance (Kesner and Johnson, 1990; Monks and Minow, 2004). Incorporating board process provides a more comprehensive picture to understanding the relationship between structure and performance.Third, the three conceptual models regarding the relationship among board structure, process and performance give an alternative avenue for researchers to explain inconsistencies in past board research. This study also complements the growing interest in opening up the \"black-box\" of decision making that has been manifested in studies involving top management teams (Lawrence, 1997).Understanding the nature of effective board role performance is among the most important areas in management research. When boards of directors are seen as stewards of organisational resources that impact, for better or for worse, the whole of society, the importance of understanding and improving the way they discharge their roles becomes apparent. Hence, by treating boards as decision-making groups and drawing in existing knowledge of board and group dynamics, researchers can focus directly on what boards need to do in order to perform their roles more effectively.\n",
"role": "user"
},
{
"content": "Purpose: - Past literature on board research has centred on board structure and firm performance. Since empirical studies do not reveal a conclusive relationship between the two, attention has shifted towards board role performance. This paper aims to investigate this issue.\nDesign/methodology/approach: - The paper examines three conceptual models for the studying of board role performance: structure, process, and mediation.\nFindings: - Current literature provides little consensus as to the specific configuration for effective board role performance.\nOriginality/value: - First, the study examines various aspects of board role performance. This is in contrast with previous research which largely investigates board role performance in general or under the dominant agency perspective which emphasises the board's monitoring role. Second, unlike traditional governance models, the role of board process is explicitly advocated here. Third, the three conceptual models regarding the relationship among board structure, process and performance give an alternative avenue for researchers to explain inconsistencies in past board research. They complement the growing interest in opening up the \"black-box\" of decision making that has been manifested in studies involving top management teams.\n",
"role": "assistant"
}
] |
A cyclical evaluation model of information security maturity
|
[
"Security",
"Risk",
"Maturity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The critical and methodical evaluation of information security (IS)-related controls becomes necessary because technologies, business processes and people change constantly, altering the current level of risk and creating new risks to the organization (Jirasek, 2012).\nRelated work: The work of Karokola et al. (2011) describes a proposal of information security maturity model (ISMM) for secure e-government services (implementation and service delivery). Basically, the model is based on the findings from the critical analysis of ISMMs of the literature. Five maturity levels with their respective security control dimensions were defined: level 1 (undefined), level 2 (defined), level 3 (managed), level 4 (controlled) and level 5 (optimized). Maturity level 1 is the lowest and maturity level 5 is the highest. This paper is a theoretical proposal and does not show a case study of application of the proposed model. The paper is limited to defining five levels to evaluate the maturity of an ISMS specifically for secure e-government services, without presenting a method for the measurement of the current situation of security or the monitoring and evolution of security and related processes.\nMain technical standards related to IS: The main normative references are the standards in the \"27000 family\" from the International Organization for Standardization (ISO), which are specific for the management of IS (Cowan, 2011).\nSecurity maturity models: A security maturity model provides a guide for a full security program. It also defines the order in which security elements must be implemented, encourages the use of standards of best practices and provides a means to compare security programs (Chapin and Akridge, 2005; The Open Group, 2011).\nThe model of evaluation by maturity levels: The model presented in this paper aims to evaluate IS in a way consistent with organizational goals. The main characteristics of the model are:\nCase study of the evaluation of IS maturity: A case study was performed for the application of the model of evaluation of IS maturity level. The organization that participated in the study is real, but will be named Company X.\nConclusion: The detailing of a method for the management of IS through periodical evaluation of maturity and continuous improvement of controls was shown. The use of the maturity scale allied to the cyclical process of evaluation provided the generation of instantaneous and temporal indicators for the management of IS.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims at presenting a cyclical evaluation model of information security (IS) maturity. The lack of a security evaluation method might expose organizations to several risky situations.\nDesign/methodology/approach: This model was developed through the definition of a set of steps to be followed to obtain periodical evaluation of maturity and continuous improvement of controls.\nFindings: This model, based on controls present in ISO/IEC 27002, provides a means to measure the current situation of IS management through the use of a maturity model and provides a subsidy to take appropriate and feasible improvement actions, based on risks. A case study is performed, and the results indicate that the method is efficient for evaluating the current state of IS, to support IS management, risks identification and business and internal control processes.\nResearch limitations/implications: It is possible that modifications to the process may be needed where there is less understanding of security requirements, such as in a less mature organization.\nOriginality/value: This paper presents a generic model applicable to all kinds of organizations. The main contribution of this paper is the use of a maturity scale allied to the cyclical process of evaluation, providing the generation of immediate indicators for the management of IS.\n",
"role": "assistant"
}
] |
Investigating online consumers' responses to product presentation modes: Does product type really matter?
|
[
"Purchase intention",
"Product type",
"Product knowledge",
"Electronic commerce (e-commerce)",
"Product presentation mode",
"Virtual reality (VR)"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Product information has been extensively acknowledged as a substantial factor in terms of determining consumers' purchase decisions (Brown and Dant, 2009). Illustrating the attributes and functions of a product can help increase consumer comprehension. Of the various research studies concerning product information, one major research stream concentrates on the presentation mode to communicate product information (Li et al., 2016). Product information may be delivered via a single source of sensory stimulus, such as verbal, visual, aural, scent, taste or touch, or some combination of these. In this regard, presentation modes range from text to multimedia to direct product trials. Prior studies in an offline context have found that different presentation modes produce distinct impacts on consumers' beliefs and product evaluations (Kempf, 1999; Li et al., 2012). With internet content being rich and easily accessible and becoming a crucial information source for consumers (Sumita and Zuo, 2010; Tseng et al., 2017), how consumers respond to various online presentation modes is in need of further investigation (Jiang and Benbasat, 2007a; Park et al., 2005).\n2. Theoretical background: 2.1 Virtual reality\n3. Research model and hypotheses: 3.1 The effects of VR mode\n4. Methods: 4.1 Experimental design\n5. Results: A 3 (presentation mode: situational VR/pure VR/picture) x 3 (product type: geometric/material/mechanical) ANOVA was conducted. All cell sizes were equal to 25. The average scores and standard deviations of product knowledge and purchase intention in each shopping situation are shown in Table VII. The ANOVA results shown in Table VIII reveal that the main effect of presentation mode was significant on both product knowledge (F=11.53, p<0.001) and purchase intention (F=8.71, p<0.001). The various types of presentation mode influenced consumers' responses differentially. Post hoc mean comparisons among the modes were conducted using Tukey's HSD test, as displayed in Tables IX and X. Both the pure VR mode and the situational VR mode contributed to a higher level of perceived product knowledge than did the picture mode (pure VR vs picture: t=0.33, p<0.01; situational VR vs picture: t=0.52, p<0.001). The same mode effects were also observed in terms of influencing consumers' purchase intention (pure VR vs picture: t=0.40, p<0.05; situational VR vs picture: t=0.56, p<0.001). Therefore, H1a and H1b were supported. From Table VII, although the situational VR mode had a higher product knowledge and purchase intention than did the pure VR mode in terms of mean values, the mean differences were not significant (t=0.19, p>0.05; t=0.16, p>0.05). Thus, H2a and H2b were not supported. Furthermore, this indicated that both the situational mode and the pure VR mode facilitated a higher degree of product knowledge and purchase intention than did the picture mode. This suggested that the non-immersive VR system induced a certain degree of immersion into the virtual environment among respondents.\n6. Discussion: With the proliferation of VR applications available within the internet marketing context, this study explored the effect of product presentation mode (i.e. picture, pure VR and situational VR) on consumers' responses (i.e. product knowledge and purchase intention). Specifically, this study examined whether VR with situational factors (i.e. situational VR) facilitate better consumer responses than VR lacking any background decoration (i.e. pure VR) and the static mode of online product presentations (i.e. picture). We also performed a moderation test to understand the congruity of the effect of presentation mode on consumers' responses across different product types (i.e. geometric, material and mechanical). The findings of this study provide several important theoretical and practical implications for online marketing.\n7. Conclusions: Based on the prolific use of VR within the internet marketing context, this study explored the main and interaction effects of presentation modes and product types on consumers' product knowledge and purchase intention through an experimental design. The results of this study indicated that both situational VR and pure VR modes were found to have a stronger effect on consumers' perception of product knowledge and purchase intention than the picture mode across the three product types (i.e. geometric, material and mechanical). More specifically, the pattern of VR modes superiority was found to be consistent across all three product types.\n",
"role": "user"
},
{
"content": "Purpose: With the proliferation of virtual reality (VR) applications in electronic commerce, investigations on the effects of VR on consumer responses are important. The purpose of this paper is to examine the effects of online presentation modes (i.e. situational VR, pure VR and picture) on consumer responses for three product types (i.e. geometric, material and mechanical).\nDesign/methodology/approach: This study conducted a 3x3 between-subjects experiment to validate the research model and hypotheses.\nFindings: The results revealed that both the situational VR mode and the pure VR mode had a greater impact on product knowledge and purchase intention than the picture mode. The situational VR mode yielded a higher level of product knowledge and purchase intention than the pure VR mode although it was not statistically significant. Furthermore, the pattern of VR modes superiority was found to be consistent across geometric, material and mechanical product types.\nOriginality/value: This research study contributes to the VR literature by investigating a new type of VR: situational VR, and offering a more comprehensive picture of consumer responses to online product presentations. The authors then drew the implications from the findings to suggest guidelines for practitioners to efficiently allocate resources and maximize the effectiveness of online presentation modes.\n",
"role": "assistant"
}
] |
Devil wears (counterfeit) Prada: a study of antecedents and outcomes of attitudes towards counterfeits of luxury brands
|
[
"Counterfeiting",
"Consumers",
"Premier brands",
"China"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\nBackground: Counterfeiting in China\nRelevant literature and hypotheses development: Counterfeits defined\nMethodology: Data collection\nFindings and discussion: Samples\nDiscussion and implications: Normative susceptibility is shown to influence consumer perceptions of counterfeits and purchase intentions. Consumers may possess the desire to own luxury brands to impress others. However, due to hefty prices of originals, consumers often resort to counterfeits. Luxury brand companies can offer greater affordability to consumers through brand extensions and special licensed products (Wee et al., 1995). If prices could be lower and made more affordable to Chinese consumers, there might be less inclination for them to purchase to counterfeits and instead purchase the originals. However, brand owners need to be cautioned that this may carry the risk of eroding prestige and exclusivity for consumers who pursue the brand value that premium luxury products exude (Wee et al., 1995; Chaudhuri and Majumdar, 2006). If properly executed, there are successful cases of brand extensions such as Armani-Exchange (parent brand House of Armani) and Miu Miu (parent brand Prada) that have further accentuated the parent brand. It is ultimately disturbing that consumers are unable to discern between the originals and the counterfeits. Therefore, it is important for luxury firms and advertisers to ensure that there is a differentiation between originals and counterfeits in terms of quality, functionality and reliability.In terms of information susceptibility, it is also shown that consumers who rely on the expert opinion of others to purchase products would be less inclined to purchase counterfeits of luxury brands. As such, the policy makers of relevant governmental bodies should educate consumers about the negative impacts. For instance, consumers should be aware of the quality and safety standards of counterfeit products. Certain forms of endorsement such as celebrities and government officials would project credibility.It is interesting to note that in contrast to previous studies (Ang et al., 2001; Wang et al., 2005), collectivism did not play a role in influencing consumer attitudes and purchase intention towards counterfeits of luxury brands. This could be attributed to the fact that Shanghai is increasingly becoming more cosmopolitan with strong Western influences. This may have encouraged a certain degree of individualism (Li and Su, 2007). In such a case, advertising initiatives should focus more on the individual self rather than the collectivistic self. However, this result has to be taken with some caution, as it does not represent the whole of China.The findings have also shown that value conscious consumers are more likely to purchase counterfeits of luxury brands. One way to combat this behavior is to use repetition of advertising messages such as \"the best clone would not be close to an original\" to reinforce the quality and value of the original brand (Wee et al., 1995). These are offerings that counterfeit products are unable to imitate. For certain luxury product categories, lifetime warranties could ascertain the utmost best quality that the brand offers, adding value to the consumer's purchase. The original luxury brand manufacturers can widen the gap between risks of purchasing a low quality, fault ridden counterfeit with a sound and valuable original (Cordell et al., 1996). Loyal or current consumers can be encouraged to purchase original brands through promotional campaigns that aim at reinforcing positive attitudes towards the brand (Cheung and Prendergast, 2006).Integrity is found to be an extremely important factor in influencing social consequences of consumer attitudes towards counterfeits of luxury brands. This reflects similar results from previous research (such as Ang et al., 2001; Wang et al., 2005). As such, from the social policy point of view the emphasis on educating consumers on ethics and morals cannot be taken lightly. Educational programs should not only be limited to schools, but also to employees of multinational companies, tourism-related businesses, and other domestic businesses (Simone, 2006). It needs to start at the grass root level, especially during the age when learning is still developing at a faster rate. From a managerial perspective, as suggested by other studies (Wee et al., 1995; Tom et al., 1998; Ang et al., 2001; Wang et al., 2005), by putting a more \"human face\" on the influences would gain more empathy from consumers. This can be introduced through counterfeiting advertisements showing negative economic impacts and potential unemployment as long-term repercussions of counterfeiting activities.It is important to note that the counterfeiting problem does not only lie with the manufacturers but also with the purchasers of counterfeits (Bloch et al., 1993). Consumers' integrity towards counterfeits of luxury brands may be diminished by the perception that luxury brand manufacturers are profiting excessively from the exorbitant prices of such goods (Penz and Stottinger, 2005). As such, consumers justify their purchase by a self-fulfilling prophecy that they are doing no wrong. Companies of luxury brands should consider building an image of strong ethical values and social responsibility to win consumers over.Status consumption seems to consistently influence both \"perceptions of counterfeits\" and \"social consequences\" from the measure of attitudes towards counterfeits of luxury brands. But it is interesting to note that status seeking consumers are more likely to have positive attitudes towards counterfeits of luxury brands. The only explanation could be that many Chinese consumers are becoming wealthier and the conscious need to display such wealth is higher. However, the gap between the rich and the poor is widening as a result of strong economic growth in China and the region (Chadha, 2007). For some, the desire for luxury brands is not achieved due to the lack of affordability to exorbitant price tags of original luxury brands. This contributes to a dissonance whereby consumers resort to counterfeits to satisfy their status seeking needs.Similarly, the results reflect that status consumption influences purchase intention of counterfeits of luxury brands. Luxury brand owners should emphasize the prestige and status symbol that the brand connotes. One of the important aspects to dissuade counterfeiting would be to emphasize on the concepts of face or \"mianzi\" that the Chinese consumers value (Zhou and Belk, 2004). The consequences of being caught consuming counterfeits would be embarrassment and humiliation. This would probably be a strong deterrence against purchasing counterfeits (Wee et al., 1995; Zhou and Belk, 2004; Cheung and Prendergast, 2006).One of the issues much debated in the past is the difference in perceptions in quality, reliability and functionality between counterfeit and original brands (Eisend and Schuchert-Guler, 2006). Admittedly, counterfeits of luxury brands in China are of unexpectedly high quality, whereby there are even \"grading systems\" attached to them to segregate the better quality counterfeits from the less superior ones (Gentry et al., 2006). Therefore, it is a bigger incentive for consumers to purchase a counterfeit since prices are lower and yet attributes of the original version are compatible (Wee et al., 1995). As such, the major cue for companies is to take into consideration processes such as, designing or investing in stronger differentiating product attributes and other innovations that are harder to imitate and costly to counterfeit. This would better enforce the belief that companies do invest to improve products for consumer benefits, instead of producing relatively uninspiring products, but charging exorbitant prices. In most cases, consumers might be deceived into believing they are purchasing original articles when they are really counterfeits. Educating consumers on differentiating original items from counterfeits by providing a list of authorized retailers is very important (Gentry et al., 2006). Even though many manufacturers are afraid that this might bring more negative attention that benefit counterfeiters (Wee et al., 1995), it would be an inevitable move for a long-term success.Other social policy initiatives could also be implemented. It is regrettable that buyers and non-buyers share the same sentiments towards the social and ethical issues pertaining to counterfeits, such as the infringement of intellectual property rights. Although instilling fear towards penalty and criminal punishments attached to counterfeiting is useful, other dimensions to change consumer behavior may also be explored. The Chinese government should enforce a policy whereby buyers and sellers should both be prosecuted for buying or selling counterfeits. Such strategies should be in place to discourage the \"demand\" and \"supply\" side of counterfeiting as both parties should be held responsible (Bloch et al., 1993).\nConcluding comments: In summary, it is evident that consumer attitudes towards counterfeits of luxury brands play an important role in influencing consumer purchase intention. Consumers are more influenced by the perceptions of counterfeits of luxury brands than by the ethical and legal considerations. \"Integrity\" and \"status consumption\" are the most significant factors influencing attitudes and purchase intention of consumers.It is noted that buyers hold more positive views of counterfeits in terms of quality, reliability and functionality of counterfeits than non-buyers, which are consistent with previous research (e.g. Wee et al., 1995; Nia and Zaichkowsky, 2000; Ang et al., 2001; Wang et al., 2005). Results are also consistent with prior findings by Wang et al. (2005), that there is no significant relationship in buyer and non-buyer attitudes when evaluating social consequences of counterfeiting.There are a number of limitations worthy of improvement and future research. The study was conducted using mall intercept method, which may limit the populations that could be reached. Those who may purchase may not be regular shoppers at a shopping mall but may be in wholesale markets where counterfeit products are largely sold. As the study is a snapshot of the Chinese consumers in the coastal areas of the newly rich, extensions to populations of other areas in China of different socioeconomic groups and to other countries may produce different results. The addition of materialism and \"face consumption\" constructs can be further investigated to test for their influences on Chinese consumers.Another possibility is to focus on foreigners who travel or reside in China as buyers of counterfeit luxury brands. The examination of situational factors and mood states, such as tourists on holiday, may have different results as they are likely to experience a sense of excitement softening their ethical stance. Further exploration using qualitative approaches to examine consumer purchase behavior of counterfeit products may provide deeper insights. While this study reflects that the integrity of consumers plays a role in influencing purchase intention, it might differ in other product categories such as pirated CDs and software. Actual ownership can be measured to determine if buyers are also owners of counterfeit products.\n",
"role": "user"
},
{
"content": "Purpose: - This paper sets out to examine how social and personality factors influence Chinese consumers' attitudes towards counterfeits of luxury brands and how these two sets of variables influence purchase intention. It provides a profile of buyers and non-buyers of counterfeits of luxury brands.\nDesign/methodology/approach: - A self-administered questionnaire was designed using established scales. A survey was conducted in downtown Shanghai through the \"mall intercept\" method. A variety of statistical techniques were used to analyze the data.\nFindings: - Status consumption and integrity are strong influencers of purchase intention, whereas normative susceptibility, information susceptibility, personal gratification, value consciousness, and novelty seeking had weaker influencing relationships. The attitude towards counterfeits of luxury brands is found to influence purchase intention. Collectivism does not influence attitudes nor purchase intentions towards counterfeits of luxury brands.\nResearch limitations/implications: - The findings are limited to Chinese consumers in Shanghai, which cannot be generalized across the whole of China or other international markets. Further, only luxury brands are considered. Other cultural contexts and product categories should be investigated in the future.\nPractical implications: - The research provides an in-depth understanding of Chinese consumers' attitudes towards counterfeits of luxury brands. The research findings can be used to formulate strategies for academia, practitioners and, more importantly, policy makers to help eradicate, or at the very least curb, counterfeiting activities.\nOriginality/value: - The majority of previous studies focused on counterfeiting and piracy of music and other optical media, whereas this paper focused exclusively on luxury brands. Status consumption is also added as an antecedent towards attitudes and purchase intention of counterfeits.\n",
"role": "assistant"
}
] |
Asymmetric price transmission in the Chinese pork and pig market
|
[
"China",
"Asymmetric price transmission",
"Pork market",
"Price and stockpiling policy",
"Price substitution"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Few food industries in China command as much attention as pork (Wu et al., 2011). Pork accounts for three-fifths of meat consumption in China, pig production supports the livelihoods of more than 52 million Chinese rural households, while a plethora of policies seek to influence the pork market. Thus recent movements in pork prices along with the relationships between pork and pig prices are of great interest to Chinese consumers, producers and the government. The intent of this paper is to explore the transmission process between pork and pig prices, and to see whether the transmission is impacted by the level and volatility of prices, by government policies or by other factors.\n2. Literature review: Numerous studies have been conducted into agricultural product prices focusing on three main areas: price fluctuation patterns; factors influencing price fluctuations; and price transmission mechanisms. Studies into price transmission have examined co-movement and completeness, the dynamics and speed of adjustment, and asymmetric price transmission (FAO, 2003). In a meta-analysis of price transmission studies covering 282 commodities, Peltzman (2000) concluded that asymmetric price transmission mechanism exists in food markets but that the transmission differs by commodities, countries and market movements.\n3. Methods and data: As mentioned, a three-step modeling approach is used in the study to: determine the asymmetry of the price relationship between pig and pork prices (Section 3.1); identify characteristics of the asymmetric transmission (Section 3.2); and analyze factors influencing the asymmetric transmission (Section 3.3). The data used for these different models are described in Section 3.4. As price transmission is examined under different periods of price volatility, Section 3.5 outlines the process by which the two periods were identified.\n4. Results and discussion: Overall, the pork and pig price series in Figure 1 appear to move together suggesting some level of integration between the two commodities. Table I reports the results of the unit root tests in the two periods. The Augmented Dickey-Fuller (ADF) test reveals that the pig and pork price series are non-stationary but that the first difference of these series are stationary I(1) enabling a cointegration test to be conducted. A cointegration test for the price time series in the two periods (June 1994 to June 2007 and July 2007 to June 2016) was conducted with the TAR model and C-TAR model as well as the MTAR model and C-MTAR model with the results reported in Table II.\n5. Reasons behind asymmetric price transmission in the July 2006 to June 2016 period: Two main explanations emerge from the literature for asymmetric price transmission; market power and menu cost. The market power explanation focuses on imperfect competition in intermediate (wholesaler and retail) links that lead to abuse of market power and asymmetric price transmission (Pan and Li, 2015). The menu cost explanation postulates that asymmetric price transmission originates from costs (to intermediate links) of adjusting prices that can include updating computer systems, re-tagging items and setting new pricing strategies (Levy et al., 1997). Other explanations for asymmetric price include properties of commodities, ratchet effects, government policy and labor costs (Wang et al., 2010).\n6. Concluding remarks: The empirical analysis of pork prices and pig prices in China from June 1994 to June 2007 and July 2007 to June 2016 revealed a long-term equilibrium relationship between pig and pork prices in the two periods. However, there is an asymmetric price transmission effect between pork and pig prices in the July 2007 to June 2016 period.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze price transmission in the Chinese pork market between 1994 and 2016 and examine any incidence and causes of asymmetric price transmission.\nDesign/methodology/approach: The approach uses threshold autoregressive models, asymmetric error correction models and autoregressive moving average models to examine the price transmission using monthly pig and pork prices from 1994 to 2016.\nFindings: While a symmetric price transmission between pork and pig prices was identified for the period between June 1994 and June 2007, an asymmetric price transmission response between pork and pig prices was found for the period July 2007 to June 2016. Key factors behind the asymmetric price transmission include the chicken price and China's provisional purchasing and stockpiling policy which is having a counter-productive impact on prices.\nOriginality/value: The paper contributes to the literature by examining price transmission in two different periods: 1994 to 2007 where prices are lower and more stable; and 2007 to 2016 where prices are higher and volatile. The paper examines the impact of production and market policies on price transmission in the Chinese pork and pig market, with several policy implications.\n",
"role": "assistant"
}
] |
School leadership, teacher's psychological empowerment and work-related outcomes
|
[
"Job satisfaction",
"School leadership",
"Trust in leader",
"Teachers’ empowerment",
"Workplace attachment"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Teacher's job is one of the most stressful, being at the intersection of the sometimes-conflicting expectations of students, parents, school leaders and politicians. The shortage and turnover of qualified teachers constitutes a problem in many countries (World Bank, 2018). School as a workplace is not sufficiently attractive even in the countries with good academic performance, including Estonia.\n2. Literature review and hypothesis: 2.1 Leadership style and trust in leader\n3. Methodology: 3.1 Context of the research, sample and data collection\n4. Results: Based on the correlation analysis (Table I), we can state that there was a significant positive correlation between teachers' perception of principal's empowering leadership style and teachers' psychological empowerment subscales autonomy (r=0.40; p<0.01), impact (r=0.33; p<0.01), meaning (r=0.30; p<0.01) and competence (r=0.17; p<0.01).\n5. Discussion and summary: The mediating role of psychological empowerment revealed by our study is consistent with the results obtained by Lee and Nie (2014), where teachers' psychological empowerment appeared as a mediating variable between school leadership characteristics and job satisfaction and organizational commitment. The mediating role of psychological empowerment has also been investigated outside the school environment and it has been found that psychological empowerment plays a mediating role between the leader's leadership style and job satisfaction (Aydogmus et al., 2018).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the mediating effect of teacher's psychological empowerment between school leadership style and teachers' work-related outcomes.\nDesign/methodology/approach: A total of 711 teachers from 31 Estonian schools were surveyed with a questionnaire measuring four dimensions of psychological empowerment (competence, meaning, self-determination and impact), school leadership characteristics (leadership style, leader's empowering behavior and trust in leader) and teacher's work-related outcomes (job satisfaction and workplace attachment). AMOS path analysis was used to investigate the direct and indirect relations between the teachers' perceptions of school leadership, their psychological empowerment and their workplace attachment and job satisfaction.\nFindings: This study found that psychological empowerment (subscales meaning and impact) mediates the relationship between perceived leadership empowerment behavior and teachers' work-related outcomes. Also, the psychological empowerment (meaning and impact) mediates the relationship between perceived leadership style and teachers' work-related outcomes. Trust in the principal has direct and indirect effect (through psychological empowerment) on job satisfaction, whereas there only seems to be indirect effect on workplace attachment through two components of psychological empowerment.\nPractical implications: The mediating role of psychological empowerment includes an important message for school principals - in order to empower employees it is not sufficient to merely delegate formal power and decision-making rights. To facilitate the development of psychological empowerment, it is important to provide employees with an opportunity to experience agency, to experience that their voice and opinions are taken into account (perceived impact) and the purpose and targets of the whole organization are discussed with the employees and formulated in collaboration with them (perceived meaning).\nOriginality/value: Psychological empowerment as a mediating variable has not been widely researched, especially in school environment. The results will provide important signals for school principals, where and how to find leverage to improve teachers' job satisfaction and workplace attachment.\n",
"role": "assistant"
}
] |
To couple or not to couple: A case study of institutional legitimacy relating to SaaS applications in two universities
|
[
"Universities",
"Technology adoption",
"IT service",
"Institutional theory",
"Implementation",
"Services",
"Cloud computing",
"SaaS",
"Decoupling",
"IS professionals",
"Coupling"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This is a paper about the appeal of popular software as a service (SaaS applications within two institutions of higher education (HE) in Australia and Sweden, and how such applications were employed in harnessing legitimacy as centres of innovative digital practices. SaaS applications offered in the cloud, such as commercial Google products and Dropbox, to list a few, have become widespread amongst consumers with the advent of social media platforms, and the immense popularity of ubiquitous mobile devices, such as smartphones. These digital tools have become commonplace in both personal and institutional use. Yet, the origins of SaaS applications can be traced to the rise of the concept of cloud computing, which has been a buzzword in the IT industry (MacMillan, 2009; Voas and Zhang, 2009) and global economic forums (World Economic Forum, 2010) for more than a decade, and the subject of extensive research by computer scientists (Erdogmus, 2009; Jensen et al., 2009; Pearson, 2009; Armbrust et al., 2010; Zhang et al., 2010). It has also garnered an emerging attention of scholars in the information systems (IS) area (Dwivedi and Mustafee, 2010; Marston et al., 2011; Wang et al., 2011; Sultan and Van de Bunt-Kokhuis, 2012; Stein et al., 2016; Wang et al., 2019), management (Geng et al., 2009; Babcock, 2010; Ross, 2011) and corporate law (Jaeger et al., 2008; Ryan and Loeffler, 2010). Several of these previous works looked at cloud computing from an economic-rationalist perspective. However, Kshetri (2011) applied institutional theory to cloud computing, suggesting that governments can influence how organisations legitimise themselves, such as supporting cloud services that allow data to be easily monitored from a centralised location. There is also earlier work by Melin et al. (2014) applying institutional theory on cloud-based services. A critical point made by Maclean and Behnam (2010) is the fact that internal legitimacy, pertaining internal stakeholders' perceptions of their organisation's compliance with legitimacy-enhancing initiatives, needs to be devoted as much consideration as its external legitimacy. Legitimacy is of importance to both novel and established technologies as it has implications for the mobilisation of resources, the support of regulatory bodies and the growth in influence of technology advocates (Bergek et al., 2008; Markard et al., 2016). Tilcsik (2010) discussed the phenomena of coupling and decoupling in Russian state enterprises, as part of the discourse on organisational legitimacy. It is to be noted that these works reflected the study of cloud computing as a concept related to technology deployment. On the other hand, consumer SaaS applications, though they are an integral aspect of cloud computing infrastructures, have not been approached in terms of how they are adopted and deployed by institutions. Martins et al. (2019) also applied an institutional perspective, but as elaborated on below, in combination with other perspectives. The concluded e.g. that top management support and normative pressures influenced SaaS use. They highlight that cost saving and security concerns were direct predictors of opportunities and risks, respectively. Based on the dominance of economic-rationalist perspective, and the room for an institutional theory-based view on cloud-based services, such as SaaS, we perceive this lack of research as a point for departure.\nAn overview of the institutional theory perspective: One of the central tenets of institutional theory stipulates the need to understand how and why organisations assign meanings to certain aspects of their institutional environments, but not to others. In other words, the theory can be applied to an investigation of what motivates organisations to adopt particular technologies, which has implications for understanding how institutional meaning systems are understood and interpreted by internal stakeholders (Suddaby, 2010).\nResearch approach: The initial step in our research approach is to treat the literature review as a continuous iterative process in understanding the area of decision making concerning cloud service deployment. This is consistent with Boell and Cecez-Kecmanovic's (2014) hermeneutic framework. We started with an inner hermeneutic cycle of search and acquisition. Here, we used a set of initial concepts that we understood as relevant such as cloud computing and outsourcing concepts, economic-rationalist theories such as resource based view and transaction cost economics. Going through a cycle of searching, acquiring and reading, this was then refined as we lengthened our search keywords to include non-economic justifications for deploying what could be deemed a \"fashionable\" technology (Wang, 2010). After this, by going through a wider hermeneutic cycle of analysis and interpretation, we refined our research problem by making a critical assessment to depart from looking at various economic theories to focus mainly on institutional theory concerning myths and legitimacy. After further hermeneutic cycles of search and acquisition, and of analysis and interpretation, we refined our literature even further to specifically target the concepts of coupling and decoupling in our research problem.\nCase background: The dual case studies comprise of an Australian university, UniOz and its Swedish counterpart, UniSwed. The following sections provide a brief overview of the two universities.\nAnalysis and discussion of findings: In this section, we present a cross-case analysis and discussion of findings from our empirical work in UniOz and UniSwed, derived through an interpretative approach to the investigation of the data from the formal and informal interviews and participant observations. In accordance with this, two main themes emerged, as a part of our reflexive research approach, structuring the analysis below, namely, the universities striving for: institutional legitimacy through acts of coupling; and the revelations of consequential decoupling (e.g. Meyer, 1978; Meyer et al., 2007; Tilcsik, 2010). From the analyses of the two cases via these themes, a pattern emerged that discusses how the institutions are continuity engaged in coupling and decoupling internal processes in light of the legitimacy mandate and changes in leadership, demography of staff and shifts in influences of stakeholder groups within the universities. Both themes are explained from the stakeholders in the organisational fields of the two institutions, which include IT managers, students and both academic and administrative staff. This section is summarised in Tables I and II.\nConclusions and implications: From the outset of this paper, the aim of this research is to apply an institutional theory perspective, focusing legitimacy through coupling and decoupling, to understand the meanings and interpretations behind the decision-making behaviour concerning the deployment of new technologies, namely, SaaS applications, in the two case organisations illustrated in the paper. The two universities, one in Australia (UniOz) and the other in Sweden (UniSwed) have shown how different stakeholders' viewpoints can influence the decision-making process concerning this type of deployment longitudinally. Therefore, the intent was to analyse and understand how institutional myths can come into play in such processes. Our study provides a rich and original illustration involving different stakeholder groups and services over time, using institutional theory. We illustrate a play, different acts and patterns of coupling and decoupling can reveal to us the peculiarities of drivers and motivations that lead to upholding or strengthening the legitimacy of decisions made regarding different cloud-based solutions. This analysis is a complement to previous research of cloud computing, applying economic-rationalist justification to the adoption of cloud computing in organisations. The present study, as such, is also an example of how an institutional theory framework, focusing on legitimacy, and coupling and decoupling can be applied in order to understand cloud-based, particularly popular SaaS solutions, in organisations (and in our case universities).\nLimitations and future research: The authors recognise that the case studies of two specific universities limit the ability to transfer these findings to a wider context; at least from a statistical point of view. However, the rich qualitative study of an Australian and a Swedish university reveals the intricacies of how legitimacy-enhancing strategies are played out in light of actors closely tied to revenues and cost control, and point to the need to examine the broader societal and economic forces shaping higher educational institutions in the two countries. This necessitates the continuation of this research by expanding the pool of universities to include additional cases of Australian and Swedish universities, also including continued advancements in relevant digital technologies. This will contribute towards the illustration of a more comprehensive analysis of the morphing relationships between the universities and relevant actors through images associated with the adoption and deployment of new technologies. Studies of emerging SaaS applications can also be updated including more recent advancements, bridging the time lag in the present study. On the other hand, we expect to find similar patterns of coupling and decoupling with more recent applications launched by major technology providers, but this is a question for future studies. Furthermore, such a pursuit will allow us to uncover deep underlying issues governing the business models of universities in the two nations. Another avenue for future research could be to highlight the perspectives of e-learning, focusing the learning aspects of students e.g. using cloud-based services offered by the universities, and to highlight the different governance structures that are present in the institutional contexts of the two universities addressing digital government aspects.\n",
"role": "user"
},
{
"content": "Purpose: The predominant narrative is that contemporary organisations, motivated by economic-rationalist aspirations, adopt cloud applications on the premise of achieving cost-savings and efficiency gains. However, how they actually adopt and rollover such new or emerging technologies may be influenced by acts, patterns and processes of institutional legitimacy. Therefore, the purpose of this paper is to explore the dynamics surrounding decisions on how specific cloud applications are adopted from the context of institutional theory, with a particular focus on the concepts of coupling and decoupling.\nDesign/methodology/approach: To examine this phenomenon, two qualitative case studies, using a reflexive research approach, of an Australian and a Swedish university have been undertaken, both of which adopted commercial cloud applications for e.g. e-mail, collaboration and storage (as examples of software as a service) at different points in time. One of the universities was known for its early adoption of cloud applications, but had decelerated further deployment of such services, while the other, despite its conservative reputation, has made rapid strides in this regard.\nFindings: The findings of the dual case studies reveal that organisations, contrary to economic-rationalist claims, may or may not decide to adopt particular commercial cloud-based offerings for the support of core operations, on the basis of how they perceive their institutional legitimacy being affected by a complex network of influential actors, both internally and in the external spaces. Therefore, this paper offers an institutional theory-based discourse and rich illustrations on how the role of technology is played out in enhancing relationships between an organisation and such actors in terms of legitimacy focusing acts of coupling and decoupling.\nOriginality/value: In the analysis and findings the authors, in a novel way, illustrate how organisations strive for: institutional legitimacy through acts of coupling, and the revelations of consequential decoupling. The value is based on a rich case description, analysis and application of institutional theory.\n",
"role": "assistant"
}
] |
Indigenous technical knowledge inputs for effective disaster management in the fragile Himalayan ecosystem
|
[
"Disasters",
"India",
"Sociotechnical change",
"Earthquakes",
"Landslides",
"Floods"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nThe disaster-prone terrain of Himalaya: Subduction of the Indian Plate and its consequent collision with the Tibetan Plate caused intense folding, faulting, thrusting and fracturing of the rocks, that makes this terrain environmentally highly fragile and prone to mass wastage. The ongoing northward drift of the Indian Plate makes Himalayas highly prone to earthquakes. With population and urban growth, the vulnerability of the people to seismicity has increased in this region and the earthquakes of recent times (1991 Uttarkashi and 1999 Chamoli) have exhibited the devastating potential of these tremors. A portion of the Himalayas lying between Bihar-Nepal border (epicenter of 1935 earthquake) to the east and Kangra (epicenter of 1905 earthquake) to the west has not witnessed a Great Earthquake (Magnitude > 8 on Richter Scale) for the previous 200 years, that makes it a potential site of a future catastrophic earthquake (Bilham et al., 2001).Consequent to the Himalayan uplift and subsequent stabilization of the southwest monsoon the region experiences heavy rainfall, mostly concentrated to the monsoon season. Localised heavy precipitation (cloud burst) often takes place in the region resulting in landslides, debris flow, and flash floods. These often block the streams resulting in the formation of temporary dams whose breach has caused devastation numerous times in the past. Alaknanda valley in Garhwal and Sutluj valley in Himachal have been particularly vulnerable to these. In the Alaknanda valley the urban centers of Srinagar and Chamoli have been ravaged by such extreme events and the district headquarter of Chamoli had to be shifted from Chamoli to Gopeshwar after the floods of 1971.The agriculture that still remains the mainstay of the economy has always been dependent on atmospheric precipitation in the region and the failure, weakening or delay of the southwest monsoon or winter rains often cause conditions of drought in the region, leading to failure of crops. To add to it, western disturbances often cause wind and hail storms in the region, damaging horticultural crops in particular. Occasional forest fires also cause a lot of environmental degradation, paving way for the onset of other erosional processes. In the recent times, with the introduction of modern means of communication, the frequency of road accidents in the region has been particularly high and these have been shown to be particularly fatal in this region.The foregoing description makes it amply clear that the Himalaya has been prone to a large number of natural disasters, besides road and other accidents that take heavy toll of human lives, economy and infrastructure. Increased anthropogenic interventions in the previous decades have enhanced the pace of erosional processes but the disasters in the region cannot be attributed to anthropogenic activities alone. The engineering excavations (mining, road and dam construction) have no doubt enhanced the threat of landslides but these have not spared areas virtually free of human interventions (Rautela and Thakur, 1999, Rautela and Paul, 2001).The frequency of these extreme events has been quite high in the region and their adverse affects on human interests have no doubt increased in the recent times but these would not have always spared the sparsely populated habitations in the past. Archeological evidences of the region show that the human habitation in the region dates back to prehistoric times. All through their habitation in this disaster-prone terrain the human populations would have suffered at the hands of these disasters. Through experience and accumulated knowledge of generations these people would have attempted ways of protecting themselves and their community from the wrath of nature. This experimentation and evolution laid foundations of a system of resource management that is fine-tuned to the fragile ecological conditions of the region and cushioned the society against natural disasters, besides optimally utilizing the resources.Besides being readily acceptable to the masses these practices are cost effective and therefore these need to be documented, studied and innovated upon so as to make these more effective.\nThe indigenous technical knowledge: Earthquake protection\nDisruption of the tradition and need for innovation: The previous couple of decades have witnessed the weakening of the social fabric of the hill communities in Himalaya and with this many age-old traditional practices of resource management that drew strength from social cohesion are losing ground.The traditional practice of house construction that had all the elements of earthquake resistance has been relegated to the back seat with the onslaught of modern construction practices and technology. This is owed to: restriction upon traditional rights to forest for timber and stone; increasing restrictions upon felling and quarrying in the name of environmental protection; increasing cost of timber due to growing demand and ease of transportation; social status attached to concrete buildings, even though these are less energy efficient and unsuited to the cold climate; and local artisans switching to concrete construction for the lack of patronization.This transformation has resulted in haphazard construction of multistoried concrete houses in the seismically active zone of Himalaya. The adoption of cement-based construction technology in the region has not been together with the adoption of engineering standards related to it; we have just changed the construction material. The previous earthquakes of 1991 (Uttarkashi), and 1999 (Chamoli) have highlighted this problem.Despite innovations and technical advancement, the populace at risk from large earthquakes is neither willing to recognize that their traditional houses had the elements of earthquake resistance nor are they willing to believe that there exist means or ability to combat such destructive power. This is the cause of their reluctance to think in terms of planning, organising and spending part of their income, as individuals or as societies, on protection measures. A massive awareness drive for popularizing the earthquake-resistant technology therefore needs to be undertaken at the grassroots level and the local masons need to be skilled in this technology. The voluntary organizations working in the remote Himalayan region can play a major role in this.Breakdown of the community solidarity and over-dependence upon the state has led to the downfall of a number of resource management practices developed over generations and the tradition of maintaining water escape channels at vulnerable locations to ward off the threat of landslides is no exception. With the passage of time the relevance of these structures has been lost into oblivion. Where ever possible, these structures should be rejuvenated, with the involvement of the community under various rural development schemes of the government. At the same time, the efficacy of constructing similar and even more efficient structures should be studied at locations vulnerable to landsliding.Differential appreciation in the cost of coarse and fine cereals together with the increasing dependence upon market has been a major disincentive for the cultivation of traditional crops and even the genetic strains of a large number of crop varieties have been lost. The price of this loss, that could have boosted the diversity of mountain agriculture, remains to be adjudged. The traditional crops of the Himalayan region therefore require patronization, together with value addition and marketing support.The Himalayan terrain houses a number of famous tourist, adventure sports, and pilgrimage centers that attract large numbers of people from across the country and abroad. The region thus witnesses a large number of drivers that are not conversant with the terrain conditions, principles of hill driving and of course the significance of these small road-side temples. It is for the safety of these people, and also to promote tourism that is one of the main sources of revenue, that road safety needs to be given more attention by the organized sector. Not leaving everything to the government, it is required that the organizations involved in insurance, tourism, hospitality and the like become partners for making hill roads safer.Moreover the state needs to take recourse to:* Awareness drives for highlighting the high accident rates in the hills together with the general driving precautions so that the drivers are cautious.* Licensing procedures have to be made stringent for hill driving and those involved in reckless driving should be immediately penalized.* Strict survillance sould be resorted to for the adherence of the traffic rules.* In the rugged Himalayan terrain even a small negligence on the part of the driver has the potential of causing fatal accidents and this is reflected in a large number of accidents that result in the vehicles falling off the road into deep gorges, cliffs and fast-flowing and deep waters. Even though the police force reaches the site of the accident, it is ill equipped and trained to rescue the victims. The delay in response thus leads to soaring fatality rates in such cases. For reducing the fatality rates of the mountain roads it is a must to train and equip those in charge of search and rescue in the aftermath of any accident. An option of raising trained cadre of volunteers can also be resorted to for this purpose.* Charged with better communication and mobility, the search and rescue personnel should be stationed at crucial sites.* A large number of accidents in the hilly region take place during the night hours, therefore night driving should be totally disallowed.* The accident-prone zones should be marked and special structural measures should be resorted to in these zones to avoid fatal accidents.Many organizations today realize the relevance of the traditional practices and are working at documenting, studying and conserving these practices before these are lost for ever. Efforts from the scientific community are however required for giving due recognition to this knowledge base and it has the responsibility of coming forward with innovations that amalgamate the tradition with the recent breakthroughs in knowledge, so as to be fine tuned and well suited to the ground realities of the region, apart from being readily acceptable to the masses. The state on its part should also promote and patronize these studies for reducing the vulnerability of the remote Himalayan communities rather than putting forth state-of-the-art technological solutions. It needs to be realized that the foundations of the prosperous and safe Himalayan communities can only be laid upon the traditional technical knowledge base of the inhabitants of the region.\n",
"role": "user"
},
{
"content": "Purpose: - Survival strategy of the masses has led to the evolution of area-specific, locally pertinent and effective ways of mitigating natural disasters. This vital knowledge base is, however, often ignored and is being fast eroded. The paper aims at highlighting the relevance of these practices that put forth most cost-effective and acceptable means of mitigating disasters.\nDesign/methodology/approach: - The paper is based on the extensive field studies and in-depth probe of the traditional resource management practices of the people of the remote Himalayan terrain by the author and it reflects his belief in the acumen of the masses.\nFindings: - The paper discusses the relevance of the various disaster management practices of the region. During the course of their habitation in the disaster-prone Himalayan terrain the indigenous people through experience, experimentation and accumulated knowledge devised ways of reducing their vulnerability to natural hazards. Studies show that their understanding was fairly evolved in the areas of earthquake, landslide and drought management and had devised efficient ways of mitigating the effects of these.\nResearch limitations/implications: - The paper reflects the author's individual understanding and interpretation of the practices based on his interactions with the masses.\nPractical implications: - Improvisation on the traditional practices of disaster management has a major role to play in putting forth cost-effective and sustainable means of shielding the community against the impact of natural disasters. This article puts forth an alternative approach of interpreting the traditional practices of the masses and would lead to appropriate innovations for better disaster management in the region.\nOriginality/value: - This paper represents a useful attempt to reinforce the importance of local knowledge in mitigating natural disasters, which is often ignored and is quick to erode.\n",
"role": "assistant"
}
] |
Stemming the tide: dealing with the imbalance of customer relationship quality with the key contact employee versus with the firm
|
[
"Customers",
"Employees",
"Service industries",
"Customer service management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\nIntroduction: In the past decade, marketing scholars have called for firms to build strong relationships with their customers because quality relationships with customers could ensure future retention, referrals, and long-term profitability (e.g. Brown, 2001; Anderson and Robertson, 1995; Morgan and Hunt, 1994). One way of building high quality relationships with customers is to encourage customer-employee bond-building (e.g. Hartline and Jones, 1996; Doney and Cannon, 1997; Ramsey and Sohi, 1997; Price and Arnould, 1999; Reynolds and Beatty, 1999; Swan et al., 1999). Previous studies have found that a customer's bond with a firm's employee will lead to positive word of mouth (WOM) (Gremler et al., 2001), stronger repurchase intention (Reynolds and Arnold, 2000), and higher levels of satisfaction with the organization (Swan et al., 1999; Reynolds and Beatty, 1999). Current wisdom prescribes that companies encourage their employees to build strong personal customer-employee bonds (e.g. Berry and Parasuraman, 1991; Beatty et al., 1996; Price and Arnould, 1999; Gremler et al., 2001). Corporate America has responded by devoting a significant amount of resources towards employee training and recruiting systems that facilitate building strong interpersonal customer-employee relationships. Thus, a quality relationship between the customer and the key contact employee (Stanley, 1985) appears to represent another source of sustained competitive advantage (Hansen et al., 2003).However, previous studies on customer-employee and customer-firm relationships assume that the key contact employee continues working for the firm. Few studies have examined the potential harm caused by the termination of an imbalanced customer-employee relationship, where customer-firm relationships are stronger than customer-employee relationship. We use the imbalance of relationship quality to depict such phenomena (Bendapudi and Leone, 2002). Current thinking in this area suggests that the consequences of terminating customer-employee relationships in this situation are daunting. For example, American Express has projected that on average, 30 percent of a financial advisor's clients would move with their advisor if he or she were to leave the firm (Tax and Brown, 1998). Similar transferring events after a key contact employee's leaving have been reported in several other contexts (e.g. Sherden, 1992; Gutek, 1995; Payne et al., 1995; Beatty et al., 1996). Even when switching costs prohibit customers from defecting to a competitor, the dissolution of an imbalanced customer-employee relationship may cause customers to change their attitudinal loyalty to the service organization (Gamble, 1988), to re-evaluate their relationship with the firm (Anderson and Robertson, 1995; Duboff and Heaton, 1999), or to become receptive to competitive promotional activity.A two-year long study by Bendapudi and Leone (2002) suggested that traditional strategies aimed at preventing the termination of customer-employee relationships or customer defection resulting from the termination (e.g. encouraging employee retention and requiring \"non-compete\" agreements with employees) have been shown to be largely ineffective. Service firms are especially vulnerable to the effects of an imbalanced customer-employee relationship. The purchase of a service is thought to be a process that depends in part upon the interpersonal interaction between the service provider and the customer (cf. Berry, 1983; Crosby et al., 1990; Iacobucci and Ostrom, 1993; Ostrom and Iacobucci, 1998). Therefore, the attributes of the service context are much more likely to create an imbalanced customer-employee relationship, especially when the services are intense, close, and long-term (e.g. psychotherapist-client, patient-doctor, client-attorney).Despite the negative impact and the high managerial relevance associated with the issue of imbalanced customer-employee relationship, very little research has investigated the causes of its origin (Bendapudi and Leone, 2001, 2002). Previous studies on customer-employee (service worker) relationships focus mainly on the bright side of such relationships (cf. Bove and Johnson, 2001). Little is known about the customers' emotional and cognitive reactions when the relationship with his/her favorite employee terminates. However, understanding such reactions could enable service organizations to take precautious actions to prevent switching events and to provide better customer care when facing the termination of an imbalanced customer relationship. We address this gap in the literature by identifying the factors that cause the imbalance of customer relationship quality and by examining customers' cognitive and emotional reactions to such imbalance upon the loss of the key contact employee.The paper is organized such that it first provides a brief overview of the literatures relevant to the study. It then presents three categories of factors leading to the imbalance of customer RQ, and customers' cognitive and emotional reactions to the leave of the key contact employee under an imbalanced relationship situation. The research methods and results are then presented. The paper concludes with managerial implications and suggestions for future research.\nConceptual background and hypotheses development: The service triangle model and imbalance of customer RQ\nMethods: Sample and survey administration\nAnalysis and results: Measurement model analysis\nDiscussion: Although previous studies suggest that the imbalance of customer RQ in favor of the employee may well pose vulnerability to the service organization, especially when the employee is no longer available (Bendapudi and Leone, 2001, 2002), little research has been conducted to understand what causes such imbalance. Few studies have investigated into the potential consequences of such imbalance upon the leave of the key contact employee. Drawing upon the attribution theory and relationship marketing literature, we identified six factors from three categories as important antecedents of the imbalance of customer RQ in favor of the employee and three consequences reflecting customers' cognitive and emotional reactions to the leave of the key contact employee. The predictive powers of all these factors and the proposed consequences were tested by surveying 780 customer-employee pairs. The results confirmed our propositions. These findings, as well as the implications to service organizations, will be elaborated as follows.Three consequences\nLimitations and future research: While useful insights have been obtained through this study, there are a number of limitations that could be addressed with future research. First, our study sampled a large number of service organizations (67) and varied number of customer-employee pairs from each. This design certainly enhanced the generalizability of our findings, but it also posed the question of extraneous factors outside the study such as organizational size effects, industry effects[5], or even the regional effects. Although we believe that the large sample size, both at organizational level and at customer-employee level, will cancel out any of these effects, future studies should be conducted to look at the potential impact of factors outside this study. It would be important for both service researchers and service providers to know how the organization size, nature of service, types of service, and other factors moderate the predictive power of these six antecedents studied in this study.Second, we have focused on six antecedents and three consequences of the imbalance of customer RQ, but these factors by no means are inclusive. For each category, there are other antecedents that future studies should investigate. For example, customer relationship proneness may be another useful variable in understanding this phenomenon (cf. Wulf et al., 2001). Besides gender, other demographic factors from both customer and employee perspectives could lead to more insights. Since our sample had more females (64.4 percent) than males, our findings about customer gender may need to be re-tested in future studies. From an organizational perspective, several other variables such as employee-management styles, compensation structure, and perceived organizational justice may be highly relevant in our understanding of imbalanced customer RQ. More importantly, all three consequences that we studied are from the customer's perspective. The other side of the story could also be interesting: how will the key employee react when he(she) cannot serve his(her) close customers any more? Future research should address this unanswered question.Third, this study only focuses on the imbalance of customer RQ with the individual employee versus with the firm, and such asymmetries may be extended to other pairs of business relationships such as customer-brand and customer-product/service relationships (Martin, 1996). Previous studies suggest that customers may develop a person-like relationship with brands and/or products (e.g. Fournier, 1998). Future research in this area could examine how RQ imbalance which favors the brand affects the customer's ability to form relationships with new brands launched by the company.Finally, using a cross-sectional survey design, our study took a composite view of some variables such as empowerment, extra-role performance, and multi-facet interactions. Future studies may want to use experimental designs to dissect the variables in a more elaborate fashion. Such an approach could give service providers more specific direction on how to manage relationship imbalance.\n",
"role": "user"
},
{
"content": "Purpose: - Current relationship marketing literature has stressed the importance of building customer-employee bonds. These efforts sometimes result in customer-employee relationships that overwhelm or supersede the customer-firm relationships. Previous studies report that such an imbalance weakens the firm's position with the customer. Very little research has investigated what factors contribute to the imbalance and its consequences for the firm's relationship with the customer. This research aims to look at customer relationship quality (RQ) imbalance, specifically the imbalance that favors the employee, and identifies six factors as important antecedents. Also studied are the customers' cognitive and emotional reactions to the termination of the customer-employee relationship when RQ imbalance favors the employee.\nDesign/methodology/approach: - In total, 780 customer-employee pairs from 72 service firms were surveyed using a structured questionnaire.\nFindings: - Analyses of variance (ANOVA) revealed significant group difference along all six antecedents and three consequences.\nResearch limitations/implications: - The findings of this study suggest that imbalanced customer relationships might be prevented, or at least predicted, if the causes of such relationship structure are better understood.\nPractical implications: - Service organizations should be aware of the negative consequences of imbalanced customer relationships and take necessary caution in their company policies in order to eliminate the negative consequences of imbalanced customer relationships.\nOriginality/value: - This study is the first quantitative inquiry into imbalanced customer relationship issues, which are extremely important in the services industry. Thus, it enhances the literature on services management.\n",
"role": "assistant"
}
] |
Leader empathy, ethical leadership, and relations-oriented behaviors as antecedents of leader-member exchange quality
|
[
"Ethics",
"Leadership",
"Behavior"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Leader-member exchange (LMX) theory suggests that a leader will develop an exchange relationship over time with each subordinate (Dienesch and Liden, 1986; Graen and Cashman, 1975; Graen and Scandura, 1987; Graen and Uhl-Bien, 1995). Social exchange theory and role theory provide the basis for explaining how the exchange relationship develops gradually over time as a leader interacts with each subordinate and the role of the subordinate is negotiated. The quality of the relationship may vary from one subordinate to another. In a high exchange relationship there is a high level of trust, liking, and respect. The leader provides outcomes desired by the subordinate (e.g. interesting tasks, additional responsibilities, more rewards), and in exchange the subordinate is expected to be committed to the work and loyal to the leader. In low-quality exchange relationships, subordinates are only expected to perform the formal requirements of their jobs, and extra benefits are not provided by the leader.The quality of a leader's exchange relationships with subordinates has important implications for leadership effectiveness. A leader who is able to develop high quality relationships with most or all subordinates is likely to be more effective than a leader who is unable to develop high quality relationships (Graen and Uhl-Bien, 1995). Empirical studies have found a positive correlation between LMX quality and several indicators of leadership effectiveness (e.g. Graen and Uhl-Bien, 1995; Harris et al., 2009; Schriesheim et al., 1999). In a meta-analytic study Gerstner and Day (1997) found that a favorable exchange relationship was related to higher subordinate satisfaction, greater organizational commitment, better job performance, and lower turnover. Erdogan and Liden (2002) reported additional positive outcomes of a favorable exchange relationship, such as more creativity, less job stress, and better workplace safety.More research has been conducted on the outcomes of LMX than on its determinants, but several antecedents have been identified (Liden et al., 1997; Nahrgang et al., 2009; Schriesheim et al., 1999). A favorable exchange relationship is more likely when the subordinate is perceived to be competent and dependable, and the subordinate's values, attitudes, and demographic attributes are similar to those of the leader. Some personality traits for the leader and subordinate (e.g. agreeableness, extroversion, positive affectivity) may also be related to LMX. However, the number of studies on traits is too small to reach any firm conclusions, and the studies did not include mediating variables such as leader behavior to explain the relationship.The development of LMX may also be affected by contextual variables (Liden et al., 1997). It may be more difficult for the leader to develop favorable exchange relationships when the work unit or team has many members, when the members are only temporarily assigned to the team, when the members are widely dispersed and seldom interact with the leader, when the leader is overloaded with responsibilities and has little time for interaction with individual members, or when the leader has little power to provide rewards and benefits desired by members. The extent to which leaders develop different LMX relationships with their subordinates is probably affected by other aspects of the situation as well, such as the organizational culture, human resource practices, and the type of team or work unit (Henderson et al., 2009).Another antecedent of LMX is leader behavior, but the theory does not clearly explain how this behavior is related to the quality of the exchange relationship with a subordinate. Causality in both directions is possible, because leader behavior can influence how the exchange relationship develops, but leader perception of a subordinate's competence and loyalty influences the choice of behavior (Graen and Scandura, 1987; Keller and Dansereau, 1995). A recent study found that LMX is correlated more with relations-oriented behaviors than with other types of leadership behaviors (Yukl et al., 2009). The relations-oriented behaviors included providing psychological support, recognizing subordinate contributions, developing subordinate skills, consulting with subordinates to learn about their ideas and concerns, and delegating more authority and responsibility to subordinates.Two other leader characteristics that appear relevant for the development of a favorable exchange relationship involve leader skills and values rather than leader behaviors. Empathy is the ability to recognize and understand the emotions and feelings of others, and this interpersonal skill can make it easier to develop a cooperative relationship of mutual trust with a subordinate. Ethical leader values include concern for the welfare of subordinates and willingness to protect, help, develop, and empower them. A leader with these values will have more desire to develop and maintain a cooperative relationship with subordinates.Empathy, ethical values, and relations-oriented behavior are distinct constructs that may jointly influence the development of exchange relationships. No prior study included more than one of these three types of antecedents, and it is not evident whether they have independent or joint effects. An implicit assumption in much of the theory and research on effective leadership is that behaviors mediate the effects of leader traits and skills on outcomes such as subordinate satisfaction, motivation, and performance (Yukl, 2010). Based on the limited theory and research involving leader empathy and ethical values, it is plausible that the effects of empathy and ethical values on LMX are mediated by the leader's relations-oriented behaviors. The purpose of our study was to examine these relationships and test alternative models about them. The full mediation model shown in Figure 1 is compared to a direct effects model with no mediation, and the possibility that a partial mediation model provides a better fit to the data is also considered.\nLiterature review and hypotheses: The number of directly relevant studies on leader empathy, ethical values, and relations-oriented behavior as antecedents of LMX is still very limited. Much of this research has emphasized broadly-defined constructs such as transformational leadership and emotional intelligence, which include other components in addition to the predictor variables emphasized in our study. The results in such studies are difficult to interpret, but they can provide clues about the likely relevance of our predictor variables. The following sections will review literature that provides the rationale for our proposed hypotheses about relations-oriented behavior, empathy, and ethical values as antecedents of LMX.Relations-oriented behavior and LMX\nMethod: Sample\nResults: Descriptive statistics and preliminary analyses\nDiscussion: Several prior studies provided evidence that leader empathy, ethical leadership values, and relations-oriented behaviors are antecedents of LMX, but they did not examine joint effects for more than one of these antecedents. Our study was conducted to examine multiple mediators simultaneously, and our proposed mediation model was based on the common assumption that the effects of leader values and skills are mediated by leader behaviors. The results from two different types of analysis (hierarchical regression and SEM path analysis) were consistent with a causal model in which relations-oriented behavior fully mediates the effects of leader empathy on LMX and partially mediates the effect of ethical leadership.The direct effect of ethical values on LMX was not included in the full mediation model we initially proposed, and it may be explained by mediators that are not components of relations-oriented behavior. For example, a leader with strong ethical values is likely to set an example of ethical behavior for subordinates, set firm standards of ethical conduct, and hold subordinates accountable for unethical behavior. Other types of leader behavior that may mediate the effects of ethical values on LMX include dispensing rewards and benefits fairly, mediating conflicts impartially, and taking personal risks to defend subordinates who are in danger of being exploited or mistreated by others. Future research on mediators of ethical leadership should include these other likely mediators in addition to the relations-oriented behaviors in our study.Limitations\n",
"role": "user"
},
{
"content": "Purpose: - Leader empathy, ethical values, and relations-oriented behavior all appear to be relevant for effective leadership, but nobody has examined how all three variables are jointly related to leader-member exchange quality (LMX). The purpose of this study is to examine these relationships and test a proposed model describing them.\nDesign/methodology/approach: - Data were collected with a questionnaire from subordinates of leaders in several organizations, and SEM path analysis was used to test alternative models.\nFindings: - The results indicated that a leader's relations-oriented behavior fully mediated the relationship between leader empathy on LMX and partially mediated the relationship between ethical leadership and LMX. Recommendations for future research to verify and extend the results were provided.\nOriginality/value: - This is the first empirical study to simultaneously examine the complex relationships among these four different aspects of leadership (i.e. a skill, values, behaviors, and exchange relationship). The authors' measure of ethical leadership minimizes confounding with other constructs, which is a problem in earlier studies. The practical implication is that leader empathy and ethical leadership should be included in leadership selection and development programs.\n",
"role": "assistant"
}
] |
From transitional to radical knowledge-based development
|
[
"Knowledge management",
"Cities"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: With this issue, the Journal of Knowledge Management begins a series of annual special issues devoted to knowledge-based development (KBD). Since a first attempt in 2002, when a special issue on knowledge-based development was published as Vol. 6 No. 4. the topic has been consolidating thanks to the continued support of Rory Chase, the journal's general editor and Rachel Murawa, managing editor. Our major concern was, in that first special issue (SI), to lay down a sound foundation for the bridging of knowledge management (KM) and KBD. The challenge was to introduce KBD to the KM community as both a naturally related and significant topic.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this article is to introduce the 2006 annual special issue on knowledge-based development, focusing on the progress of this emerging field and on the direction it seems to be taking.\nDesign/methodology/approach: - From this viewpoint, a brief outline of the volume contributions is provided.\nFindings: - A number of characteristics of the contributions along this and the two preceding special issues are identified: multidisciplinarity, theoretical and methodological variety, systems approach, empirical bases and particularly a growing strategic perspective.\nOriginality/value: - This volume may contribute to stimulate further research into the full potential of knowledge-based social value systems\n",
"role": "assistant"
}
] |
Rapid video assessment for monitoring testing facility fraud
|
[
"Accreditation",
"Fraud",
"Testing",
"Anomaly detection",
"Motion detection"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Testing is a process that determines specified characteristics of an object. It provides the means for verifying if certain expectations about the quality, safety, reliability, efficiency, and environmental sustainability of products and services are delivered. Some typical testing services relate to water quality, clinical tests, emissions, and vehicle safety. To ensure that a testing facility has the competence and independence needed for carrying out their activities, accreditation is used. The accreditation process is a third-party attestation related to a testing facility conveying formal demonstration of its competence to carry out specific tasks. An accreditation body is usually granted its authority by the government of a country. The accreditation body, using trained and qualified experts, periodically evaluates the accredited organizations to verify their compliance with technical regulations. For these evaluations, assessors could use video records and photographs of the tests performed by the facility to judge their conformity to technical requirements.\n2. Related work: Because of a lack of research about fraud detection using video surveillance in the accreditation context, this section presents a literature review of some methodologies developed for monitoring specific tasks performance.\n3. Background: This section describes the basic concepts underlying the proposed procedure.\n4. Rapid video assessment procedure: This study proposes a rapid processing technique to find unusual patterns in accredited testing facility video records of their services. The main contribution described in this paper is the process of rapidly finding unusual patterns in videos of accredited tests with a minimum of manual effort. This is achieved by identifying video frames that are not statistically redundant.\n5. Illustrative case: To validate the proposed method, a real-world data set with video records from accredited testing facilities in the field of vehicle safety in Brazil was used. The processing time of the proposed procedure was compared with the time needed to process the video records in a traditional fashion, i.e., where an expert analyzes the video records minute-by-minute with all his/her attention focused on the videos to determine if the test followed the full protocol.\n6. Conclusion: This paper presented a technique to identify unusual patterns in accredited testing facilities using video records of their services. This technique enables identification of video records that contain anomalous behavior in accredited testing facilities.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present a procedure for finding unusual patterns in accredited tests using a rapid processing method for analyzing video records. The procedure uses the temporal differencing technique for object tracking and considers only frames not identified as statistically redundant.\nDesign/methodology/approach: An accreditation organization is responsible for accrediting facilities to undertake testing and calibration activities. Periodically, such organizations evaluate accredited testing facilities. These evaluations could use video records and photographs of the tests performed by the facility to judge their conformity to technical requirements. To validate the proposed procedure, a real-world data set with video records from accredited testing facilities in the field of vehicle safety in Brazil was used. The processing time of this proposed procedure was compared with the time needed to process the video records in a traditional fashion.\nFindings: With an appropriate threshold value, the proposed procedure could successfully identify video records of fraudulent services. Processing time was faster than when a traditional method was employed.\nOriginality/value: Manually evaluating video records is time consuming and tedious. This paper proposes a procedure to rapidly find unusual patterns in videos of accredited tests with a minimum of manual effort.\n",
"role": "assistant"
}
] |
Wage formation and gender wage gaps: is there a role for job-task evaluation schemes?
|
[
"Gender wage gap",
"Decomposition",
"Human capital",
"Job-task evaluation",
"Technology industry",
"Wage formation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Both the private and the public sectors in Finland have, over the past decade or so, seen rapidly expanding use of various modes of performance-related pay and job-task evaluation schemes. The technology industry has been a forerunner in adopting and implementing performance-pay systems as well as in developing job-task evaluation schemes. Indeed, the growing use within the technology industry of various modes of performance-related pay has resulted in an increasingly diverging trend between especially white-collar workers' basic (normal) and total wages. Moreover, in 2002 the technology industry introduced an innovative system for evaluating the job tasks of white-collar workers, in a similar fashion across all of the industry's establishments. This evaluation is to be entirely based on the efforts, responsibilities and specific skills required for performing the working tasks related to each job, without formal skills such as education and work experience influencing the assignment of individual job-task levels. Hence, the system is not to be seen as a kind of proxy for standard human capital endowments. Concomitantly, because of the industry's fundamental - mainly globalization-induced - restructuring, including outsourcing and off-shoring, also its workforce has undergone substantial changes, the most conspicuous being a clear strengthening in the dominance of high-educated employees.\n2. Data and descriptive statistics: The data is extracted from the full records of the Confederation of Finnish Industries EK. The confederation gathers, on a regular basis, information on wages and worker attributes directly from its member companies, which guarantees that the information is of high quality. It is mandatory for the companies to respond to the EK survey implying that the non-response bias is practically non-existent. The EK files are supplemented with information on, inter alia, completed educational degrees as recorded in the official registers of Statistics Finland.\n3. Estimation method and some initial results: The decomposition method based on unconditional quantile regression techniques applied in the subsequent analysis is developed by Melly (2005, 2006a, 2006b). It extends the decomposition procedures based on quantile regression proposed by Gosling et al. (2000) and is very similar to that presented by Machado and Mata (2005). As shown by Melly (2006b), his estimator is, in effect, numerically identical to the Machado-Mata estimator when the number of simulations used in the Machado-Mata approach goes to infinity. A major reason for choosing Melly's method instead of the Machado-Mata estimator is that it is more straightforward to implement. For details on the differences between the Melly estimator and previously suggested methods (see Melly, 2006b).\n4. Sources underlying real wage growth and increased wage differentials: decomposition results by gender: The effects of changes in coefficients (price effect) and in workforce characteristics (composition effect) on the observed changes in the distribution of total hourly real wages among male and female white-collar workers employed in the technology industry are first estimated by inclusion of merely standard human capital measures (years of schooling, potential work experience and seniority) as explanatory variables. The effect of the residuals is persistently negligible and is therefore not depicted in the two graphs of Figure 1.\n5. Sources underlying the gender wage gap: comparison of decomposition results for 2002 and 2007: Have these similarities and differentials in real total-wage growth and increased wage dispersions across genders affected the male-female white-collar total-wage gap of the Finnish technology industry? Most importantly, has the implementation of the job-task evaluation scheme introduced in 2002 had a clear-cut impact on the industry's gender total-wage gap already by 2007? The approach used to answer these questions is identical to the one applied in the previous section: the effects of gender-specific differences in coefficients (price effect) and characteristics (composition effect) on the gender total hourly wage gap are first estimated with account being made for standard human capital endowments only, then by supplementing the estimated model with information on job-task evaluation levels and, finally, with a set of occupation dummy indicators.\n6. Concluding remarks: This paper has investigated major factors underlying the observed patterns and trends in male-female wages and wage gaps among white-collar workers employed in the Finnish technology industry. Special attention has been paid to the role played by standard human capital endowments (years of schooling, work experience and seniority) as compared to alternative ways of measuring competencies (the technology industry's job-task evaluation scheme). The methodology applied allows the whole wage distribution to be decomposed into the effects of characteristics (composition effect) and of coefficients (price effect).\n",
"role": "user"
},
{
"content": "Purpose: Both academia and policymakers express a strong belief in higher average education levels exerting a narrowing impact on wage inequality in general and gender wage gaps in particular. The purpose of this paper is to scrutinize whether or not this effect extends to R&D- and export-intensive branches such as the technology industry.\nDesign/methodology/approach: In exploring the impact of individual and job-related background factors and, especially, of job-task evaluation schemes on the size and change in gender wage gaps in the technology industry, the paper applies an elaborated decomposition method based on unconditional quantile regression techniques.\nFindings: While changes in standard human capital endowments can explain little, if anything, of the growth in real wages or the widening of wage dispersion among the Finnish technology industry's white-collar workers, a new job-task evaluation scheme introduced in 2002 seems to have succeeded, at least in part, to make the wage-setting process more transparent by re-allocating especially the technology industry's female white-collar workers in a way that better reflects their skills, efforts and responsibilities.\nPractical implications: One crucial implication of this finding is that improving the standard human capital of women closer to that of men will not suffice to narrow the gender wage gap in the advanced parts of the economy and, hence, not also the overall gender wage gap. The reason is obvious: concomitant with rising average education levels, other skill aspects have received increasing attention in working life. Consequently, a conscious combination of formal and informal competencies as laid down in well-designed job-task evaluation schemes may, in many instances, offer a more powerful path for tackling the gender wage gap.\nOriginality/value: While the existing evidence on the impact of performance-related pay on gender wage gaps is still scarce but growing the authors know of no empirical studies analyzing the gender pay-gap effect of job-task evaluation systems.\n",
"role": "assistant"
}
] |
Threshold concepts in economics
|
[
"Economics",
"Threshold concepts"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: How can business educators assist their students to learn and understand the basics of their academic field? This deceptively simple question is common to all the academic disciplines - including those taught within business and management schools. It is deceptively simple, because to answer this adequately requires academics to not only know their own discipline content and how to communicate it, but also to understand what it means to learn that content and to recognise how it changes the way learners see the world. It means being able to identify the fundamental conceptual break-throughs novice learners need to achieve to genuinely understand a discipline's perspective. It is also a challenging question for university educators, because it demands they understand how their students learn their discipline and can identify when students genuinely grasp new perspectives (as opposed to mimicking jargon and \"key concepts\"). Threshold concepts offer important insights into these and many other aspects of student learning that can assist business educators in their efforts to help students understand.\nTheoretical background: Meyer and Land (2003, 2005, 2006) and Land and Meyer (2010) and others, introduced and then developed the idea of the threshold concept - the fundamental conceptual change/learning that a learner must make, in order to see the world from a particular discipline's perspective. For the learner it represents a \"threshold\" which must be crossed, and when achieved transforms that student's perspective.\nThreshold concepts in economics: Shanahan et al. (2006, 2008), and others including Davies and Mangan (2007, 2008, 2010) and Davies (2011) have argued that the idea of \"opportunity cost\" is a threshold concept in economics. \"Opportunity cost is the evaluation placed on the most highly valued of the rejected alternatives or opportunities\" (Eatwell et al., 1998, p. 719). When this concept is fully and deeply understood, it can transform the way in which individuals view how to evaluate choices. Economics, which regards the existence of scarcity as a basic phenomenon, places the necessity to evaluate and make choices between competing objectives of scarce resources at the core of its discipline. \"Opportunity cost, the value placed on the rejected option by the chooser, is the obstacle to choice; it is that which must be considered, evaluated and ultimately rejected before the preferred option is chosen. Opportunity cost in any particular choice is, of course, influenced by prior choices that have been made, but with respect to this choice itself, opportunity cost is choice-influencing rather than choice-influenced\" (Emphasis in original; Eatwell et al., 1998).\nSome examples of teaching and assessing opportunity cost[2]: Lecturers whose focus is to teach and assess a threshold concept such as opportunity cost must immediately face the challenge that simple questions that encourage, or even accept as passable, simplistic, or \"rote-based\" answers, will not suffice[3]. Questions that attempt to \"get at\" student understanding, therefore, are both deceptively simple, and require considerable judgment to evaluate. While these questions may take more time to mark, they can also serve as the basis for class exercises. Consider the following example, taken from the \"Embedding Threshold Concepts Project\".\nConclusion: The idea of threshold concepts is a potentially powerful tool that can assist lecturers to help their students learn economics. Discipline-specific pedagogies developed around this idea have the potential to deepen student learning and add considerable value to a student's university education. They also require considerable commitment on the part of the discipline lecturer.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine threshold concepts in the context of teaching and learning first-year university economics. It outlines some of the arguments for using threshold concepts and provides examples using opportunity cost as an exemplar in economics.\nDesign/methodology/approach: The paper provides an overview of the theoretical literature around threshold concepts in economics and provides exemplars from previous research, and current practice, to illustrate how they might be applied in economics.\nFindings: The notion of threshold concepts such as opportunity cost can be extremely useful in pedagogic design. They focus on how learning concepts changes student thinking, rather than the standard approach which treats many concepts as equal in importance. Designing appropriate instructional materials around threshold concepts is time consuming and this may prove a barrier to the notions widespread adoption.\nOriginality/value: This paper contributes to the literature discussing the practical application of threshold concepts in teaching first-year economics as well as identifying possible reasons for its relatively slow adoption as a teaching tool in university.\n",
"role": "assistant"
}
] |
When ICT meets schools: differentiation, complexity and adaptability
|
[
"Communication technologies",
"Schools",
"Functional differentiation",
"Organizational structures",
"Innovation",
"Israel"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: \"Happy families are all alike; every unhappy family is unhappy in its own way\", thus says Lev Tolstoy at the beginning of his celebrated book, Anna Karenina Tolstoy, 2000). This study is about \"happy\" organizations such as those that have successfully implemented information communication technology (ICT). Are they also all alike? Woodward (1965) found that organizations applying the same technology are \"all alike\" as a consequence of sharing similar structure. But when ICT is the subject in question, the picture is changed. Two international studies that examined more than 200 schools chosen by 30 countries as demonstrating the case for successfully implemented ICT-based pedagogy, found a wide variety of methods, practices and processes (Kozma, 2003; Venezky and Davis, 2002). It seems that the end of Tolstoy's sentence is more appropriate for describing the effect of ICT on school's organizational structure: every school integrates ICT in its own way. Why is this so? What effect does ICT have on the school's organizational structure and function? How does ICT serve school adjustment? And what lesson can be learned for other organizations? These are the questions this paper addresses.\nTheoretical background: The German scholar Niklas Luhmann, a former student of Talcott Parsons, was the founder of the sociological systems theory. Some of his highly abstract and analytical theories, especially those dealing with system complexity (Kallinikos, 2003; Hernes and Bakken, 2003; Bakken and Hernes, 2003), seem to be applicable to the issue of information technology and organizational form. I begin this theoretical background with a brief presentation of Luhmann's system theory, then examine it against current organizational theory and clarify the new direction it offers, and finally introduce its relevancy to the case of ICT and schools.Luhmann's system theory\nResearch design: The \"what\" types of the above questions and the school arena which contain too many uncontrolled complicating factors suggest an exploratory case study (Yin, 1989), the importance of the perspectives of those being studied suggests qualitative methods (Merriam, 1990), and Luhmann's too abstract theory suggests a hypothesis-generating study, rather than a hypothesis-testing one (Merriam, 1990; Eisenhardt, 1989). Thus, I have applied Eisenhardt's (1989) guidelines for a theory-building case study research along with the guidelines for an instrumental case study aimed at analyzing underlying issues, relationships and causes offered by Stake (1995). By employing this method, I was able to conduct an in-depth study of nine schools, looking at factors and processes involved in successful ICT integration, and conclude a model which has to be further validated with quantitative methodology. Following Majchrzak et al. (2004), in Table I, I have summarized my study design, comparing it to recommendations made by Eisenhardt (1989).School selection procedure\nResults: Luhmann's theory provided me with abstract concepts which served as a general framework in which I answered the research questions: The first part presents the findings for the first question regarding schools' communication that allows ICT integration; the second part shows the findings regarding the conditions for differentiation; and the third part introduces the effect of ICT on school adaptability. The results reflect the generalized level of analysis required for theory-building (Eisenhardt, 1989), based of course on the raw findings from each school, which are all present on the research web sites (http://muse.tau.ac.il/ict).Communication\nDiscussion: In this study I have proposed an analytical model for ICT integration into the school, while aiming to offer generic propositions that portray a set of connections among variables (Turner, 1986). By analyzing nine schools that have successfully implemented ICT-based pedagogical innovation, I found that ICT integration is related to school communication and contingent conditions that create different types of differentiation in the schools, which in turn increase internal complexity and enhance school adaptability. In a recursive process, ICT integration affects school communication and its sensitivity toward further ICT integration.Being extreme cases makes the chosen schools perfect for demonstrating ICT integration (Merriam, 1990), however, this might limit the generalizability of the findings. But since the aim of this study is analytic generalization (Yin, 1989, p. 38), the nine cases provide the opportunity to further develop a theory regarding the substantive area relating to ICT and school structure, and even more formal theory regarding information technology and organization structure (Strauss and Corbin, 1990). Therefore, this study should be judged by the degree to which it fosters new insights and raises new questions and propositions for further research (Eisenhardt, 1989), quantitative as well as qualitative. I use these questions and propositions to summarize my findings and to indicate their relevance for new areas.Role of communication\nConclusion: According to my model, I have proposed several research questions and propositions to stimulate further research. The unique context of extreme cases of ICT integration raises a challenge for further research to test the generalizability of these propositions both in regular schools and other budgeted organizations. I found that positive communication regarding ICT, which reflects a successful history of innovation, high self-esteem, stakeholder interest, ready infrastructure, and a certain leadership style, can ease the introduction of ICT into schools. Assuming that these factors are effective in all service and budgeted organizations, I would expect to see similar results in other cases. One more important question is how culture affects interaction between schools and ICT. According to the two international studies, the Israeli schools were found to present quite widespread ICT-based innovation (Kozma, 2003; Venezky and Davis, 2002). Still, further study is needed to analyze the cultural effects.I also found that certain variables correlate with each type of differentiation, and probably apply to different kinds of internal patterns in schools. Further research is needed to find how this pattern is created and maintained, and what is the role played by each variable.One more important finding is that the internal complexity raised some of the organizational dilemmas. Since this study was carried out at one point in time, it was not suitable for detecting the mechanisms of balancing the enduring dilemmas. As loose couples were found to be a kind of solution for professional-bureaucratic control (Weick, 1976), more research is needed to find how the school responds to dilemmas in their entirety.In addition to suggesting a working model, my findings also have practical implications. First, the importance of school communication as a precondition for successful ICT integration indicates the need to prepare schools for ICT. This means that natural diffusion from the environment (top-down) is not enough, and neither is sporadic teacher training. Preparing schools further suggests intervention at the top management level (principal, deputy, coordinators) and a constant effort to assimilate ICT into the school's decisions. Since ICT integration is the best way to reinforce communication regarding ICT integration, supporting an initial ICT-based project might also be a good idea for enhancing further significant ICT usage in schools.Second, awareness of the conditions affecting each type of differentiation enables schools a better choice of ICT usage, and enhances school control over the implementation process. Finally, the findings indicate that to stabilize organizational dilemmas amplified due to ICT integration, additional recourses can be of help: appropriate rewards for teachers, special training for students, professional autonomy for teachers, and additional budgets for an ICT-based learning environment.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this study is to explore the interaction between information communication technology (ICT) and the school's organizational structure, and propose an analytical model based both on Luhmann's system theory and empirical findings.\nDesign/methodology/approach: - The approach of building a theory from a case study research along with an instrumental multi-case study method were applied to analyzing nine Israeli schools that successfully implemented ICT-based pedagogical innovation.\nFindings: - The findings suggest that ICT generates three kinds of differentiation within the school's structure: segmentation, stratification and functional differentiation. The type of differentiation correlates with the school's communication and set of contingencies which includes ICT usage types, leadership style, time and space arrangement, source of expertise, and the champions - those who bear the burden. All the differentiation types were found to increase internal complexity and enhance school adaptability, and in a recursive process, affect school communication and its sensitivity toward further ICT integration.\nResearch limitations/ implications: - The fact that the chosen schools demonstrated extreme cases of successful ICT integration might limit the generalizability of the findings, but nonetheless it offers a significant contribution to the development of a substantive ICT and school structure theory.\nPractical implications: - Supporting ICT integration called for intervention at the school's top management level; supporting an initial ICT-based project enhances the school's control over the implementation process, and provides schools with additional resources.\nOriginality/value: - This study bears a substantive theory regarding ICT and school structure, and fosters new insights and propositions for further research.\n",
"role": "assistant"
}
] |
Decision-making training for occupational choice and early turnover: a field experiment
|
[
"Career development",
"Employee turnover",
"Decision making"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: New entrants are highly responsive to organizational influence during their initial encounters with the organization. The relative confusion and disorientation that characterize this phase increase their susceptibility to cognitive-informational cues as well as to emotional and social support cues (e.g. Louis, 1980; Morrison, 1993; Ostroff and Kozlowski, 1992). The majority of studies that examined organizational entry interventions have focused on the cognitive-informational aspects of these interventions, and confirmed the benefits of providing accurate information early on. Considerable research shows that the Realistic Job Preview (termed RJP) results in increasing commitment and reducing turnover (for reviews, see Meglino et al., 2000; Phillips, 1998).\nMethod: In a constructive replication, the same hypothesized relationships among the same theoretical constructs are tested, while varying their operationalization. We varied three such aspects.First, we varied the timing of the DMT provision and of its outcome measurement. In the Ganzach et al. study DMT was administered eight months prior to entering the organization. The relatively long period that elapsed between the intervention and entry might have obscured a sense-making interpretation. In the present study DMT is administered a few weeks after organizational entry, upon entry into the particular route of training, and its effect is measured twice, after four months and then ten months.Second, whereas the sample in the previous study was drawn primarily from a highly motivated population, the present sample is drawn primarily from the lower end of the population, in terms of prior attitudes towards the organization. The dynamics of reducing turnover among recruits whose initial level of commitment is low might be qualitatively different than that needed to boost growing levels of commitment (Bretz and Judge, 1998).Third, by measuring an additional variable, Perceived Organizational Support, the present study can better distinguish between social exchange and self-determination. The social exchange explanation suggests that the influence of DMT on turnover depends on enhanced Perceived Organizational Support. On the other hand, the self-determination explanation suggests that the provision of DMT reduces turnover without necessarily increasing the perception of support.Context and procedure\nResults: The percentages of the dependent variables - the two kinds of turnover - and the means (standard deviations) of Perceived Organizational Support are presented in Table I by experimental condition.Supporting our prediction, the results show that Program Turnover in the DMT condition was lower than in control, indicating a large effect of DMT on turnover - a decrease of about a third. Despite the low power of kh2 in detecting differences in proportions that deviate from 50 percent, the difference between Program Turnover in the two groups was significant, kh2=4.2, p<0.05.Six months after completion of training, Occupation Turnover among the programs' graduates - a statistically independent measure for the effect of the experimental manipulation - was considerably lower in the DMT group compared to control (1.9 per cent versus 4.2 per cent). However, the difference did not reach statistical significance (perhaps because of the overall low levels of Occupation Turnover2).On the other hand, we did not find a significant difference in Perceived Organizational Support between the DMT and the control group, t(443)=0.7.\nDiscussion: The present study provides further evidence to the usefulness of a simple intervention that reduces early turnover. We found that a structured opportunity to reflect upon occupational options via decision training results in stronger adherence among new entrants to their choice and reduced turnover. This finding is now generalized to a different population in terms of prior motivational profile (its lower end) and to different intervention timing, though further research is certainly needed to generalize it to non-military settings and to other kinds of organizational contexts and occupations.From a theoretical standpoint, the results of this study are not consistent with the social exchange rationale. The predicted elevation in Perceived Organizational Support among participants who participated in the DMT was not supported. On the other hand, the results are consistent with the self-determination rationale. The stronger behavioral perseverance of the DMT participants can be accounted by post-decisional influence, that justify prior choice. DMT participants were presumably more conscious of their choice process. Aligning their attitude to their awareness that their own choice had led them to this program, the DMT participants persevered in the face of hardships. This interpretation is in line with the view of behavior as shaping future attitudes and actions, and of people as engaging in retrospective processes of self-justification (Brockner, 1992; Salancik, 1977; Salancik and Pfeffer, 1978). In the context of organizational entry, this view suggests that the manner in which a job choice decision is made affects the decision maker's subsequent behavioral commitment (see also O'Reilly and Caldwell, 1981).Two points need to be made with regard to retrospective justification in these circumstances. First, self-determination is especially relevant when expressed desires are fulfilled. If one gets the assignment that one has reflectively chosen, one becomes more committed, retrospectively, to this assignment (compared to a non-reflective, unaware chooser). On the other hand, when assignment is inconsistent with choice, there is no reason to expect retrospective binding or commitment. In fact, based on the escalation of commitment literature, we might even expect reduced commitment among reflective compared to non-reflective choosers, as a way to retrospectively justify what one did not get (Schoorman, 1988; Schoorman and Holahan, 1996; Staw, 1981). For technical constraints that are common in field studies, we could not analyze the turnover data separately according to the degree to which single preferences had been met, but in future research such an analysis is needed in order to further clarify the self-determination explanation.Second, the supportive evidence that favors the self-determination over the social exchange interpretation should lead to a careful re-definition of the construct that underlies the DMT effect. The obtained higher perseverance in this study should be viewed as stronger commitment to a course of action, or as commitment to professional training, rather than as organizational commitment, which the social exchange argument had predicted.Our evidence rules out an alternative (prospective) explanation, according to which the lower turnover among the DMT participants was caused by more desirable assignments and better person - job fit. First, the two groups did not differ in the degree of correspondence between the participants' top preferences (as marked on the Preference Form) and the actual job that they eventually received (kh2 =0.052). On average, 64 per cent of participants in this sample were assigned to one of their three top choices. Second, mean assignment satisfaction (a measure that was obtained during the first days of training program, after the assessment week) was similar for the two groups [t (462)=0.61]. However, our evidence at this point cannot rule out the possibility that the choices of the DMT participants were more appropriate for them, so that among those that did get what they wanted, there might have been a better person - job fit (e.g. Chen et al., 2004; Wolniak and Pascarella, 2005).Returning to the social exchange explanation, the lack of support for this explanation in the present study does not contradict the operation of social exchange processes in employment contexts in general (Birdi et al., 1997; Gaertner and Nollen, 1989; Nordhaug, 1989), and during organizational entry in particular (Ganzach et al., 2002; Meglino et al., 1993; Meglino et al., 1988). The lack of significant elevation in Perceived Organizational Support among the DMT participants in the present study might be an indication for the robustness of POS among the entire pool of participants. A single organizational gesture might not be enough to alter the perception of that organization. The perception of organizational support in this study was probably shaped by the first few weeks of service, in which the DMT was but one event. It is possible that a short DMT intervention might be more effective in enhancing POS when it occurs prior to any substantial contact with the employer (as was the case in the Ganzach et al., study).Finally, an important implication of this study is that encouraging the initial reflection and deliberation of new entrants regarding their own direction in the organization, holds promise of getting a more committed workforce. According to the career management literature, increased awareness of self and of alternative occupations leads to more effective goal setting (e.g. Greenhaus et al., 2000), which, at an early career phase, involves a well-thought decision to enter a particular occupation. The DMT tool supports this process. The results of this and the previous study suggest that its provision increases commitment to the chosen goal. It is therefore recommended to consider DMT-type programs as an addition to early socialization practices.\n",
"role": "user"
},
{
"content": "Purpose: - The study seeks to examine how a short intervention, aimed at enhancing occupational choice skills, influences turnover during the early stages of organizational membership. It seeks to explore two theoretical rationales for this effect: social exchange and self-determination.\nDesign/methodology/approach: - The study is a \"constructive replication\" of previous research, and it employed a similar field experimentation methodology. Groups of candidates in the Technical School of the Israeli Air Force were randomly assigned to experimental (Decision Making Training) and control groups. Perceived Organizational Support and turnover were measured at two points in time.\nFindings: - The results showed that the intervention reduced turnover (relative to a control group) as measured at two points in time - at the end of a technical training program and six months later into their military service. Perceived Organizational Support was not enhanced by the intervention. The experimental results were consistent with the self-determination explanation more than with the social exchange explanation.\nPractical implications: - The practical benefits of Decision Making Training during early encounters with the organization are discussed. Decision Making Training is recommended as an effective tool during the career exploration phase, to facilitate the integration of self and environment awareness and to enhance self-determined career goals.\nOriginality/value: - Provides further evidence of the usefulness of a simple intervention that reduces early turnover.\n",
"role": "assistant"
}
] |
Evaluation of section heads' performance at Kuwait secondary schools
|
[
"Job evaluation",
"Principals",
"Teachers",
"Secondary schools",
"Kuwait"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The human element is the major instrument of change and development. Therefore, employees need guidance and supervision to optimize the level of their performance, improve their services and consequently increase their production. The educational process, where school leaders play a central role, is supposed to lead the change and development (Ahmad, 1987). Hence, evaluation of performance is essential for determining the real achievement rates, during which staff performance will be measured and evaluated to collect information needed for future development of their performance skills and for increasing their chances for promotion (Obaidat, 1995). On account of this, the performance evaluation of heads of educational sections at secondary schools is an important educational process intended to observe and identify the points of strength and weakness in administrative practices to ameliorate the educational process (Brown and Rutherford, 1999).In the past, performance evaluation was a confidential job; in other words, the job was done by the principal. Although, it was the section heads' responsibility to undertake this task, they were ignored (Olaim, 2000). Since performance reports are key documents in the professional life of employees demanding the full attention of the evaluators and supervisors, the goal of this research was to study the performance report for section heads at secondary schools in Kuwait; as more talk and observations are constantly being made on this issue (Al-Ali et al., 2001). This called for the need to know the opinion of the section heads' principals at secondary schools in Kuwait.\nBackground of the study: The evaluation of concerted efforts exerted by the heads of educational sections reveals how successful they have been in meeting the sought goals. The process is governed by a set of criteria incorporated into an \"evaluation form\" and graded by the supervisor and principal, in addition to the section head in certain parts of the form. Since much importance is attached to the evaluation form, it is essential to know what the evaluators think and who is undertaking the evaluation process by looking at certain aspects relating to the evaluation of the section head. Such aspects should include: the applicability of some efficiency factors, the scales of measure used, training requirements and self-evaluation. Therefore, the study attempts to answer the following questions:* What do the section head, supervisor and principal think about the performance evaluation form designed for the section head at secondary schools in the State of Kuwait?* What are the proposals that can develop the instrument of evaluation used for section heads at secondary schools in Kuwait?* Are there any significant statistical differences among the answers given by the sample with regard to title, gender, study system, type of school and educational district?\nObjectives: The study attempts to examine the viewpoints of those involved in evaluating the performance of section heads; mainly section heads themselves, supervisors and principals. It sets out to determine the strength and weaknesses in the performance evaluation form designed for section heads. Also, the study examines the differences among the answers given by the sample. Finally, it provides recommendations for developing the performance evaluation form design.\nSignificance of the study: This study is important for several reasons. First, according to researchers, this is the first field study that examines the viewpoints of section heads, supervisors and principals collectively about the performance evaluation form designed for section heads at secondary schools in the State of Kuwait. Second, it provides recommendations and proposals for designing an evaluation form for section heads based on the outcome of a field survey and previous studies. Third, it helps lay down criteria for selecting and training section heads. Finally, it examines the viewpoints of researched participants involved in the evaluation of section heads, showing potential differences of opinion with regard to the performance evaluation form designed for section heads.\nOverview of the performance evaluation form designed for section heads in secondary schools: The Performance Evaluation Follow-up Committee at the Ministry of Education, State of Kuwait, has designed one form for evaluating the performance of the section head and the teacher. The purpose is to develop a clear-cut and fair policy for monitoring the professional growth of the section head and the teacher based on their promotional growth. The idea is also to enhance the performance of the section head and the teacher through bolstering their self-confidence and joint work with the school management to optimize the educational process and eliminate negativities (Ministry of Education, 2002). The performance evaluation form consists of five sections as follows:1. The first section contains information about the section head and the teacher. It describes their professional life, their commitment to work, the penalties they incurred and violations they committed throughout the study year. It also lists the training courses they took over the last five years.2. The second section contains self-evaluation data by the section head and the teacher. In this section, each carries out self-evaluation. They supply information in three areas: main activities and contributions they made throughout the study year (conferences, symposia, others); training they require at work; and proposed recommendations for work development.3. The third section contains comments on the early part of the study year, which delineates the positive and the negative aspects of the section head and the teacher: their character and administrative and skills.4. The fourth section contains the ten points against which performance is evaluated with regard to character and administrative skills. The points are measured on a five-item scale.5. The fifth section contains end of year comments, incorporating the opinions of the principal and the supervisor on the major positive and negative aspects of the section head and the teacher.The evaluation process is carried out as follows:1. The school prepares the performance evaluation form at the start of the study year and fills it out with job details.2. At the end of the first semester, both the section head and the teacher fill out the part on self-evaluation.3. The principal and the supervisor will write down their comments on the first period. Then those comments will be revealed to the section head and the teacher.4. At the end of the study year, the principal and the supervisor will provide their ratings in figures in the fourth section.5. The principal and the supervisor fill out the fifth section on the major positive and negative aspects of the section head and the teacher.Once approved, the performance evaluation form will be submitted to the educational district.\nPerformance evaluation ethics: Judging a certain subject is a part of the overall evaluation process. It is based on specific clear-cut criteria (Haertel, 1986). It may be part of the objectives set for it. It may also be comprehensive, covering all levels and objectives sought. Evaluation is considered an essential component of the educational process. Therefore, it is a responsibility of the section heads to measure the results of the learning process and find out the objectives so far achieved, through establishing the learning objectives in a behavioral manner, selecting the right experiences for students and selecting the right learning means in a given learning environment.In its modern perspective, the idea of supervision seeks to improve the learning process that presupposes an innovative mutual relationship (Ahmad, 1987). The educational justification for the supervision rests with the fact that teachers, regardless of how highly-skilled in their jobs they may be, still need to interact with more experienced teachers so that they may adopt constant programs to improve their performance through a mutual process of interaction between the section head and the teacher on the one hand and the educational environment on the other (UNISCO, 1991), judging by the value of learning objectives sought by the school. The section head is considered a key player in educational reform and development processes. He serves as a link between the teacher and the educational management (Carlson and Kaczkowski, 2001). Also, he guides, facilitates, and informally monitors the instruction of new teachers, and gives ongoing emotional support during the first year of teaching (Richardson, 2003) new teachers often find themselves in the role of survivors rather than of producers. The applicability of the performance evaluation system to the section head presupposes convenient schemes and programs, which requires administrators who are capable of making the needed educational change. This is usually done by weighing up the points of strength in the scheme subject to evaluation (Al-Rasheed, 1986).\nReview of the literature: A review of the research on this topic was conducted to summarize the major recommendations of previous studies.Arabic studies\nMethod: The scope of study\nFindings: The study attempts to find out the opinions of section heads, supervisors and principals in the performance of section heads at secondary schools in the State of Kuwait. From Table I, it is clear that over 70 percent of the study sample believe that the performance evaluation form designed for the section head enables self-evaluation (73.3 percent); its content contributes to the principle of self-evaluation (70.5 percent); it encourages discipline at work (75.3 percent) by minimizing absence (71.2 percent); it enables them to develop their performance in the next semester of the study year (71.2 percent); it contributes to defining training requirements (72.3 percent) and indeed it encourages taking training courses (72.2 percent), as noted by Salmon (1977).The reason for the high percentage of agreement in the study sample is that the evaluation form contains areas on the visits made to the section head and the rate of his commitment to working hours (with respect to attendance, absence and going out by permission) in addition to a list of his violations and penalties. However, it is believed that the list of violations and penalties needs to be removed from the evaluation and placed in a separate folder.The form also contains a list of the training courses taken and passed by the section head over the past five years. It also contains data on self-evaluation, where he may point out his works and proposals for work development. However, Al-Ali et al. (2001) believe that candor needs to be continued.On one side, the study sample individuals believe that candid revelation of the evaluation result to the section head may raise problems between section head, supervisor and principal (59.3 percent). They maintain that low grading below what is expected by the section head leads to laziness (59.3 percent). Therefore, they believe that the candid revelation of the evaluation results should be done at the end of the first semester if the score grade was \"weak\" (46.4 percent). The same was emphasized by Al-Ali et al. (2001).The sample individuals also believe that the evaluation grade weights should be reduced from five to three (54.6 percent). This will give the section head a greater chance to score \"outstanding\" (72.9 percent). The same was confirmed by Al-Thubaity (1995). He recommended the placement of three grades beside every evaluation element.As to the part on positive and negative points in the performance evaluation form, there is somewhat of a proximity between agreement (45.8 percent) and disagreement (46.1 percent) as to the sufficiency of mentioning the opinion of the supervisor. This is because the supervisor, who is the principal, is fully aware of the section head's daily activities. Although the sample individuals believe that the evaluation form measures different sides of the activities (78.8 percent); however they are of the opinion that the total number of elements on the form need to be reduced (78.8 percent), since these cover all character, administrative and technical aspects. The same was emphasized by Brown and Rutherford (1999), Turner (1996) and Grumbles (1994).Demographic differences in responses\nConclusions and recommendations: In the light of the findings, the two researchers recommend the following:* Print a specially-designed performance evaluation form for the section head, different from that currently available, with a teacher evaluation form.* Retain the part on positive and negative aspects, but replace the word \"negative\" with the expression \"areas requiring improvement\".* Retain the part on the supervisor's opinion with regard to positive and negative aspects (areas requiring improvement). There is no need for the opinion of the next person in command.* At the end of the first semester, it will be enough to mention the positive areas and the areas requiring improvement without the need for oral or digital results to reduce related problems.* Reduce the evaluation weights from five to three.* Reduce the elements of evaluation to a convenient number.\n",
"role": "user"
},
{
"content": "Purpose: - The study attempts to examine the viewpoints of those involved in evaluating the performance of section heads in Kuwait secondary schools; mainly section heads themselves, supervisors and principals. It sets out to determine the strength and weaknesses in the performance evaluation form designed for section heads.\nDesign/methodology/approach: - A study was conducted on a sample of section heads, principals, and supervisors at secondary schools, six educational districts in Kuwait via a 15-item questionnaire.\nFindings: - The sample believe that the performance evaluation form designed for the section head: enables self-evaluation; contributes by content to the principle of self-evaluation; encourages discipline at work by minimizing absence; enables them to develop their performance in the next semester of the study year; contributes to defining training requirement; and encourages taking training courses.\nOriginality/value: - This study is important for several reasons. First, according to researchers, this is the first field study that examines the viewpoints of section heads, supervisors and principals collectively about the performance evaluation form designed for section heads at secondary schools in the State of Kuwait. Second, it provides recommendations and proposals for designing an evaluation form for section heads based on the outcome of a field survey and previous studies. Third, it helps lay down criteria for selecting and training section heads. Finally, it examines the viewpoints of researched participants involved in the evaluation of section heads, showing potential differences of opinion with regard to the performance evaluation form designed for section heads.\n",
"role": "assistant"
}
] |
Does a brand have to be consistent?
|
[
"Brands",
"Quality",
"Wines"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\n1. Introduction: It tends to be a given in much contemporary academic discussion about brand management that a brand will be consistent - that the product attributes will not vary. Thus the assumption of total quality management (TQM) (Reeves and Bednar, 1994) is that the consumer will receive precisely the same benefit whenever they consume a product. This assumption has also influenced approaches to the management of services (Reichheld and Sasser, 1990), and it is now suggested that marketing communications, as much as production, should be imbued with the TQM ethos (Haynes et al., 1999). Indeed, it has been claimed that \"a side effect of market globalisation is product similarity where brand differentiation is crucial to survival for companies\" (Matthiesen and Phau, 2005, p. 325). As Askegaard (2006) has noted, the academic focus on branding has concentrated on a standardisation versus adaptation debate rather than any broader attempt at interpretation. The purpose of this paper is to discuss whether or not that apparent assumption - that brands should be consistent in order to sustain consumer loyalty - is valid. The paper focuses on the consumption and branding of wine, and specifically \"high-quality\" wine, to exemplify the issues surrounding the discussion, but as will be seen, it is claimed that the argument has relevance to food, fashion and tourism brands, as well as to many products with a substantial aesthetic dimension, such as musical performers or dramatic companies. It may also have relevance to luxury goods, which can have a focus on being crafted rather than mass-produced (Beverland, 2005a; Kapferer, 2006).The paper is not the direct result of a single empirical study. Rather it is conceptual, resulting from reflection not only on the literature surrounding the nature of brands, but also on a series of studies and analyses undertaken by the author over a number of years (Charters, 2004, 2006; Charters and Pettigrew, 2006, 2007, 2008). It is also the result of a close reading of and immersion in many contemporary texts on wine written by wine critics in a number of Anglophone countries. As a result, there is no section of the paper covering in detail the methodology utilised (and readers are referred to the papers noted above for detailed methodological information). Rather, one section covers the context of the nature of brands. This is then followed by an analysis of the presentation and consumption of quality wine focusing on the importance of consistency in the process. The conclusion then pulls the various themes together.\n2. Context: 2.1. The nature of brands\n3. Do wine brands have to be consistent?: This section of the paper aims to do two things: first, to consider the notion of quality wine and second to examine the pricing of wine and what it reveals about consumer approaches to consistency.3.1. The nature of quality wine\n4. Discussion and conclusion: With most goods there is a clear desire for identical attributes, and the consumer will search for consistency from consumption event to consumption event. Indeed, with many goods (such as airplanes or cars) consistent quality is not just desirable but essential for the security of the consumer. However, there is a range of goods where both the nature of production and the expectations of the consumer mean that precise consistency in the product attributes is not only impossible, but also undesirable. While this paper has examined wine, it is probable that the argument could be extended to some foodstuffs (cheese, for instance, is said to vary in taste and style depending on whether or not the cows have fed on spring, summer or autumn grass) and aesthetic and ludic products.As was stated earlier, Aaker (1991) has suggested that a brand needs recognition, quality, reputation and loyalty. Aaker does not add consistency and it could be that the key term here is quality. For some consumers quality lies in the way that the product is delivered in varying contexts. It may therefore be necessary to distinguish between delivering the same product, and similar, consistently high quality brand values. The consumer who buys Chateau Leoville las Cases from year to year is not buying an identical product, but is buying into the same brand values of quality, production integrity and a style that reflects a similar blend of grape varieties and vineyards. At the same time, the subtle variations from year to year may in fact enhance the value of the brand to the consumer. Beverland (2005b), in his discussion of the production of luxury brands suggests that stylistic consistency is a core value in such a brand and consistency rather than duplication may be the key in this case. Another way of expressing this may be to say that customers distinguish changing brand attributes, but expect (within the constraints of the weather conditions or other variable production factors, the same level of brand quality. In each case, however, customer expectations are the key. It is not consistent content, but a consistency of quality and the ability to deliver up to and beyond expectations that is crucial for the brand. This is reflected across the range of products with an aesthetic impact; the orchestra must deliver consistent quality, whilst the attributes of the brand vary in the light of the external constraints of interpretation.Only recently have some researchers shifted to a debate about interpretation and co-creation debate (Askegaard, 2006). An interesting question for the future is how variable brands will operate. As a younger generation becomes involved in wine seeking reassurance in a highly fragmented market - a group who are said to be very \"brand savvy\" (Nowak et al., 2006) - it could be that a lack of consistency becomes more of a problem. On the other hand, as authenticity becomes increasingly important as a core brand value (Boyle, 2003; Brown et al., 2003; Jones et al., 2005) the kind of products which flaunt their authenticity by refusing to be identical each time they are consumed may become more popular. Furthermore, if younger consumers are indeed sceptical about the influence of brand managers (Holt, 2002), then they may, paradoxically, be more positive about brands where consistency is not a given, and where variability becomes part of the attraction of the product, allowing the \"co-creation\" of brand meaning with brand managers (Brown et al., 2003). Wines can be chosen because their variability, stemming from particular agricultural (or even production-related) factors which represent, for example, a year over which the consumer is nostalgic; this inconsistency can mirror historical memories that are significant for the drinker, and which they desire to relive or recreate. Musical performances may offer similar advantages for high-involvement consumers, reflecting various moods, events or life-experiences.Beyond this, if Holt's (2002, p. 87) argument about the \"post-post-modern consumer\" is accepted, and brands become significant as a \"cultural resource\" allowing the consumer to gain personal control over the use of a product, then again a brand which is variable, perhaps in multiple guises, allows great opportunities for the individual to shape the meaning the brand provides. Indeed, it could be argued that with the co-creation of brands (Holt, 2002; Schroeder and Salzer-Morling, 2006) consistency may actually be an impediment - its continual \"sameness\" gets in the way of the consumer's ability to reshape the brand for their own symbolic or experiential purposes.\n5. Managerial implications and applications: What, then, are the practical implications of this conclusion? The first point is that those involved in managing creative, aesthetic, perhaps even ludic brands, which are consumed substantially by high-involvement consumers, need to place less emphasis on the consistency of their offering than on overall quality, within the context of what they are delivering. Interest is important as a dimension of quality with these products and needs to be developed and underlined. Thus consumers may enjoy some quirkiness or variation in the brand if it adds to the interest of the product and their enjoyment of its consumption.By way of example, the Berlin Philharmonic Orchestra has one of the best classical music brand reputations in the world, but its performance will vary, albeit subtly, on discs representing different performances of the same piece of music. Where different conductors are involved, with shifting interpretations of the score, those variations may be deliberately sought out by the most knowledgeable consumers, who like to be able to compare changes in tempo, timbre or tone. Similar suggestions can be made for dramatic productions or the various fine art interpretations. Tourism experiences and events also fall into this category, and tourism managers may seek to stress variation and the unexpected rather than consistency and sameness. As suggested earlier, these kinds of products have a strong hedonic and perhaps symbolic dimension, where physiological and aesthetic taste, symbolic meaning and self-actualisation are particularly important. Brand managers need to be aware of the desire of some consumers to be involved in the co-creation of performance and the consumption event. However, this does make creating and managing the brand necessarily complicated.Thus it can be suggested that in the future more consumers will be actively searching for this kind of product variation if it enables them to become involved in shaping the brand, the way they use it and the way it is perceived in their brand community (Holt, 2002). Brand managers with this type of product who are ahead of the game may actually want to build variation and inconsistency into their product to encourage such an interactive consumer response. This comes with risks, but also offers substantial potential benefits when it is well handled.\n",
"role": "user"
},
{
"content": "Purpose: - This conceptual paper seeks to examine the issue of whether or not a brand needs to be consistent. Whilst this assumption often seems to be made within the marketing literature, it has not previously been examined.\nDesign/methodology/approach: - The paper uses the example of wine to explore the consumer's desire for consistency. The paper is not based on a specific research project, but rather reflects the findings of a number of projects, and of an engagement with commentators on wine over a number of years.\nFindings: - The paper suggests that rather than wine invariably being the same it is necessary for it to offer a regular standard of quality within varying external constraints, and that consumers may actually search out such inconsistency. By extension, it is argued that this is relevant for other goods, such as a foods and fashions, and aesthetic and ludic products.\nPractical implications: - It may be relevant for managers to build inconsistency into product design in some cases to give consumers the chance of involvement in brand co-creation.\nOriginality/value: - The paper examines an aspect of brands (consistency) not previously considered, and moves on the current debate about consumer participation in the development of brand meaning.\n",
"role": "assistant"
}
] |
Effects of quality management on hospitality performance in different contexts
|
[
"Innovativeness",
"Customers’ performance",
"Employees’ performance",
"Management attitude",
"Quality deployment",
"Quality management standard"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The hospitality industry faces multiple changes. The primary changes are economic issues, greater competition and increasing customer demands. In addition, human resources management has become more difficult due to certain employees' attitudes (Testa and Sipe, 2012). The challenge is to do more with less (Wang and Wang, 2009). Given that the hospitality industry contains high levels of service and customer interactions in production, service quality is the primary means of ensuring hotels' long-term managerial success (Testa and Sipe, 2012; Ostrom et al., 2015).\n2. Literature review: 2.1. Quality management and quality performance in the hospitality industry\n3. Methodology: 3.1. Sample and method\n4. Data analysis and statistical results: The results are presented in three sections. First, the validity and reliability of the measurement scales are analyzed. Second, the working model is analyzed using a structural equation model (SEM). Third, an additional contextual analysis is conducted to determine whether certain contextual factors (quality standards and innovative behavior) influenced the validation of the hypotheses.\n5. Discussion of results: 5.1. Results\n6. Conclusions: The findings of this study suggest a number of conclusions and recommendations for the service sector in general and for the hospitality industry in particular.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is threefold: first, to test the role of the managerial team's commitment to quality deployment, quality performance and firm performance; second, to shed light on the effects of a certified quality management system on this chain; and third, to analyze the effects of changes in this chain on different management team attitudes, namely, innovativeness.\nDesign/methodology/approach: Based on a survey conducted in a sample of 370 hotels from the region of Madrid (Spain), the model and hypotheses were validated using structural equation modeling analysis.\nFindings: The results reveal the importance of a commitment to quality for quality deployment. In addition, quality deployment has a strong impact on operations and employee and customer performance. However, only impacts derived from customers' perceptions of service quality affect firm performance. In addition, a contextual analysis was performed to verify whether various contextual factors (certification in a quality management standard and the innovative attitude of the management team) could modify the previously obtained results for the full sample.\nPractical implications: The management team should understand that its attitude toward quality management could affect the company's overall performance. On the one hand, the team should be aware that the cost of implementing quality practices could be a burden for employees in the performance of their daily activities. On the other hand, there is a positive relationship between involvement in addressing customers' requirements and financial performance. Therefore, a balance between the efforts to maintain a certain level of quality by the firm's employees and the service quality offered to the customers is crucial to achieving better financial performance.\nOriginality/value: The impact of the implementation of quality management practices on performance has been widely studied; however, the role of intangible factors has not been commonly discussed in service industries in general and even less so in the hospitality sector. This paper analyses the influence of the quality management team's attitudes toward innovativeness and the implementation of a quality management standard.\n",
"role": "assistant"
}
] |
A conceptual model of consumers' online tourism confusion
|
[
"Social media",
"Information processing",
"Internet experience",
"Online information confusion",
"Online travel information search",
"Tolerance for ambiguity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: As the rapid technological developments have enabled companies to provide an excessive amount of information about their products on the Internet, online search has become the dominant mode for consumers to acquire tourism information (Xiang and Gretzel, 2010). Consumers use the Internet to access and process travel-related information, to buy tourism products and services and to express their opinions about different tourism products (Martin and Woodside, 2012; Morosan, 2014). Although there are benefits for consumers to access enormous amount of information from different online sources such as industry suppliers (e.g. hotels, airlines, restaurants), intermediaries (e.g. travel agents), controllers (e.g. governments and administrative sectors), user-generated online social networks and many non-profit organizations (e.g. destination marketing organizations) (Xiang and Gretzel, 2010), there is also a greater chance for confusion and information overload as consumers must decipher and discriminate between sources offering what is often similar, complicated or ambiguous information.\nDiscussions and conclusions: Theoretical implications\n",
"role": "user"
},
{
"content": "Purpose: This study aims to develop a conceptual model demonstrating the antecedents and outcomes of consumers' online tourism information confusion.\nDesign/methodology/approach: A deductive approach was utilized to propose eight variables as antecedents of online confusion and five confusion reduction strategies as outcomes of consumers' online tourism confusion. The underlying mechanisms in which these variables might lead to consumers' online tourism information confusion are elaborated using elaboration likelihood model (ELM) (Petty and Cacioppo, 1986) as the major theoretical underpinning.\nFindings: The model indicates that consumers could experience overload, similarity and ambiguity confusion when the information acquired is too much, too similar and/or too vague. In addition, as suggested by the ELM (Petty and Cacioppo, 1986), online users who are low in learning orientation, price consciousness, cognition need and Internet experience and high in ambiguity tolerance are more likely to experience confusion because of their lower motivation/ability to process external stimuli.\nResearch limitations/implications: One limitation of this study is the lack of empirical test of the proposed model. Another limitation is that only five individual characteristics that might make online consumers prone to confusion were included. Other variables related to individual differences that could influence confusion should be explored as well.\nPractical implications: This paper provides valuable implications for online tourism marketers to address consumers' confusion during information search process. Five individual characteristics proposed as important antecedents of consumers' confusion can be utilized by online tourism marketers to develop customized online communication strategies for different segments.\nOriginality/value: This paper is one of the few studies that connect the concept of consumers' confusion to the online tourism field as well as discuss the concept of consumers' confusion through the integration of information provider and recipients' perspective.\n",
"role": "assistant"
}
] |
Intercultural training for US business expatriates in Taiwan
|
[
"Business expatriates",
"Intercultural training",
"International HRD",
"United States of America",
"Taiwan"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Continually improving information and communication technologies are accelerating business globalization. Human resource development (HRD) is a key factor in companies' success in the inevitable move to globalization (McLean, 2001). As businesses expand into global marketplaces, the number of expatriate employees grows. The success of employees who accept foreign assignments is becoming more important to the headquarters in their home countries. To cope with global competition, businesspeople need to be able to work effectively with people from a diversity of countries and cultures (Friedman, 2005; Harvey et al., 2009).As the first little dragon to emerge from Asia, Taiwan took advantage of its close strategic ties with the USA to become one of its major trading partners (Vogel, 1991). Researchers have found significant differences between the Chinese culture and Western cultures (Blackman, 1997; Brahm, 2003; Hofstede, 2001; Osman-Gani and Rockstuhl, 2009). Business interactions between people from different countries often fail due to cultural differences and lack of training about those differences. Many Western business expatriates experience serious problems due to cultural differences while on assignment in Asia (Black and Mendenhall, 1990; Dunbar, 1992; Shen and Lang, 2009; Tharenou and Harvey, 2006). Expatriates sent by US multinational companies have a higher failure rate than those of European or Japanese multinational companies (Harzing, 2001; Tung, 1988). To reduce problems in this area, it is necessary for US businesspeople to obtain effective intercultural training related to an overseas assignment.Although there has been an abundance of literature on intercultural training, rarely has research been done on Taiwan cultural-specific training. The purpose of this study was to discover cultural-specific training needs of US business expatriates on assignments to Taiwan. The study assessed Taiwan culture training needs of US business expatriates from the perspectives of both US expatriates and their Taiwanese colleagues, and compared the perceived importance of these intercultural training needs between these two groups. The study focused on finding answers to the following questions:* To what extent are specific intercultural knowledge and skills perceived to be needed by US business expatriates in Taiwan?* Do differences exist between US expatriates and their Taiwanese colleagues in terms of the perceived importance of these intercultural training needs?\nReview of literature: Today, Western businesspeople visiting Taiwan's major cities, such as Taipei or Kaohsiung, may find few differences between these modern cities and those in the West on the surface. However, how people conduct business is embedded in their traditions and cultural norms.Taiwan's business culture\nMethods: The methods used in the study are provided in this section, including participants, instrumentation, procedures, and limitations.Participants\nResults: On a 1-5 scale, the overall means and standard deviations of items in each category for US and Taiwanese respondents combined are:* Knowledge of the nation, M=3.55; SD=0.88.* Relationship building, M=4.01; SD =0.77.* Interpersonal communication, M=3.02; SD=1.07.* Business protocol, M=4.07; SD=0.75.* Legal issue, M=3.80; SD=0.77.* Living in Taiwan, M=3.50; SD=0.77.The mean of all 26 items rated by US respondents is 3.64 with standard deviation 0.93 (n=78). The mean of all 26 items rated by Taiwanese respondents is 3.53 with standard deviation 0.92 (n=78).To what extent are specific intercultural knowledge and skills perceived to be needed by US business expatriates in Taiwan?\nDiscussion: This study found that both US expatriates and their Taiwanese colleagues believed that it is necessary for US expatriates to have knowledge and skills in all six areas of the survey. US participants generally believed that all items are needed more than their Taiwanese colleagues did. They particularly perceived more need in categories \"knowledge of the nation,\" \"relationship building,\" \"business protocol,\" and \"legal issue.\" The differences between these two groups can be explained perhaps by the fact that, as the US participants themselves are business expatriates, naturally they experience a strong need to have cross-cultural knowledge and skills, particularly in those four areas. In addition, 48.7 percent of US participants have at least ten years of international work experience. These experienced business expatriates have already developed sophisticated cross-cultural competencies and become globally oriented. Therefore, they have higher expectation of others. On the other hand, the Taiwanese participants have less expectation of US business expatriates. As Taiwan is an island and the USA is a big country, the Taiwanese are more tolerant toward US expatriates' lack of understanding of Taiwan's culture.Knowledge of the nation\nConclusion: This study found that both US expatriates and their Taiwanese colleagues believed that US expatriates need to acquire knowledge and skills through education and training in knowledge of the nation, relationship building, interpersonal communication, business protocol, legal issues, and living in Taiwan. US expatriates believed that they had a need for more knowledge overall and in four of the areas than did their Taiwanese colleagues: knowledge of the nation, relationship building, business protocol, and legal issue. The study also found that US expatriates are in need of intercultural training; however, their companies have not paid enough attention to this area. Not only is cultural-specific training needed, but there is also a lack of cultural-general training among US expatriates in Taiwan.\nRecommendations for practice: The following suggestions are recommended for HRD practitioners when developing and implementing intercultural training for US business expatriates going on assignment to Taiwan:* For expatriates who are being assigned to Taiwan, training should include general knowledge or facts about Taiwan. US expatriates need to know the history, geography, and politics of Taiwan. They need to know the different groups of people in Taiwan, as well as regional cultural differences of Taiwan. It is particularly important for US expatriates to know the relationship between Taiwan and Mainland China, Taiwan and the USA, and the role of the USA in Taiwan-China relations.* The training should provide US expatriates with the knowledge and skills in building relationships with their Taiwan associates. US expatriates need to understand the importance of family and friends to the Taiwanese. They need to know the role of social relationships in business and learn how to build and maintain good relationships with their Taiwanese associates.* The training should include the knowledge and skills in interpersonal communication. US expatriates need to understand that there are different communication styles between cultures and know the communication patterns of Taiwanese people. It is very important for them to know how to deal with situations when communication problems occur between them and their Taiwanese associates.* The training should include Taiwan's business protocol. This includes the way business is usually conducted in Taiwan, how business meetings are usually conducted, and the negotiation strategies used by Taiwanese business people.* The training should include legal issues related to doing business in Taiwan. US expatriates need to know about Taiwan's legal system, the use of contracts in Taiwan, business ethical standards in Taiwan, and Taiwan's government regulations for business.* The training should provide US expatriates with information about living in Taiwan. This includes the life styles of people in Taiwan, the food and dining conditions, the commuting conditions, the housing conditions, and the schooling conditions in Taiwan. The training should provide information about the social support available to US expatriates in Taiwan and how to access it. The training also should inform US expatriates about how to cope with natural disasters, such as typhoons and earthquakes.* US business expatriates need intercultural training. Companies working across country borders should pay more attention to this training. In addition to cultural-specific training, companies should provide cultural-general training for their expatriates. They might also consider these concerns when selecting employees; those with a good liberal arts background might well be expected to bring some of these skills and knowledge with them to the job.\nImplications for research: There is a need to study a broader group of US expatriates in Taiwan. This will add validity and expand generalizability of these findings. Future research should expand the questionnaire used to include the items that surfaced in this study in response to the open-ended questions. To reduce the response bias that might exist in the current study, future research needs to compare the occupational and positional differences between US expatriates and their Taiwanese colleagues. Second, a model for Taiwan culture training should be developed and its effectiveness should be evaluated. Future studies can focus on the development and refinement of the model.Taiwan's business culture is significantly different from that in the USA. Although there has been a close economic relationship between Taiwan and the USA, limited research has addressed intercultural training needs for expatriates in Taiwan. A study in this area can help human resource practitioners in developing expatriate training programs. Research results can contribute to the knowledge base of expatriate training and development, as well as the development of theories in this area. There is a lack of theory developed in the literature on intercultural training. More research needs to be done to generate knowledge that can contribute to theory building in intercultural training, especially in Taiwan cultural-specific training. Finally, more studies need to be conducted to investigate expatriate failure among multinational companies and to examine the relationship between intercultural training and expatriate success.\n",
"role": "user"
},
{
"content": "Purpose: - This study aims to explore the intercultural training needs for US business expatriates on assignment in Taiwan. The study assesses Taiwan culture-specific training needs of US expatriates from the perspectives of both US expatriates and their Taiwanese colleagues and compares the perceived importance of these intercultural training needs between these two groups.\nDesign/methodology/approach: - This study used the survey method to assess the opinions of US business expatriates as well as their Taiwanese colleagues. A questionnaire was developed for the study. A total of 26 items were identified as knowledge and skills needed for US business expatriates in Taiwan. The items all fall within six categories: knowledge of the nation, relationship building, interpersonal communication, business protocol, legal issues, and living in Taiwan.\nFindings: - Data collected from 78 US respondents and 78 Taiwanese respondents were analyzed using matched pairs t-tests. Between-group differences for the overall 26 items and each category were examined. Results indicate that there was a significant difference between the US and Taiwanese respondents in the perceived importance of the overall items.\nOriginality/value: - Although there has been an abundance of literature on intercultural training, rarely has research been done on Taiwan cultural-specific training. A study in this area can help human resource practitioners in developing expatriate training programs. Research results can contribute to the knowledge base of expatriate training and development, as well as the development of theories in this area.\n",
"role": "assistant"
}
] |
Re-assessing minimum-age standards for children's work
|
[
"Children (age groups)",
"Labour",
"Labour law",
"Education",
"Human rights"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: This article calls for re-thinking the \"minimum-age\" approach to problems of child labour, exemplified in the ILO Convention Concerning Minimum-Age for Admission to Employment (No. 138 of 1973), known as the Minimum-Age Convention, and similar national regulations in many countries. We argue that a universalized policy of excluding children below a given age from \"employment or work in any occupation\" (article 2.1) is unjustified for at least the following reasons:\nILO Convention 138: an unreasoned hangover from history: The minimum-age approach derives from a historical legacy of romantic ideas of childhood seen against the treatment of children in industrial Europe. The ILO began in 1919 to set standards in relation to child employment with the prohibition of specific branches and forms of work, in particular factory labour. Fifty years later ILO's Convention 138 expanded these selective sector-specific bans into an almost absolute prohibition on the employment of school-aged children in all forms of work, specifying a minimum-age of 15 years (or the age of completion of compulsory schooling, where this is higher) below which no child \"shall be admitted to employment or work in any occupation\" (art. 1)[1]. Work and school were thus presented as fundamentally incompatible in children's development, distinctions between harmful and beneficial forms of work were neglected, and for the first time work itself became the thing to be abolished. In 1998, the ILO's Declaration on Fundamental Principles and Rights at Work included Convention 138 as one of seven \"core conventions\", meaning that all member states must \"respect, promote and realise\" it, regardless of whether they have ratified it or not.Convention 138 allows authorities to specify permissible \"light work\" for children as young as 13 (or 12 in less developed countries with poorer education facilities) that is part-time, safe and does not impede schooling, but not for younger children. The process of specifying and registering light work is usually so complicated that it is ignored (e.g. McKechnie and Hobbs, 1999, p. 94). So in practice part-time work, even an hour a week, remains technically illegal. \"Work\" is defined to include virtually all economic activities, paid or unpaid, in or outside the home, including many chores traditional for children and generally considered part of acceptable child-rearing practice. As a result, the policy needlessly affects many children whose limited part-time or seasonal work is not harmful and would not be characterized by most people as \"child labour\".Legal historian Marianne Dahlen has shown that between 1919 and 1973 the ILO's Minimum-Age approach - originally inspired by the 19th century Factory Acts in various European countries - failed to adjust to the contemporary realities of working children, although \"the members of the ILO knew a lot about the difficulties and inefficiency of minimum-age legislation when Convention No. 138 was adopted\". The global survey conducted by the ILO in preparation for Convention 138 provided ample evidence of the failure of previous Minimum-Age Conventions in abolishing child labour, and that the majority of child labour was performed in contexts of small-scale, informal sectors difficult to control by laws and labour inspection. The convention's approach, however, remained unaffected: \"the minimum-age campaign suffered from a permanent 'hangover from history'\" (Dahlen, 2007, pp. 287, 330-1; see also Woodhead, 2007).In both rich and developing countries, many have found the provisions of Convention 138 so sweeping as to interfere with acceptable child-rearing. Several countries have pushed for a change of main strategic focus, and in the mid 1990s the ILO began emphasizing harmful forms of work, adopting in 1999 the \"worst forms of child labour\" Convention 182. This convention demands the elimination \"as a matter of urgency\" of \"all work which by its nature or the circumstances in which it is carried out is likely to harm the health, safety or morals of children\" (leaving only harmless work covered by 138 and not by 182). While this implicit differentiation between acceptable and unacceptable, benign and harmful work seemed to many a step forward, not everyone agreed. The ILO continues to espouse universal ratification and application of Convention 138 as an \"overarching goal\" (ILO, 2006, pp. 6, 23). To combat child labour's \"worst forms\" defined in Convention 182, and at the same time to insist on a universal minimum-age policy is an awkward pairing of \"regulatory\" and \"abolitionist\" approaches that are not easily compatible.\nThe need for policy analysis: Minimum-age policies reflect a paradigm that assumes that children benefit from being withdrawn or excluded from work. There is little empirical evidence to support this assumption and considerable evidence to question it. As already noted, minimum-age policy has yet to be subjected - internationally or nationally - to comprehensive policy analysis to determine how its legislation and enforcement affect both children and society, who benefits and who loses, and whether the benefits are worth the costs. Good governance normally submits social and economic policies to such rigorous analysis based on careful research, but this accountability has not yet been extended to child labour policies.In the absence of comprehensive analysis, the research of several historians, economists, and social scientists provides a wealth of suggestive, though not systematic, evidence about the effects of child work, and interventions in it. For example, historians have considered how children of industrializing Britain fared under child labour laws of that time: they are sceptical as to whether such laws were of net benefit to children (e.g. Cunningham, 1991; Nardinelli, 1990, 2000). Economists have analyzed data in national Living Standards Measurement Surveys from various countries to determine the effects of work, schooling, poverty and interventions: they recommend economic incentives rather than prohibitions to improve the lot of working children (e.g. Grootaert and Patrinos, 1999). Many sociological and anthropological studies do not find work to be inherently negative, though it certainly may be harmful in specific contexts and conditions, and therefore question abolitionist perspectives (Nieuwenhuys, 1994, 2005; Mortimer, 2003; Leonard, 2003; Hobbes et al., 2007).This situation calls for research to determine how children fare under each policy approach, assessing and comparing their observed impact on children's safety, health, development, socialization and future prospects. The results should be used to ground policy in what is demonstrably in the children's best interests. Meanwhile, in the face of evidence that the minimum-age approach as currently practiced drives child labour underground, offers no protection to children who are exploited, and interferes with useful child-rearing practices, we suggest that Convention 138 should for the time being be suspended.\nFocusing on evidence: The minimum-age debate seems dominated, on the one hand, by largely normative assertions with little or no basis in research about the evils of child work and the virtues of minimum-age interventions, and, on the other, by empirically based equivocations and refutations. We sketch a few of the more important arguments, in each case questioning a common assertion in defence of minimum-age polices in the light of the weight of existing evidence. In this brief overview, we provide a few references as illustrations, not proof, of the points raised.Child work and poverty\nThe need for policy analysis on the impact of minimum-age policies on children: The questions raised above provide sufficient grounds to investigate rigorously the impact of ILO Convention 138 and the national laws deriving from its guidelines on children. Policy analysis should be broad-based and multi-disciplinary, investigating the extent to which the policy is achieving its purpose, distinguishing between \"winners\" and \"losers\" under the policy, and weighing its benefits against its costs.Benefits and harm are not easy to measure. Ideally, long-term controlled studies are required to assess how work in childhood and adolescence affects adult lives. Too often harm is assumed when work challenges cultural ideals of childhood, or when adults\" interests are threatened. While assumptions can be counterbalanced by paying attention to the views and understandings of the children and adolescents concerned, such perceptions do not provide conclusive evidence of long-term benefits or harm. Although suggestions have been made for conceptualizing differences between the negative, beneficial, and neutral effects of work on children, and by implication the effects of interventions (McKechnie and Hobbs, 1998: especially 41; White, 1994; Woodhead, 2004), these have not yet produced the tools required to analyse efficiently the complexity of real-life child work situations.\nThe legality of minimum-age laws: A further serious challenge - seldom recognised - to ILO Convention 138 and national minimum-age laws is the strong, but so far untested, possibility that this convention violates the fundamental human rights of children and is therefore open to formal legal challenge[3].The 1948 Universal Declaration of Human Rights and the 1966 Covenants clearly articulate everyone's right, without qualification by age, to work and to join worker associations. The Covenants, and after them the United Nations Convention on the Rights of the Child, do however provide for children's protection, which could perhaps be interpreted as permitting an abridgement of children's human rights in order to protect them. Superficially, this might seem to authorize a minimum-age policy that excludes children from the right to work.However, abridging of children's human rights would be legitimate only if it were demonstrated to be necessary (that is, if there are no protective alternatives available that would impinge less on their rights), and to be effective in providing the intended protection. Convention 138 meets neither of these conditions; it has demonstrated neither that the minimum-age approach is the uniquely necessary remedy, nor that the approach achieves its protective aims. It is not clear, therefore, that the ILO or its member states can legally exclude children from rights plainly granted to all without exception in the \"International Bill of Human Rights\".Furthermore, it is inconceivable that the fundamental human right of workers to organize could be legally denied to children in any circumstances whatever. Refusing them trade union protection is an egregious violation of their rights that has no defensible basis in human rights law. That such denial is based on an ILO international convention declaring that \"underage\" children should not be workers at all is inconsequential. The stronger human rights provision prevails. The policy analysis we recommend to investigate the impact of minimum-age policies on children should include these human rights issues.In conclusion, we concur with Cunningham and Stromquist (2005, p. 79):One challenge for the twenty-first century is to imagine a childhood as free as possible of the romantic conception. This will mean a childhood that does not place a taboo on productive activity, but at the same time ensures that the child worker is not exploited.\nNotes: 1. A lower age of 14 is allowed, as a temporary measure, in countries where the economy and educational facilities are insufficiently developed.2. For example, Minimum-Age (Agriculture) Convention 10 of 1921, Article 2 provides no minimum-age for agricultural work so long as the work takes place \"outside the hours fixed for school attendance\" and \"is not such as to prejudice attendance at school\".3. There is some discussion of this issue in Hanson and Vandaele (2003, pp. 122-31).\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to call for re-thinking of the universal minimum-age approach to problems of child labour.\nDesign/methodology/approach: - The authors point out that there has been no serious policy analysis on universal minimum-age approaches, and question common assumptions concerning such policies by reviewing available knowledge on the impact of work on children.\nFindings: - Available research does not support a presumption that blanket minimum-age laws are beneficial. In some cases, it is clear that they are injurious to children, underlining the need for systematic policy analysis.\nPractical implications: - The promotion of universalized minimum-age policies should cease until their effect on children has been reliably assessed. In the meantime, more energy and investment should be devoted to alternative, proven ways of combating forms and conditions of work that are genuinely likely to cause harm, and to promoting access to education.\nOriginality/value: - This paper contributes towards introducing more appropriate policy on children's work.\n",
"role": "assistant"
}
] |
External technology acquisition of SMEs in the engineering industry of Bangalore: What prompts them to move faster for acquisition?
|
[
"Economics",
"Technology transfer",
"Manufacturing industry",
"Small and medium sized enterprises",
"Industrial performance",
"Engineering"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Small and medium enterprises (SMEs, hereafter) have a strategic appeal to policy makers globally in view of their quantitative significance - number of enterprises, number of employment, value of production and value of exports, among others. In addition, they are the seedbed of entrepreneurship development, innovation and risk taking behavior and therefore provide the foundation for long-term growth dynamics and the transition toward larger enterprises (UNIDO, 2002). These are general merits observed with reference to SMEs, which vary in terms of size, structure, technologies, products, inter-firm relationship, internationalization and competitiveness from country to country.\n2. Literature review: Manufacturing is strategically important to the world economy (Wang et al., 2012). But manufacturing industries around the world are being affected profoundly by emerging technologies resulting in a significant increase in competition in local, regional and global markets (Chowdary and George, 2012). To meet these challenges, manufacturing enterprise will require new technologies, systems and paradigms that will enable them to enhance their manufacturing capabilities (Thomas et al., 2012).\n3. Objectives, scope and methodology: In the light of the above literature survey and identification of research gaps, we formulate our research objectives. The present study has the following objectives:\n4. Analysis and results: 4.1 Public policy for SME technology up-gradation in India\n5. Summary and conclusions: The present study has examined the dimensions of public policy support currently extended to SMEs in India for facilitating external technology acquisition achieve technology up-gradation and modernization, followed by ascertaining the key features of technology change achieved by 64 SMEs in Bangalore city through ETA. Finally, through Cox regression analysis, the key factors which influenced the first ETA of these 64 SMEs have been analyzed.\n6. Policy implications:: The above discussed empirical findings have an important policy implication. India has been a net importer of machine tools and its exports are only marginal (EXIM Bank, 2001; IVG Partners, 2010). The major inhibiting factors for increasing India's machine tool exports are absence of modern technology and inadequate product certification levels (EXIM Bank, 2001). While large machine manufacturers cater to India's heavy and medium industries, SMEs meet the demand of ancillary and other units (IVG Partners, 2010). Given the current demand-supply gap, there is a clear need for adding capacities in this sector. But the industry is moving toward increasingly sophisticated CNC machines, driven by demand from key user segments, such as automobiles and consumer durables. Given this, machine tool manufacturers, particularly SMEs, need to develop capabilities to cater to this demand and investments in this area could yield long-term benefits (IVG Partners, 2010).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to ascertain: first, India's public policy support for small and medium enterprises (SMEs) for external technology acquisition (ETA); second, objectives of SMEs for ETA; and third, factors which induced them to obtain their first ETA.\nDesign/methodology/approach: Public policy is examined through survey of literature, whereas objectives and factors influencing first ETAs by SMEs are analyzed based on primary data collected from 64 SMEs in Bangalore. Objectives of ETAs are analyzed descriptively whereas factors which facilitated/hindered early ETAs are examined through Cox regression analysis.\nFindings: Public policy for ETAs by SMEs includes technology information, assistance and fiscal incentives. The technology focus of these SMEs has been shifting from conventional lathes to computer numerical control (CNC) machines. Most of the SMEs have gone for technology up-gradation with the objective of improving product quality, scale expansion, and meeting customer demand. Majority of these SMEs have obtained their first ETA within six years of their inception. Firm level factors have significantly influenced the time taken by these SMEs for their first ETAs. Overall, when technology is well developed and easily accessible, SMEs would hardly look for external support for ETAs.\nResearch limitations/implications: The shifting technology focus from conventional lathes to CNC machines is a welcome development, which is driven by the need for \"competitiveness enhancement\". Since there is no major obstacle for ETAs, policy makers may focus more on providing SMEs with market information and market developments.\nPractical implications: The shifting technology focus from conventional lathes to CNC machines in Indian SMEs is a welcome development, which is largely driven by the need for better product quality, scale expansion and customer demand, and internal factors played a crucial role in the time taken by these SMEs in accomplishing their first ETA. As such there is no major obstacle for these SMEs in going for ETAs since technology suppliers are available at the door-step and finance is available from the banks. Therefore, policy makers may focus more on providing SMEs with market information and market developments in the domestic as well as international market.\nOriginality/value: This is a first attempt to examine public policy, objectives and factors influencing SMEs for ETAs in India, after 1991.\n",
"role": "assistant"
}
] |
The Rosarium Project: Building a digital collection on the genus Rosa using <oXygen/> and the TEI
|
[
"Digital humanities",
"",
"Digital collections",
"Roses",
"TEI",
"Text encoding"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The Rosarium Project will collect, highlight, digitally curate and bring to the Web historical primary sources about the genus Rosa. The project focuses on non-fiction materials written about the rose in English from the sixteenth century up to 1923. These resources will include books, pamphlets, ephemera and articles from popular magazines, scholarly journals and newspapers. The digital texts - with their accompanying images, a rose glossary and a bibliography - will be fully searchable on the project's own website. Researchers will be able to browse the books and articles, search by subject, rose variety and color via drop-down menus or do a keyword search of the whole collection. Users will have freely available online access to historical documents about roses.\nA little bit about the TEI: The TEI is an internationally accepted standard for encoding text in the digital humanities. The TEI specifies:\nPhase One Part A: The Rosarium Project kicked off in March of 2015. The principal researcher conducted initial research using online databases such as the American Antiquarian Society's Historical Periodical Collection and The Reader's Guide Retrospective, online catalogs such as Harvard's Hollis+ catalog and WorldCat, digital repositories such as the Hathi Trust, Google Books and the Internet Archive and the two English language bibliographies written on roses. Bibliographic records for materials discovered were created and stored in the online bibliographic utility RefWorks. Over 900 potential sources were identified in this round of research.\nPhase One Part B: In November of 2015, the principal researcher luckily was able to attend a workshop on basic TEI encoding held at Northeastern University by Julia Flanders and Syd Bauman, the principals behind the WWP. This workshop was crucial because it provided a general overview to the TEI mindset, explained basic encoding decision-making strategy and provided an opportunity to get detailed answers to specific questions from TEI professionals. Materials from the workshop also provided links to additional resources, which have proven particularly helpful in moving the project forward (www.wwp.northeastern.edu/outreach/seminars/intro_2015-11/presentations/welcome_and_wrapup/wrapup_02.xhtml). Among the resources was a link to an especially helpful website - TEI by Example (http://teibyexample.org/TBE.htm).\nPhase One Part C: The last part of Phase One of the Rosarium Project has begun. It will be the longest to date because it will be the transcription and encoding of the remaining articles of the original 163 identified through the Reader's Guide Retrospective. It is expected to continue for over a year, assuming the daily responsibilities of the principal researcher remain constant. A major concern is the amount of research needed to provide contextual information for these articles. Indeed, this research may well consume more time than the encoding itself.\nFuture work: Phase Two and beyond: In Phase Two in the Rosarium Project, the principal researcher will return to the publications indexed by the Reader's Guide Retrospective. She has discovered, while working with some of the primary resources, that the Reader's Guide was not completely rigorous in its indexing of materials written on the rose. This will mean reviewing each issue of the journals indexed to identify further material. Some of this work may be done online using the Hathi Trust, Google Books and the Internet Archive, but some titles will require hands-on research at libraries with major horticultural and popular culture special collections.\nList of resources: American Antiquarian Society Historical Periodicals Collection: www.ebscohost.com/archives/aas-thematic-collection\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to describe the Rosarium Project, a digital humanities project being undertaken at the Phillips Memorial Library + Commons of Providence College in Providence, Rhode Island. The project focuses on a collection of English language non-fiction writings about the genus Rosa. The collection will comprise books, pamphlets, catalogs and articles from popular magazines, scholarly journals and newspapers written on the rose published before 1923. The source material is being encoded using the Text Encoding Initiative (TEI) Consortium's P5 guidelines and the extensible markup language (XML) editor software <oXygen/>.\nDesign/methodology/approach: This paper outlines the Rosarium Project and describes its workflow. This paper demonstrates how to create TEI-encoded files for digital curation using the XML editing software <oXygen/> and the TEI Archiving Publishing and Access Service (TAPAS) Project. The paper provides information on the purpose, scope, audience and phases of the project. It also identifies the resources - hardware, software and membership - needed for undertaking such a project.\nFindings: This paper shows how straightforward it is to encode transcriptions of primary sources using the TEI and XML editing software and to make the resulting digital resources available on the Web.\nOriginality/value: This paper presents a case study of how a research project transitioned from traditional printed bibliography to a web-accessible resource by capitalizing on the tools in the TEI toolkit using specialized XML editing software. The details of the project can be a guide for librarians and researchers contemplating digitally curating primary resources and making them available on the Web.\n",
"role": "assistant"
}
] |
Extremes of acceptance: employee attitudes toward artificial intelligence
|
[
"Innovation",
"Employee attitudes",
"Digitalization",
"Artificial intelligence",
"Digital transformation",
"Not-invented-here"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Many companies are currently in a digital transformation process, and the full impact of this change may not be apparent. Consequently, many employees are still getting used to the growing importance of data and software for their established business lines in industries such as automotive, chemicals, electronics, machinery and fast-moving consumer goods. While the impact of digital transformation has not yet reached its inflection point, many firms' technological evolution moves to the next level based on the growing use of artificial intelligence (AI). By applying a variety of intelligent algorithms and advanced data analytics solutions, executives leverage the value of the data collected in the digitalization initiatives. While this is a natural move from a strategic perspective, the growing use of AI constitutes an additional challenge in many firms' transformation processes. In this regard, some employees are open to adopting the latest solutions because they see the opportunities these technologies offer. Many other employees, in contrast, are reluctant to accept these technological evolutions and may even be frightened by them.\nAttitudes to technology: Employee attitudes play a key role in the adoption of new technologies. There has been a substantial body of research examining this topic. The technology acceptance model suggests that various factors, including the usefulness of a new technology - as perceived by employees - affect the decision about using it or not (Davis, 1989). Several factors influence employee attitudes to technology, and these factors are related to the new technology, to the employee, to the human-machine interface, as well as other fields. Beyond the general importance of employee attitudes, not-invented-here attitudes are well known, and they refer to negative attitudes toward acquiring technology from external sources (Katz and Allen, 1982). As a result, employees may be reluctant to collaborate with external partners, and they prefer internal technology development in make-or-buy decisions (Lichtenthaler and Ernst, 2006).\nPositive and negative attitudes to artificial intelligence: The AI initiatives of most companies have focused on increasing efficiency based on replacing human work by means of advanced technology. This substitution of human work only comprises a minor part of the potential competitive advantages because the complementary use of human intelligence and AI will most likely offer more important benefits in the long run than the mere substitution of human work (Lichtenthaler, 2018b). Nonetheless, this substitution has also received major public attention with speculations about the number of jobs that may be lost in this transformation (Winick, 2018). Even in the case of complementarity of human intelligence and AI, executives in firms may face a human-or-machine decision at least for some of the key activities that could be completed by AI or human work in the future. Beyond this human-or-machine decision, there will be a growing relevance of new human-machine interfaces in light of intelligent algorithms and advanced data analytics.\nRoots of attitudes: Negative attitudes toward automation may be traced back to multiple roots, which favor the emergence of a specific degree of negativity in a given situation. These roots falls into five categories, which reflect the human-machine interface, as well as the dynamic evolution of technologies and typical behaviors in an employee's environment (Muethel, 2013; Lichtenthaler and Ernst, 2006). Figure 1 provides an overview of these selected factors. The first group of factors refers to the human that is involved in the interaction with AI. Here, the affinity toward new technology plays a critical role. Some persons are very technology-driven and are typically among the early adopters of a new technology, whereas many others are more reluctant with regard to applying the latest technology, including advanced analytics. An additional key driver is a person's convenience orientation (Marquis, 2005). This orientation reflects to what degree a person is rationally trying to ease and optimize activities rather than being emotionally bound to human interactions. Employees with these characteristics are more likely to exhibit positive attitudes in a given situation.\nImplications and outlook: The new concepts of no-human-interaction and IA attitudes deepen our understanding of firms' success and problems in implementing advanced data analytics. Moreover, the paradox of negative and positive attitudes among the same employees helps to reconcile diverging findings in extant studies about the perception of the opportunities and risks of AI (Hirsch, 2017; Winick, 2018). At the present evolutionary stage, positive attitudes may play a role in some situations, but in most business contexts, negative attitudes are more prevalent, and many executives face the challenge of shifting employees' negative attitudes toward neutral or IA attitudes. The groups of factors influencing the emergence of negative or positive attitudes also offer important starting points for actively managing employee attitudes. With regard to the human factors, executives may place a high emphasis on attitudes to new technology in the hiring process of employees, especially for those departments where a strong interaction with AI is required. In addition, executives may assign the role of gatekeepers to selected employees, who formally or informally support a balanced use of intelligent solutions in the organization. By championing the use of AI, these gatekeepers may become key influencers for other employees.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present paradoxical employee attitudes towards interacting with artificial intelligence (AI).\nDesign/methodology/approach: This is a conceptual paper, which builds on prior research, especially on the widely accepted notion of not-invented-here attitudes in technology adoption.\nFindings: Many companies experience barriers in implementing AI owing to negative attitudes among their employees. This paper develops the concept of no-human-interaction attitudes, which describe employees' preference to collaborate with real humans rather than having virtual colleagues. If they perceive a benefit from voluntarily using AI, however, many employees exhibit positive attitudes, leading to the concept of intelligent-automation attitudes. Jointly, these attitudes lead to the paradox that the same persons may have positive or negative attitudes to AI, depending on the particular situation. Firms need to address these attitudes because the interface of human and AI will be a key driver of competitive advantage in the future.\nOriginality/value: The new concepts of negative and positive employee attitudes contribute to our understanding of firms' success and problems in implementing AI. Moreover, the paradox of negative and positive attitudes among the same employees helps to reconcile partly diverging findings in extant studies. A thorough understanding of the roots of these employee attitudes, along with several examples, further provides immediate starting points for actively influencing these attitudes in practice.\n",
"role": "assistant"
}
] |
Willingness to pay for carbon footprint on foods
|
[
"Contingent valuation",
"Carbon footprint",
"Food products",
"Chile"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Worldwide concerns regarding climate change has brought many governments and international agencies to propose policies focused on reducing the greenhouse gases (GHG) emissions (e.g. Kyoto Protocol). Thus, several GHG reduction programs have been implemented, such as carbon markets, air miles, and eco-labels, among others. However, the most advanced and recent initiative it is the carbon footprint (CF), which is defined as \"the measurement of the total greenhouse gas emissions caused directly and indirectly by a person, organization, event or product\" (Carbon Trust, 2011). Labeling the CF on products allows consumers to make informative purchase decisions, given them the option of choosing products with low gas emissions.\n2. Theoretical and econometric framework: WTP can be defined as the maximum amount of money that an individual would pay for a product (or a new product's attribute) that is not in the marketplace, and hence, it does not have a market price. Eliciting the consumers' WTP requires especial procedures, such as travel cost techniques, hedonic price approaches or contingent valuation (CV) methods. The latter has been extensively used and consists of survey methods where people are asked to express their WTP for a product, \"contingent\" to a hypothetical marketplace (Haab and McConnell, 2002). Several survey methods (e.g. in-person interviews, telephone and mail surveys) and formats (e.g. open-ended questions, iterative bidding approach, and dichotomous format) are used in the CV methods.\n3. Questionnaire design and data collection: Eliciting the WTP in a CV method requires that consumers do not get confuse with the good or attribute that has to be valued. In this case, the attribute (the CF value) it is not known by consumers. Thus, the product and its price have to be well known, so consumers can focus only on the valuation of the new attribute. Certainly, the WTP for the attribute will be closer to the price the consumer would pay in a real market situation if he/she is familiar with the product and its price (Ajzen et al., 2004). With this in mind, and considering the products' budget shares, we choose two products: fluid milk and bread. Prices of both products are well known by most Chilean people. Indeed, fluid milk is sold in boxes (plastic coated paperboard) of 1 liter, and although there are several varieties of milk, their prices are very similar. On the other hand, the most common bread (the \"hallulla\") is presented in small containers from where people can pick up and buy the wanted units, and price by 1 kilogram is well known. Besides, these two products exhibit different shares of monthly expense. According to the Chilean Institute of Statistics (Instituto Nacional de Estadisticas, 2011) bread occupies a 2.04 percent of an average monthly basket (the highest among food products), and fluid milk only a 0.50 percent.\n4. Results: As expected, most consumers are not familiar with the concept of CF (only 13 percent of them know the meaning of CF). However, once consumers receive an explanation about what the CF and CF labeling are, 93 percent of them would like to know the CF in food products. Moreover, a 91 percent of them would prefer to purchase foods with low CF instead of those with high CF. Table I shows the percentage of positive answers (\"yes\") to these questions by some consumers' characteristics (sex, education, activity, household income and age).\n5. Conclusions: This study has shown that, contrary to the general belief, Chilean consumers are very responsive to the global warming problem. In fact, although they are not familiar with the concept of CF, they want to receive more information related with this topic and they also would prefer foods with low GHG emissions. Moreover, they would be willing to pay an extra amount for food products with low CF values.\n",
"role": "user"
},
{
"content": "Purpose: The aim of this paper is to elicit the willingness to pay (WTP) of Chilean consumers towards the carbon footprint of food products (fluid milk and bread), controlling for several consumer's attributes.\nDesign/methodology/approach: A contingent valuation method (double bounded dichotomous choice) was used to analyze the WTP, wherein personal interviews were applied to a sample of 774 supermarket consumers.\nFindings: Chilean consumers show a positive attitude towards the carbon footprint concept. Consumers are willing to pay 29 percent for fluid milk and 10 percent for bread, over their average prices. It seems that WTP for foods is negatively related to the share of the product on the monthly household expense.\nResearch limitations/implications: Based on the results of this work, the WTP is unique for each product, and hence, any attempt to extrapolate these results must be done with caution. Further research should include a wider set of products.\nPractical implications: Local Chilean companies could obtain an advantage over their competitors (e.g. gain market share) if their products exhibit the carbon footprint. Moreover, if a company has a particular product with a competitive (low) CF, it could charge prices higher than those of its competitors.\nOriginality/value: This is the first work that provides an analysis of the WTP for carbon footprint labeling of foods in developing countries.\n",
"role": "assistant"
}
] |
Informal and formal learning of general practitioners
|
[
"Informal learning",
"Behavioural change",
"Formal learning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Lifelong learning is important to keep track of developments in professional domains because knowledge and insights develop fast, leading to a transition in professional's work routines (Fenwick, 2013; Tynjala, 2008). Even though opinions differ about the boundaries of formal and informal learning (Fenwick, 2010; Malcolm et al., 2003), learning that takes place while performing the job seems to be more efficient than formal training alone to learn job-related skills because skills are often situation-specific and need practice at the workplace (Eraut, 2004a). Formal learning is not sufficient for solid transfer of training and as a consequence, formal learning alone is probably not the most efficient way to accomplish behavioural changes (Boud and Middleton, 2003). Informal learning provides the learner with opportunities to practice the knowledge from the formal training in an everyday, workplace situation. Studies in informal learning have been done in many professional contexts, (Bjork et al., 2013; Goldman et al., 2009; Wofford et al., 2013), but studies about informal learning of general practitioners (GPs) are scarce (Campion-Smith et al., 1998; Stanley et al., 1993).\nMethods: For this study we chose a qualitative design to obtain insight in learning processes that might have changed prescribing behaviour of GPs. What and how the GPs thought they had learned during and after the online training was investigated with an interview. The script of the interview was first tested during a pilot with one GP to test whether the questions were understandable and made sense for GPs. A convenience sample was used (Neuman, 2009). During the process of interviewing the number of participants was determined (Glaser and Strauss, 1967). After nineteen interviews it was decided that saturation was reached and no other GPs were interviewed. The Netherlands Association for Medical Education (NVMO) and the Ethical Review Board gave permission for this research design.\nResults: During the analyses of the interviews themes were identified and those most relevant for the research questions were selected. Quotes illustrate the findings. The web-based training was thought to have influenced prescribing behaviour, also as a result of several informal learning activities. Additionally, being a research participant and/or a supervisor appeared as themes that influence prescribing and consultation behaviour. In the following sections these themes will be discussed.\nDiscussion: In our study, we explored the ways GPs had learned from a web-based training, through formal and/or informal learning. The results indicate that GPs who undertook more informal learning activities expected that their prescription behaviour had changed more and that GPs mostly learned from the content provided in the formal training during interaction with patients in routine consultations. This way of informal learning has been indicated already by Stanley et al. (1993). These results support the notion that it is most efficient to combine formal training with informal learning activities (Slotte et al., 2004). Furthermore, it became clear that the complete intervention was important for a behavioural change of GPs, not only the formal web-based training, but also the information booklet for parents, being a research (trial) participant and the informal learning activities afterwards. GPs supervising a GP trainee had more informal learning opportunities at their disposal because these GPs proved to be more used to reflection and open for discussions about prescribing decisions.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to understand the influence of formal learning from a web-based training and informal (workplace) learning afterwards on the behaviour of general practitioners (GPs) with respect to prescription of antibiotics.\nDesign/methodology/approach: To obtain insight in various learning processes, semi-structured interviews were conducted with 19 GPs. These interviews were transcribed and analysed with a theory-based template, that had been defined beforehand, but with an open mind for emerging themes.\nFindings: The web-based training was perceived by GPs to change their prescription behaviour, mostly as a result of informal learning processes. Being a research participant and being a supervisor appeared to create most opportunities for informal learning.\nPractical implications: The current research shows that being a research participant and/or a supervisor enhance informal learning activities, for example, reflection and social interaction, and thereby formal training becomes more effective. It is recommended to remind GPs regularly to reflect on their prescribing behaviour and to stimulate them to reflect and seek social interaction besides participating in formal training.\nOriginality/value: Our study adds to the existing literature by considering informal learning processes in an evaluation of the perceived effects of formal training. Our findings have implications for the design and evaluation of formal trainings with the purpose of behavioural change of doctors.\n",
"role": "assistant"
}
] |
Effects of customer perceptions in multichannel retail banking
|
[
"Engagement",
"Retail banking",
"Multichannel"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In the last decade, the retail banking industry has faced important changes. Innovations based on information technology, the threat of new entrants from the computing industry, customers who are more and more technologically knowledgeable and demand more transparency and the effects of the last financial crisis have all led banks to implement actions to tackle this new environment (Alt and Puschmann, 2012). One of these actions has been to adapt client service interfaces, investing in online channels and reducing branch networks. Investment in the online channel is a bank's response to a type of consumer that is becoming more and more familiar with online transactions and demands services and information that can be accessed at anytime and anywhere from their computer, tablet or mobile. Banks have needed to reduce the number of branches in order to remain competitive. Thus, banks need to redefine their value proposals in the current context, integrating their services in the offline and online channels (Bapat, 2017; Kingshott et al., 2018).\n2. Literature review: 2.1 Conceptual background\n3. Methodology: The hypotheses underlying the model were tested through data collected from a sample of customers in the banking sector. Specifically, data were gathered by means of a personal survey carried out in Spain by a market research company in the last quarter of 2017. Through a non-probabilistic quota sampling procedure, the sample size was proportionally adjusted to the population of each of the top 10 Spanish cities across different regions.\n4. Results: The proposed model was examined using partial least square (PLS) regression with SMART-PLS software. Due to the multidimensional nature of the offline and online service perceptions and customer engagement, these constructs were operationalized as second-order factors. In relation to convergent validity analysis, two factor loadings were below the common threshold of 0.7 and therefore these items were eliminated from the analysis (OFFESC1, ENSOC2). All the other factor loadings were statistically significant and above that limit (see Table III). Moreover, Cronbach's a, composite reliability (CR) and average variance extracted (AVE) were also above or close to the recommended thresholds of 0.8, 0.7 and 0.5, respectively (Hair et al., 2010). These results provide support for the convergent validity and reliability properties of the scales.\n5. Discussion: 5.1 Theoretical implications\n",
"role": "user"
},
{
"content": "Purpose: This paper focuses on the multichannel strategy in the banking sector and its effects on customer engagement. Specifically, the purpose of this paper is to propose a model in which customers' perceptions of offline and online channels are related to brand trust and brand commitment, which ultimately lead to customer engagement.\nDesign/methodology/approach: An empirical study was carried out on a sample of 306 individuals and data were analysed through partial least squares.\nFindings: The results show that offline experience is more important than online experience in terms of impact on trust and commitment, which are closely linked to customer engagement. Online experience does not have a significant direct influence on brand commitment and its effect on brand trust is moderated by the customer's familiarity with the channel.\nOriginality/value: These findings contribute to the advance in the current knowledge of the joint role of online and offline channels with the aim of strengthening customer relationships. From a managerial viewpoint, customer perceptions formed by their experiences in bank branches are more important than customer perceptions of the website's performance in the explanation of trust and commitment.\n",
"role": "assistant"
}
] |
The impact of service orientation in corporate culture on business performance in manufacturing companies
|
[
"Services",
"Organizational culture",
"Business performance",
"Manufacturing industries",
"Germany",
"Switzerland"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Product manufacturing companies increasingly exploit the financial, strategic and marketing opportunities associated with services (Martin and Horne, 1992; Oliva and Kallenberg, 2003; Brax, 2005) and extend the share of revenue and profit attributed to services. Companies such as IBM, Otis, SKF or Xerox, for example, create more than 30 percent of their revenues through services and the share of profit attributed to services often exceeds 50 percent.To achieve these service-related performance outcomes, companies have become more responsive to service business and have developed the service orientation in organizational structure, human resources, processes and measurement, and rewards and business performance. Although researchers agree on these organizational factors, the role of service orientation in the corporate culture is not without debate (Neu and Brown, 2005; Oliva and Kallenberg, 2003; Davies, 2004; Galbraight, 2002; Homburg et al., 2003; Jacob and Ulaga, 2008). On one hand, service culture functions as a resource that is difficult to imitate, creates competitive advantages, and enhances performance of manufacturing companies (Homburg et al., 2003; Matthyssens and Vandenbempt, 1998). Service culture, for example, could be beneficial for increasing a market and customer-centric orientation, which in turn would contribute to performance (Neu and Brown, 2005, 2008). In other words, service culture is a hard-to-imitate resource resulting from customers and employees who co-create service (Vargo and Lusch, 2008a).On the other hand, service culture can have an eroding effect on resource advantages based on product-oriented culture (Bowen et al., 1989). Value, meanings, beliefs and goals related to services may create conflicts with product-oriented organizational units leading an internal resistance, which may inhibit a company's exploitation of service opportunities (Mathieu, 2001; Gregory, 1983).There is a lack of understanding about dividing service culture into multiple and interrelated dimensions (e.g. service value, service behavior, service norms, etc.). In addition, service culture is an embedded organizational structure and may depend on whether the product organization and service organization are isolated or whether they are integrated (Oliva and Kallenberg, 2003). As Oliva and Kallenberg (2003, p. 171) state:It is not clear [...] if the success of the isolated service organizations was due to the additional managerial focus these organizations received, or if, as we expected, cultural and managerial biases are responsible for thwarting the service development efforts in product-centric cultures.Consequently, the interaction among service culture, organizational structure and performance remains fuzzy.This paper develops hypotheses about the relationships between service culture components and overall performance. These hypotheses are tested in two different organizational contexts: the service organization integrated with product organization and the service organization separated from product organization. In more detail, we first determine if the service orientation in corporate culture is comprised of causal relationships among corporate values and behavior across managerial and employee levels. Second, the association between the service orientation of corporate culture and business performance is tested. Third, the associations between corporate values and behavior, and between the service orientation of corporate culture and overall performance, are tested for interaction or moderating effects on the organizational structure of the service business.In the next section, a conceptual framework with the corresponding hypotheses is presented. The third section introduces the research methodology and the results are presented in the fourth section. The last section includes a discussion of managerial implications and further research opportunities.\nConceptual framework and hypotheses development: An organizational (corporate) culture consists of the attitudes, experiences, beliefs, and values of an organization (Hill and Jones, 2001; Hofstede, 1980; Deal and Kennedy, 1982). This culture is a valuable resource that is neither perfectly imitable nor substitutable without great effort (Barney, 1991; Hoopes et al., 2003). Product and service orientation represent two types of organizational cultures and they can both influence the performance of manufacturing companies (Schein, 2004). In general, one is not superior to the other. In the past, product-oriented culture has created competitive advantages in manufacturing companies by favoring an engineering mentality, that is, the value and behavior focus on product-related issues such as product form, features, performance and conformance quality, durability, reliability, style and design. Higher market complexity forced manufacturing companies to be more service-oriented and to focus on service-related issues such as ordering, delivery, installing, customer training, customer consulting as well as maintenance and repair (Bowen et al., 1989; Kotler and Keller, 2009).According to the discussion about manufacturing companies, the transition from selling products to providing services and a more service-oriented culture will lead to above-average performance. Service-oriented culture in manufacturing companies contains the values and behaviors associated with an entrepreneurial orientation, real problem-solving eagerness, innovativeness, and flexibility of service employees (Matthyssens and Vandenbempt, 1998; Bowen et al., 1989). This culture can be separated into two dimensions (Deshpande and Webster, 1989; Hofstede et al., 1990; Homburg et al., 2003). The first dimension refers to the (abstract) value of services as perceived by managers and employees who experience and express the value in their work and how it is carried out. The second dimension concerns the degree to which managers and employees behave in a service-oriented manner. This behavior reflects how much the (abstract) value of services is experienced in the organization (Homburg et al., 2003).Against this background, Figure 1 shows the conceptual framework that guides our research effort. The framework consists of three different components:1. hypothesized main effects among different elements of corporate culture;2. hypothesized main effects of service orientation in the corporate culture on business performance; and3. hypothesized moderating effect of organizational structure on these main effects.The service orientation of corporate culture is conceptualized through the two previously mentioned dimensions. As shown in Figure 1, we differentiate the service orientation of corporate culture into service orientation of values and service orientation of behavior (Homburg et al., 2003). In addition, we divide service orientation in the corporate culture into two levels: management and employee (Huber and Glick, 1993). As a result, the service orientation of corporate culture can be described as consisting of four elements: service orientation of management values, service orientation of employee values, service orientation of management behavior, and service orientation of employee behavior.The first element, service orientation of management values, expresses how the potential benefits of services can influence managers' focus. Managers are an important facet of offering service-related value propositions that support customers' value creation (Vargo and Lusch, 2004, 2008a, b). Management values are also an essential part of the commitment to services and the development of the service business (Donaldson, 1995; Gebauer et al., 2005). Managers are motivated to develop the service business if they perceive a high (abstract) value of services.In addition, it is important to examine the second element of the service orientation of employee values. This element refers to the degree to which employees actively understand the benefits of an extended service business. Oliva and Kallenberg (2003) argue that employees misunderstand the abstract value of services because they often give services (installation, commissioning, etc.) away free of charge during the negotiations to sell the product. Changing corporate culture towards service orientation requires employees who understand the value of services.Our framework next considers the third element of the service orientation of behavior at the management level. This element highlights the relevance of coaching and empowering employees, as well as setting up necessary organizational elements such as human resources, organizational structure, and measurement and rewards for the service business (Neu and Brown, 2005; Johnestone et al., 2008). Managers in the service field traditionally concentrate on recovering from customer complaints, but do not treat service business as a business on its own (Gebauer and Friedli, 2005). If managers want to extend the service business, their behavior has to change from managing customer complaints to leading a service business. They should develop service strategies and emphasize entrepreneurial orientation for achieving service revenue and profit growths (Matthyssens and Vandendempt, 1998). To accomplish the service revenue and profit targets, managers have to coach employees in selling, delivering and billing for services.The forth element, the service orientation of employee behavior, involves the role of front-line employees in complex markets (Neu and Brown, 2005). In order to sell, deliver and bill services, employees must become trusted advisers who develop a learning relationship with customers, solve customer problems collaboratively, and tailor products and services to address individual customer needs (Neu and Brown, 2005; Homburg et al., 2003; Matthyssens and Vandenbempt, 1998).In the second part of the framework, overall performance is based on the profitability achieved by manufacturing firms through providing the whole range of products and services. Typical profitability measures include return on sales, return on investment and return on assets (Chakravarthy, 1986; Homburg et al., 2003).The third part of the framework examines the type of organizational structure. Based on the relevant literature (Oliva and Kallenberg, 2003; Neu and Brown, 2005; Auguste et al., 2006; Baveja et al., 2004; Jacob and Ulaga, 2008), two types of organizational structures are proposed:1. integrating a service-organization in the strategic business units (SBU) for products; and2. setting up a separate service-organization or a separate SBU for services in addition to the product SBUs.An SBU is a relatively autonomous unit with the SBU management having control of at least three of the following functions: marketing, sales, manufacturing, R&D, accounting/finance, and human resources (Homburg et al., 1999).Integrating services into product SBUs means that the services are attached to the product functions and the product SBU takes over service responsibilities. Product sales, therefore, are also responsible for the service sales and the service development process is a sub-process of the product innovation process. Service delivery is organized within the product manufacturing. Product SBUs control the costs attributable to services. The profit- and-loss calculation is made for the total offering of products and services based on total sales, as well as product and service costs (Mathieu, 2001; Auguste et al., 2006; Oliva and Kallenberg, 2003).Separating services from the product business means that firms create a distinct SBU for services that fully control the targeting of customers and the development, pricing, selling, and delivery of service offerings. As a distinctive SBU, the service organization takes over the financial responsibilities for profit-and-loss in the service business (Auguste et al., 2006; Oliva and Kallenberg, 2003).Hypothesis development\nResearch methodology: Data collection and analysis\nResults: Measurement validation\nDiscussion: Theoretical implications\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to examine if there is a relationship between the interdependence among different components of service culture and performance. This paper also contributes to the discussion on how integrating or separating the service organization from the product organization interacts with service culture.\nDesign/methodology/approach: - Data are collected through a survey and structural equation modeling is used for the data analysis.\nFindings: - The three main findings are: identified interactions among different elements of service orientation in corporate culture, the positive association between the service orientation of corporate culture and business performance, and the moderating effects of the type of organizational structure on interactions among service-orientation elements in corporate culture.\nResearch limitations/implications: - There may be alternative explanations for the correlations in the model. There are limitations in survey research in measuring service orientation in service culture.\nPractical implications: - Separating the service and product organization strengthens the creation of a service orientation in the corporate culture, but at the same time, does not affect the link between culture and performance.\nOriginality/value: - The originality of the paper comes from combining service orientation in corporate culture with types of organizational structure. In addition, compared to the dominant single dimension approach to service orientation in corporate culture, four dimensions are used (service orientation of management values, management behavior, employee values, and employee behavior).\n",
"role": "assistant"
}
] |
Cross-border nursing education: questions, qualms and quality assurance
|
[
"Transnational",
"Quality assurance",
"Transnational education",
"Cross-border nursing education",
"Border-less education"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: We aim to identify inherent challenges associated with cross-border nursing education, and illustrate pertinent questions raised in pedagogical practice, when academic leader and student cross-border experiences are explored. From a quality assurance perspective, we suggest that cross-border nursing education programs are still challenged owing to deficient governance structures. Our literature search revealed many publications relevant to international higher education collaboration but cross-border nursing education in higher education, remains an under-researched area. According to Baumann and Blythe (2008), there are no international governing bodies that specifically manage cross-border nursing education programs. Our literature search also noted that there was potential for developing a more robust quality assurance culture with national and international nursing education institutions, stemming from a gap in knowledge and understanding relevant to quality assurance processes, which provide the rationale for this study and allowed researchers to examine actual challenges that arise in cross-border initiatives.\nBackground: \"Cross-border education\" and \"transnational education\" (TNE) are being used interchangeably in quality assurance circles, and are synonymous with \"offshore education,\" \"borderless education\" or \"international collaborative education\" (Knight, 2006). Cross-border nursing education or transnational nursing education includes all higher education nursing programs, or educational offerings in which student nurses are in a country different from the one where the awarding institution is based. Cross-border initiatives are rewarding for many global nursing education institutions. Apart from increasing the institution's profile, these educational programs provide openings for intercontinental business opportunities, consultative activities and collaborative research. Knight (2006) also agreed that cross-border education can be costly with little reward for the provider if not governed by proper quality assurance mechanisms. Fearnside and Chung (2017) found that cross-border education can be defined as study programs or educational services in which students are located in a country different from the one where the awarding institution is based, which involves student, academic, institutional and program mobility across countries and continents. This deems it necessary for collaborators to have clear rationale for their engagement in cross-border ventures, ensuring that they are \"fit-for-purpose\" and in keeping with the requirements and legalities governing international educator and student mobility (World Health Organization, 2013). These findings have undoubtedly prompted more attention to be paid by cross border education providers about the regulation and quality assurance practices. Therefore, raising the questions, shown in the below list, stimulates discussion and establishes connectivity between quality assurance practices and cross-border nursing programs.\nQuality assurance in cross-border programs: Quality assurance is undoubtedly a significant \"player\" in higher education. Countries around the world have different quality assurance systems for their education delivery, as necessitated by their unique national contexts. Quality assurance systems play a vital role in and are vital in any cross-border education program, which ensures that academic criteria such as course material, monitoring and evaluation strategies and staff calibre are at high academic standards. Students involved in cross border programs often categorize quality indicators from an academic and non-academic viewpoint. Academic staff, program quality and university reputation are considered academic factors. Financial assistance, counseling, support services, job placement services are considered non-academic quality indicators (Gamage et al., 2008).\nResearch findings from cross-border nursing education program facilitation and delivery: We used an interpretive research paradigm together with a qualitative multiple case-study approach to identify challenges and best operating practices in cross-border nursing education. The case study approach allowed us to gain insight into participant cross-border experiences and described the challenges that confront staff in educational institutions. The impetus to sample, national and international participants arose from the global TNE challenges we identified our during literature search. The sample therefore, comprised national and international nursing education institutions, nursing graduates and academic leaders (Tables I and II), including international nursing education project directors, nursing faculty and departmental heads and lecturers, which helps to draw a comparison between national and international TNE practices. During the interviews, participants articulated their responses by giving an in-depth account and narrated their experiences surrounding their participation in TNE. Other relevant information and documentation, such as provider manuals, guidelines, policies or procedure manuals, were used to gather data (\"Interview guide for academic leaders,\" \"Interview guide for graduates\" and \"Documents used during the record review process\" for interview and record review guides).\nValidity and trustworthiness: Validity and trustworthiness in this study was achieved by sampling participants from both national and international settings, which maximized the opportunities for gathering data collection. Pre-testing data collection tools was completed prior to the main study to test and refine the proposed tools. Structured questions helped guide purposive sampling, whilst probing type questions posed to the interviewee stimulated \"thick\" actual cross-border education experiences. This process continued until we reached data saturation. Credibility was attained by comparing our findings with the published literature, whilst transferability was promoted by ensuring that the interview process was described accurately. This process ensured that data were collected, interpreted and reflected upon truthfully. Minimal input by the researcher during interviews confirmed that views expressed were the interviewees' and not influenced by the researcher in any way. Documentation in this study was analyzed for descriptions and patterns related to these primary areas:\nSummary: Recruitment and selection\nThis study's contribution to knowledge: Partnerships and collaboration in cross-border education appear to be the way forward in this century. We provide a platform for academics and graduates to lend their voices to promoting effective cross-border delivery. Findings and lessons learnt in this study signify the important role that quality assurance plays in cross-border nursing education programs. These findings are also relevant to all those engaged in international partnerships related to non-nursing cross border programs offered at a tertiary institution. Adherence to quality assurance practices will benefit the tertiary institution by enhancing their reputation and recognizing their effectiveness and capacity. It expands and supports stakeholder knowledge, strengthening teaching and learning quality. The researchers recommend further research from a clinical training specialist perspective related to their experiences with cross-border nursing students. Other research areas stemming from this study are examining institutional best operating practices, such as quality assurance programs, to identify if educational institution staff accept proposed, planned and implemented changes.\nConclusion: This paper, which forms part of a larger study that proposed guidelines for cross-border nursing education programs, concludes that successful and reputable cross-border nursing education programs enhance an institution's prestige especially if the awarding university has rigid quality control measures in place, which means that all programs offered should be in line with the institution's vision, mission and philosophies and the same quality assurance principles must be applicable to its satellite learning centers or learning institutions set up elsewhere. Such principles can mean having suitably qualified lecturers or administrators to co-ordinate and facilitate particular programs and its delivery. Components such as transparency, recognition and accreditation, which we found to be noticeably fragmented and uncoordinated, have an inherent role in assuring TNE quality.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to share insights, research findings and discuss key issues relating to quality practices and quality assurance in cross-border nursing education program development and implementation.\nDesign/methodology/approach: The authors used a qualitative, multiple case-study approach, by sampling local, national and international nursing education institutions, academia and nurse graduates to identify challenges and best operating practices in implementing and facilitating cross-border education.\nFindings: The authors reveal that quality assurance affects cross-border nursing education program design, delivery and implementation.\nResearch limitations/implications: Quality assurance plays an important role in cross-border nursing education, by enhancing the reputation and recognizing the effectiveness and capacity of the educational institution. These findings of this study can offer valuable insight to forthcoming as well as existing nursing education curriculum developers who plan to engage in national or international educational partnerships.\nPractical implications: Quality assurance plays an important role in cross-border nursing education, by enhancing the reputation and recognizing the educational institution's effectiveness and capacity. The findings offer valuable insight into forthcoming and existing nursing education for curriculum developers who plan to engage in national or international educational partnerships.\nOriginality/value: This paper explores inherent challenges in cross-border nursing education and maximized data collection opportunities by sampling participants from both national and international settings.\n",
"role": "assistant"
}
] |
A survey of empirical findings on unconventional central bank policies
|
[
"Unconventional monetary policy",
"Quantitative easing",
"Survey",
"Empirical findings",
"E52",
"E58"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: This study is a survey that builds upon the findings of empirical research about unconventional monetary policy and the impact that such actions by major central banks have brought about in national, global financial markets and real economies. Previous papers such as Ugai (2007), Martin and Milas (2012), Joyce et al. (2012a), Fawley and Neely (2013), Bhattarai and Neely (2016), Bundesbank (2016), Belke and Fahrholz (2018), Thornton (2017), Lombardi et al. (2018) provide reviews of the academic literature and a meta-analysis (Papadamou et al., 2019a) on the economics of extraordinary measures by monetary authorities. Such effects are usually examined in event studies or VAR specifications. Moreover, global spillover impacts to emerging markets are investigated. Although in line to these papers in a respectable degree, our study is the only one to provide every aspect of the prevailing methodological approaches and the full spectrum of empirical findings regarding the burgeoning literature of non-conventional monetary policymaking. Therefore, we add to existing reviews both in terms of amplitude in bibliography as well as concerning the battery of methodologies applied. Furthermore, interactions between financial and macroeconomic variables not yet recorded in the existing surveys, and novel datasets employed in most recent empirical work are taken into consideration. This leads to wider examination of global effects in this paper. Moreover, theoretical frameworks built upon innovative models are included in our study. It should be noted though that the task of reviewing every single paper about unconventional monetary practices is too demanding due to the astonishingly large spectrum of relevant papers. This is the reason why, in the conduct of the survey, there is the risk to leave out important studies and important facets of already included studies. The aim of this study is to cover the most important papers in the most valuable axes of research on non-conventional monetary action taking.Unconventional practices have been employed by monetary authorities when the interest rate has reached extremely low levels, very close to the Zero Lower Bound (ZLB). Due to inability of central bankers for agile movements, portfolio enlargement and rebalancing of the central bank's balance sheet have been considered as the final arrow in the quiver of major central banks such as the Federal Reserve (Fed), the Bank of England (BoE), the European Central Bank (ECB), and the Bank of Japan (BoJ) since the Lehman Brothers collapse. Interestingly, the primary aim of the Fed was to spur the US economy by using large-scale asset purchases (LSAPs), as well as by forward guidance concerning future short-term rates. The Asset purchase Facility (APF) by the BoE resembles the Fed's action taking, whereas the ECB and the BoJ proved to be keener on direct bank lending and creation of fiat money for stimulating GDP. Overall, it is found that non-conventional practices by monetary authorities reduce long-term interest rates by influencing both expectations and risk premium. The latter is responsible for the portfolio rebalancing channel, which is found to be most influential on long-term bonds. Notably, the new monetary conditions have brought up to the surface new methods of measuring the monetary policy stance. Shadow rates for the US economy have been employed, such as in Krippner (2012) and the \"new shadow policy rate\" by Lombardi and Zhu (2018) which is driven by alterations in term premia at various maturities and asset purchases by the Fed.Unconventional monetary action taking is highly important in terms of spillovers from advanced economies towards emerging and developing ones. This is the reason why the latter such as Brazil, Colombia, India, Indonesia, Peru, Philippines, Russia, South Africa, Thailand and Turkey have implemented restrictions on capital flows. It should be emphasized that even more countries have implemented prudential requirements. Surprisingly, there is evidence that macro-prudential policy implemented due to non-conventional monetary practices rather attracts than prevents capital flows. This happens because it creates higher levels of trust regarding domestic institutions and leads to forming expectations that a possible future crisis will be effectively confronted. Thereby, it serves as a proxy of good institutional quality mainly in Asian economies (Belke and Volz, 2019).The main routes through which unconventional monetary policy-making functions are: the portfolio rebalancing, signaling, liquidity, confidence and bank-lending channels that lower long-term yields and promote investment. Specifically, portfolio rebalancing acts via the \"credit easing\" policy due to the imperfect character of asset substitutability and it operates through local supply effects that trigger alterations in the composition of portfolios. Notably, the signaling channel works through expectations about short-term rates. Moreover, the liquidity channel, by increasing the liquidity available to investors, works through reductions in the liquidity premium. The confidence channel acts by boosting confidence in economic perspectives by improvement in economic conditions and aggregate demand. It should be noted that the bank lending channel promotes loans through lower lending rates.The main forms by which unconventional practices appear are: (1) forward guidance in an attempt to reduce expectations of future short rates, and (2) large-scale asset purchases, also called as \"quantitative easing,\" mainly made by Mortgage-Backed Securities (MBS) and Treasury purchases. The monetary authorities try to influence long-term interest rates not by affecting the monetary base, but rather by their policies' impact on long-term yields through the portfolio rebalancing, signaling, and liquidity channels. The portfolio rebalancing channel mainly has an influence on the term premium, which indicates how much the long-term yield is higher than the expected overnight yield. It is composed by the duration risk premium, the liquidity premium, the default risk premium, the safety premium, the local scarcity premium, and the prepayment risk premium, and it seldom takes negative values. On the other hand, the signaling channel aims at reducing the expectations part of long-term interest rates and works mainly through central bank announcements and consistency of its policy, thereby demands a high level of credibility concerning monetary authorities.Recently, there is a growing literature focusing on another unconventional monetary policy measure, the negative interest rate policies (NIRPs) and their effect on bank lending, bank profitability, exchange rates and financial market variables (for a survey on the implementation of NIRP across central banks see on (Bech and Malkhozov, 2016). In the relevant literature of the NIRP effects on the economy there is a debate between two groups of researchers. On the one hand, authors like, Blanke and Krogstrup (2016), Jobst and Lin (2016) argue that NIRPs positively affect the economy by lowering funding costs and raising asset prices. On the other hand, Borio et al. (2017) and Caruana (2016) pointing out potential adverse effects of low and negative rates on the profitability of financial institutions and financial stability [1].Since October 15, 2008, when the Lehman Brothers defaulted, a respectable and proliferating volume of literature was nourished by needs of unconventional practices regarding monetary policymaking. The main aim of this study is to summarize in the most complete and easily perceivable way how such measures were effective in each country or not, though we do not take into consideration earlier episodes of QE in Japan during the 2000-2006 period. Empirical evidence is provided from all available channels. For these purposes, our investigation is centered on empirical findings of each country's unconventional actions in order to cast light on every positive or negative aspect of such policies.The remainder of this paper is structured as follows. Section 2 offers the empirical findings of relevant studies, categorized by their transmission channels. Section 3 shows empirical evidence about non-conventional practices by being based on their intermediate targets, while in Section 4 studies about the final targets are presented. Finally, Section 5 concludes.\n2. Empirical evidence on unconventional monetary policy channels: 2.1 An overview of studies about unconventional monetary practices\n3. Asset prices and foreign exchanges rates as intermediate targets in unconventional monetary policy channels: 3.1 Bond yields\n4. GDP, inflation and unemployment as final targets in unconventional monetary policy channels: 4.1 Unconventional effects on GDP, inflation and unemployment in New-Keynesian models\n5. Conclusions: In this paper, we have provided an up-to-date survey of the main empirical research aimed at studying the links between unconventional monetary policies. From an empirical perspective, we have seen that event studies, VAR methodologies, and panel estimations have been the main tools for assessing the impacts of central banks' non-conventional practices on the financial sector and the real economy.A number of policy lessons emerge from our empirical findings. First, the majority of studies provide evidence of liquidity provisions having modest effects on output and inflation, which are stronger during the first non-conventional rounds. Second, the primary channel through which unconventional policies are transmitted is the portfolio rebalancing mechanism, followed by the expectations and liquidity channels in many occasions. Lower long-term bond yields have been the main achievement of non-ordinary policies, while indications of effects on exchange rates and stock prices, as well as bank lending rates are also significant in many studies. Third, US unconventional policies are found to exert significant spillover effects on emerging markets economies, mainly by providing higher outflows towards EMEs, lowering yields, appreciating their currencies and spurring stock prices. Furthermore, the ECB's unconventional action taking has led to impacts on non-Euro adopting European countries, such as decreasing the sovereign risk premium and bank lending rates. QE policies may lead to higher inequality during crisis in the US, the UK and the Euro area.It is against this background that we review the recent breed of unconventional policies as a promising area to understand the mechanisms by which balance sheet enlargement of monetary authorities may affect the economies of advanced and of emerging economies. This investigation is about periods of high stress in economic and financial conditions. Our results have important implications for academics, investors and policymakers. They know the main active channels via which unconventional policies are transmitted in real economy by highlighting weak points in these transmissions. Special attention should be given on forward guidance policies and spillover effects across global markets. Concerning the latter, EMEs seem to be particularly sensitive to US QE policies. We also identify the risk facing the implementation of a QE happening.This highly promising strand of monetary policy literature is what, in our thinking, renders this survey a promissory route for a more thorough understanding on the connection between liquidity provisions and the real economy. Moreover, it provides a roadmap for any further research and leads to interesting extra fundamental insights regarding modern monetary policymaking.\n",
"role": "user"
},
{
"content": "Purpose: This paper presents an integrated overview of the empirical literature on the impact of all forms of unconventional monetary policy on macroeconomic variables and on markets.\nDesign/methodology/approach: This survey covers the findings concerning portfolio rebalancing, signaling, liquidity, bank lending and confidence channels.\nFindings: The positive effect of QE announcements on stock and bond prices seems to be unified across studies. A contagion effect from US QE to other emerging markets is identified, while currency devaluation is present in most cases for the country that its central bank adopted such policies. Moreover, impacts of non-conventional practices on GDP, inflation and unemployment are examined. The studies presenting weak instead of strong positive effects on inflation are more, and these studies, also, present weak positive effects on GDP growth.\nOriginality/value: Based on the large body of research on non-conventional action taking, this is the first survey including effects of each country that adopted quantitative easing (QE) measures and that provides results from every methodology employed in order to estimate unconventional practices' impacts.\n",
"role": "assistant"
}
] |
Corporate governance changes, firm strategy and compensation mechanisms in a privatization context
|
[
"Corporate governance",
"Strategy",
"Board of directors",
"Compensation",
"Privatization"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Evidence over the last few years shows that privatization is a popular strategy for promoting national economic development as well as the center of attention in research from a macroeconomic or political perspective (Zahra et al., 2000). In contrast, the organizational and managerial implications of privatization need further examination (Zahra et al., 2000). Because privatization implies changes in corporate governance (or firm's ownership), owners' objectives should both translate into firm strategy and influence internal firm's management (Cuervo and Villalonga, 2000).\n2. Theoretical framework: Agency theory is arguably the dominant theoretical framework in corporate governance research (Lynall et al., 2003). According to this theory, the two mechanisms that align the interests of owners with those of managers are the monitoring function of the board of directors and incentive alignment mechanisms, such as pay-for-performance schemes (Eisenhardt, 1989; Fama and Jensen, 1983; Jensen and Meckling, 1976).\n3. Research methodology: To analyze the effects of corporate governance and strategy on compensation practices, we employ a multi-case research method. Table I details a description of the five companies under analysis. The five case companies have been labeled Alpha, Beta, Gamma, Delta and Sigma.\n4. Findings: Below, the examination of the firm governance structure, strategic orientation and compensation systems follows. Data from the five case studies allow comparing these variables before and after privatization. Finally, we present an analysis of the sectors to which the companies studied belong. This will allow us a better understanding of changes in the companies after privatization.\n5. Conclusions: This study analyzes how changes in a firm's ownership after privatization can have an impact on the owners' objectives, the firm's strategy and other aspects in relation to managerial compensation systems. Based on agency theory, our main assumption is that after privatization new owners would monitor the actions of management and use incentive alignment to ensure goal congruence between principals and agents (Eisenhardt, 1989; Jensen and Meckling, 1976).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze certain organizational changes that occur after privatization. More specifically, the authors analyze from an agency theory perspective how changes at the corporate governance level (ownership type/structure and board of directors) influence firm strategy, and what implications these issues have for the design of compensation mechanisms.\nDesign/methodology/approach: The methodology employed in the study can be described as a longitudinal multiple case study approach. A multiple case study enables the comparison of pre- and post-privatization periods in five Spanish companies. This methodology is especially suitable to track the organizational changes occurring in relation to the firms' internal management.\nFindings: First, the variables that traditionally relate to greater board independence in monitoring do not suffer from variation after privatization. Second, the interests of the firms' new ownership have an impact on firm strategy after privatization. Finally, compensation system design clearly aligns with firm strategy after privatization.\nResearch limitations/implications: The research is based on a multiple case study approach, which limits the scope and generalizability of the findings.\nOriginality/value: Whereas research in privatization generally adopts a macroeconomic or political perspective, organizational and managerial implications are current aspects in need of further examination. This research offers a study that integrates three sets of variables (corporate governance, strategy and compensation) that have been rarely analyzed in this context.\n",
"role": "assistant"
}
] |
The pervasive power of bespoke internal branding: Assessing the impact of employees' generational differences and organizational tenure
|
[
"Internal branding",
"Employee engagement",
"Generation X",
"Millennials",
"Brand loyalty",
"Brand citizenship"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Employees play an often uncelebrated but crucial part in the success of an organization's brand strategy, since most companies focus overly on external customers at the cost of eroding employee engagement levels. A research paper by Dechawatanapaisal, D. (2019) concentrates on uncovering the effect of an employee's brand identification on the relationship between internal branding, brand loyalty, and brand citizenship behavior. Another specific focus is on how the successful impacts of these brand phenomena are affected by variations in an employee's organizational tenure and working generation. Data for the study were gathered from the survey responses of 702 salespeople - from both Generation X (born between 1965 and 1980) and the millennial (born between 1981 and 1999) generation - who work in Thailand's construction material and cement industries.\nInterrogating the mechanisms of internal branding: A brand distinctly and identifiably differentiates a company's products and services from those of its competitors, and also acts to build a potential customer's desire for its offering. Internal branding is the practice of seeing employees as customers, while generating within them a consistently felt sense of belonging and understanding of the brand's values in order to create a competitive edge. Thereby internal branding's objective is to cultivate engaged, knowledgeable brand ambassadors or citizens, for example salespeople, who consistently celebrate the brand to others with emotional meaning, which in turn boosts the external brand. Crucially, good internal branding should translate into a company being able to retain talented employees for a longer than average tenure.\nExploring the results around branding: Through the study's survey questions, a number of the author's hypotheses were tested to reveal the following results, that:\nActions for powering an employer brand ahead: The study's results generated the following advice for managers wishing to optimize their organization's overall internal brand-related impacts. Firstly, that they should keep in the front of their minds how the effect of an internal brand varies according to organizational tenure and an employee's generational traits. As such managers are advised to gain the support of their human resources (HR) teams in shaping policies that are specifically targeted to appeal to the preferred values and work environment priorities of millennial age employees, since this dominant group of the working population tend to have the looser attachment to the organization they work for. At the same time Generation X priorities should also be incorporated into HR policies, such as skill development opportunities through which all employees can beneficially experience personal growth. Secondly, managers can add further value by teaching their team members more about the organization's brand, with the aim of elevating those team members' emotional attachment to it. Finally, managers should be aware that employees with longer tenure become immune to the effects of internal branding over time, meaning more effort in this area is needed to keep them actively engaged with the organization.\nCommentary: The review is based on \"Internal branding and employees' brand outcomes: do generational differences and organizational tenure matter?\" by Dechawatanapaisal, D. (2019), published in \"Industrial and Commercial Training.\" This research paper confirms that tailored internal branding activity drives the brand identification of employees, since employees respond to internal branding efforts differently depending on their personal characteristics. The results specifically reveal that a shorter organizational tenure translates to internal branding having a greater impact, particularly in the case of Generation X employees. Furthermore, Generation X employees were found to have a more elevated sense of brand identification and loyalty to their employing organization than the millennial age employee, since the latter tends to manage their career path more dispassionately like a financial asset.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to review the latest management developments across the globe and pinpoint practical implications from cutting-edge research and case studies.\nDesign/methodology/approach: This briefing is prepared by an independent writer who adds their own impartial comments and places the articles in context.\nFindings: This research paper confirms that tailored internal branding activity drives the brand identification of employees, since employees respond to internal branding efforts differently depending on their personal characteristics. The results specifically reveal that a shorter organizational tenure translates to internal branding having a greater impact, particularly in the case of Generation X employees. Furthermore, Generation X employees were found to have a more elevated sense of brand identification and loyalty to their employing organization than the millennial age employee, since the latter tends to manage their career path more dispassionately like a financial asset.\nOriginality/value: The briefing saves busy executives and researchers hours of reading time by selecting only the very best, most pertinent information and presenting it in a condensed and easy-to-digest format.\n",
"role": "assistant"
}
] |
Corporate governance integration with sustainability: a systematic literature review
|
[
"Leadership",
"Sustainability",
"Mission",
"Corporate governance",
"Vision",
"Framework"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In the twentieth century, corporate sustainability is fast becoming crucial toward mainstream corporate activity. The skills, knowledge and sophistication involved in leading corporate sustainability is also becoming an important part of the management decision-making process (Windolph et al., 2014), accounting practices (Gray, 2010; Schaltegger et al., 2006) and reporting practices (Guidry and Patten, 2010), both in corporations and the public sector (Adams et al., 2014).\n2. Theoretical background: The classic approach to making profits is dictated by inclusion of the stakeholder theory. The theoretical perspective on sustainability combines emerging concepts from two unrelated political science and management theories to create a new governance structure, such as emerging sustainability issues which entail accountability and transparency from collaborations (Clarke and Islam, 2004). Corporate governance, in the context of organizations, works with all of the stakeholders that are interdependent. The dependence of the interventionist state on administrative and planning systems limits the communal sense which is unsuitable for complex decision-making involving social and environmental risks and opportunities (Clarke and Islam, 2004).\n3. Methodology: The proposition that was accounted for in the subsequent literature review of this work is how corporate governance is integrated into sustainability in terms of framework, model, relations, process and strategy to organizations. Quantitative and qualitative papers on Scopus was reviewed using a systematic approach. Articles with key words in its titles or abstract were selected. This was done in April 2017 for articles published in 1999-2017, accounting for the period when international social reporting, such as Global Reporting Initiatives (GRI, 2016), was launched.\n4. Results: 4.1 Quantitative analysis of selected literature\n5. Discussion and conclusion: This article explored literature on the incorporation of corporate governance into sustainability based on stakeholder theory. Stakeholder theory gives a framework connecting corporate governance and sustainability disclosure (Huse, 2003), by recommending that each improves stakeholder engagement and in this manner organizational legitimacy (Unerman and Bennett, 2004). As showed by Deegan (2007), organizations attempt to pick up legitimacy by revealing social and environmental unquestionable information and data.\n6. Implication of practice: As a result of different aspects/research subjects of corporate governance expression on sustainability framework with time, researchers should focus on process models related to enacting sustainability and to achieve sustainability actions which provide scholars rich opportunities for theoretical advancement of sustainability and corporate governance, especially to jointly create work that explores origins and outcomes for sustainability models through corporate governance. In fact, there is no single best form of governance for sustainability. The details must vary, improving the specifics of the framework from claim to claim. This study could develop a deep understanding of sustainability models which would better inform implementations, strategies and support governmental planning structures.\n7. Limitation and future research developments: This research is limited by the common features of Scopus search, such as the choice of number and type of keywords, the resulting selection of studies, the choice of relevant outcomes the interpretation and the generalization and application of results. The literature review only included peer reviewed articles, excluding items such as books and gray literature. This is partly mitigated when the authors write a book on the same topic. Also, other synonymous articles could have ended up altering the results.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to systematically review the literature on corporate governance and sustainability integration in identifying the main rigidity, infirmity and gaps in the current literature, and also to mention future research paths.\nDesign/methodology/approach: A systematic literature review of existing international papers is used through quantitative and qualitative approach by selecting 27 articles published in Scopus.\nFindings: The review suggests although integration of governance into sustainability is interpreted differently in a geographical area, vision, mission and leadership are the most significant drivers of sustainability framework dealing corporate governance. Despite the limitation which is related to the choice of number and type of keywords and journals, outcomes and the interpretation, generalization and application of results, sustainability frameworks suggest a number of avenues for investors, policy makers and future market scenario which will increase the efficiency of companies.\nResearch limitations/implications: This research uses limited number of reviews by the common features of Scopus search as previous studies. This review study reflects corporate governance to sustainability models and provides opportunities to researchers for a more in-depth investigation into the theoretical advancement and joint work of sustainability and corporate governance which better inform strategies and implementations of governmental structures.\nOriginality/value: This paper undertakes a significant thorough systematic review for sustainability integration with corporate governance literature. It gives a written work review and reference index from1995 to 2017, useful for both academics and professionals.\n",
"role": "assistant"
}
] |
Performance measurement and management research in the hospitality and tourism industry
|
[
"Performance measurement",
"Performance management",
"Tourism",
"Critical review",
"Hospitality"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: All management perspectives, philosophies and theories (general, management, behavioral approach, total quality management, strategic management, etc.) emphasize how performance measurement (PM) and performance management is an essential part of the success of the cycle of management, or business processes. Hence, to measure what you achieve for given goals and tasks in an organization, these approaches provide their own performance management systems, or managers build their own PM and performance management frameworks to improve their systems and/or success. Few studies have been done evaluating the evolution of PM and performance management systems from a general management perspective (Bititci et al., 2012; Choong, 2014) by focusing on developments or changes in manufacturing industries. However, studies focusing on how the PM literature in the service industry has evolved are needed to help researchers, managers and graduate students to adapt PM that could follow changes in the industry.\n2. Literature review: 2.1 Overview of the performance measurement literature\n3. Methodology: 3.1 Sample\n4. State of the art: 4.1 Progress on ontological and epistemological issues\n5. Observations, gaps and future research: This study is an attempt to discuss PMs and performance management in the H/T industry. As one of the first studies in the field, this study has specific implications discussed below based on the research questions.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to provide a critical review of the performance management literature in the hospitality and tourism (H/T) industry.\nDesign/methodology/approach: The study is conducted through a critical literature review based on three dimensions: progress on ontological and epistemological issues, on the purpose of performance measurement (PM) and on the emerging contexts.\nFindings: The progress on the three topics, namely, ontological and epistemological, purpose and emerging context, is addressed. Based on these three issues, the performance management literature in the H/T industry has not had any concrete structure. Few studies have been conducted reviewing the progress on performance criteria in the hotel industry. A research agenda is recommended to build a concrete structure in the field.\nResearch limitations/implications: This study focused on only leading H/T journals to analyze the progress on the performance management literature in the H/T industry. The findings can help researchers (re)design research agendas to contribute to both mainstream performance management literature and literature related to the H/T industry.\nOriginality/value: This is one of the first studies highlighting the progress on the performance management literature in hospitality and tourism by considering ontological and epistemological issues, the purpose of PM and the emerging contexts.\n",
"role": "assistant"
}
] |
Customer satisfaction in private healthcare: Using lean bundles to improve performance
|
[
"Management",
"Performance",
"Health care",
"Lean bundles",
"Human resource management",
"Customer satisfaction"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In an ever-competitive world, businesses need strategies to enable them to succeed, combining better and more efficient operations, cutting costs, reducing waste (in terms of time and energy as well as supplies and equipment) and providing quality services to satisfied customers and stakeholders.\nLean bundles, operational performance and business performance: TQM can be thought of as hard practices, such as technical techniques or tools which lead to continuous improvement of operations. It involves a high degree of process management. Improved performance leads to reduced variability and error rates, and thus leads to increased patient satisfaction.\nInvestigating lean bundles in private health care: It can be predicted that the lean bundles each positively affect the individual performances of efficiency, quality and accessibility, and these in turn each positively affect the overall business performance. Alkhaldi and Abdallah (2010) investigated these hypotheses in 25 private hospitals in Jordan, using 260 completed questionnaires. They found that TQM led to increased quality performance (but not efficiency or accessibility), HRM increased all OP indicators, JIT increased efficiency and accessibility (but not quality), TPM supported quality and accessibility (but not efficiency). Out of the operational performance indicators, quality and accessibility led to increased business performance, but efficiency did not lead to increased business performance in Jordanian private hospitals.\nLean bundles for management: For hospital managers wanting to increase competitiveness and their market share, it is important to have satisfied and loyal customers, and a good reputation. Quality and reliable and accessible services are operational outputs that lead to increased business performance. It is important to ensure several lean bundles are used interrelatedly, in order to avoid overlooking a particular aspect. The HRM bundle is the most important in health care due to its human nature. TQM is important for continual monitoring and improvement of operations. JIT and TPM are needed to ensure adequate and sufficient availability of resources and equipment in good working order. Lean management is a good strategy to use in healthcare ensuring high-quality procedures, trained and knowledgeable staff, reduced waste and optimal inventory management, and good equipment maintenance, resulting in satisfactorily meeting the individual patient's needs, and putting them first.\nComment: The review is based on 'Lean management and operational performance in health care: Implications for business performance in private hospitals' by Alkhaldi and Abdallah (2010), published in International Journal of Productivity and Performance Management.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to review the latest management developments across the globe and pinpoint practical implications from cutting-edge research and case studies.\nDesign/methodology/approach: This briefing is prepared by an independent writer who adds their own impartial comments and places the articles in context.\nFindings: Competitive business management requires increasing performance whilst reducing waste and resource costs. Lean management, and the use of interrelated and comprehensive lean bundles (such as total quality management, human resource management, just-in-time system and total productive management) can lead to improved operational performance (efficiency, quality, accessibility) which in turn leads to increased business performance. It is found that in private health care settings, human resources and quality operations are key to improved operational performance; quality of service and accessibility for patients are important to increasing business performance.\nOriginality/value: The briefing saves busy executives and researchers hours of reading time by selecting only the very best, most pertinent information and presenting it in a condensed and easy-to-digest format.\n",
"role": "assistant"
}
] |
Does activism matter? The case of environmental policy in the US states
|
[
"Social capital",
"Institutions and institutional analysis",
"Environmental economics",
"Public goods"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Collective action, the \"relevant\" median voter, and environmental policy: Under certain assumptions, the median voter as the decisive voter is the simplest decision-making model under majority rule capable of explaining economic policy outcomes in a democracy. This paper examines the impact of median voter preferences on policy choices conditional upon the existing institutions of political participation, including collective voter engagement, in democratic societies. We specifically examine the role of activism, a form of collective action, in determining the identity of the \"green\" median voter, and the resulting impact of such voter involvement on the environmental policy outcome of choice: regulatory stringency in the US states measured by pollution abatement spending by manufacturing facilities. Our objective is to show that median preferences of the green voter, qualified by institutional choices and the quality of such institutions, are represented in states' environmental policy choices. The study adds to the existing literature by incorporating three alternative measures of environmental activism providing broader interpretation of environmental activism which are then employed to examine the impact of activism on state-level environmental policy choices.\n2. Activism measures and other data: A panel data set covering six years from 1989 to 1994 and consisting of 288 observations is compiled to test our theory. Continuous and consistent data availability drives our choice of the study period, a period marked by significant developments in environmental policy making at the federal level. The activism variables are measured in two forms; two-year lags covering years 1987-1993 and four year lags covering years 1985-1993. The control variables are concurrent with the dependent variables. Table I summarizes the variables used in regressions.\n3. Empirical methodology: We use FE panel data methodology for hypothesis testing. While FE is not necessarily the best estimator in a setting where a number of slow changing institutional variables are included in regressions, it is nonetheless the most appropriate in our case. There certainly could be omitted time invariant institutional and cultural factors at the state and regional levels that are correlated with activism and remaining regressors that are better controlled within a FE regression design. Another important reason for choosing FE is that the documented negative trend in civic engagement during the short study period can be better controlled with time FE allowing estimation of the pure correlation between environmental activism and stringency. However, one of the drawbacks of the FE method is that results do not have external validity and apply only to the sample being used.\n4. Results: FE results are reported in Tables II and IV while the corresponding b coefficients are reported in Tables III and V. All model specifications include time dummies; coefficient values for these dummy variables are excluded from results tables to preserve brevity. Results for control variables are also omitted in Table IV for the same purpose.\n5. Concluding remarks: The results of our empirical models are decidedly mixed, which may partly be driven by drawbacks in data availability and our activism measures as well as issues with the PACE data.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the role of activism in determining the identity of the \"green\" median voter and the influence of the green voter on environmental regulatory stringency in the US states.\nDesign/methodology/approach: Regulatory stringency is measured using output weighted abatement spending and an industry concentration adjusted index of state environmental compliance costs for the period 1989-1994. Activism measures include environmental initiatives, median support for pollution standards and voter ideology. Fixed-effects panel methodology is used in empirical estimation.\nFindings: The authors find that activism increases stringency in regulating overall as well as media-specific pollution. The results particularly highlight the nuances of different approaches to activism and their varied impact across pollution media.\nResearch limitations/implications: A drawback of the empirical estimation is the lack of continuous historical abatement spending data. A longer panel with alternative stringency measure(s) would add explanatory power to activism, especially since some activism measures capture slow-changing institutional factors.\nSocial implications: The study identifies the conditions under which activism can have the most impact on a society's environmental outcomes since pollution varies in damages, hence abatement costs, across pollution media.\nOriginality/value: The paper adds to the existing literature by incorporating three alternative measures of environmental activism to systematically investigate its impact on environmental stringency within a fixed-effects regression design. It also promotes a deeper understanding of the efficacy of the activism process by deconstructing policy stringency across pollution media to show that activism and its impacts are more nuanced than previously studied.\n",
"role": "assistant"
}
] |
Selection of appropriate electronic banking channel alternative: Critical analysis using analytical hierarchy process
|
[
"Analytical hierarchy process",
"Services",
"Internet banking",
"Adoption",
"Electronic banking channel"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nI. Introduction: It has been witnessed in the last few decades; rapid development of technology played an important role in the growth of business as compared to previous time. The service industry is not untouched with the technological revolution that also helped in giving new directions of delivering services. Technological developments have witnessed enormous changes in the banking sector, through which customer can avail and perform financial transactions without visiting the bank. The evolution of information and communication technology (ICT) facilitated for companies in the global ever changing market where customers are more demanding. It has been found out by many researchers, experts and pioneers in this field that the service quality can be improved through advancements in ICT. In order to increase customer satisfaction, many banks have implemented ICTs, as internet banking (IB) facilitates customers with quick fund transfer and 24x7 services.\nII. Conceptual background and literature gap: The emergence of technology in the banking sector has given birth to various distribution channels like IB, mobile banking and ATM; thus helping the banking customers to avail services at any time and at any place. Banking customers use these channels to gain in terms of time and cost. Even though the types of service one can avail from these channels are comparable, still the benefaction among them is different.\nIII. AHP: In the early 1970s the AHP approach was developed for contingency planning and allocation of scarce resources in the military. AHP is a mathematical structured method introduced by Saaty (1977) and used as a solution for complex MCDM problems.\nIV. Methodology and applying AHP in EBC selection: In this study, we tried to evaluate the priority of EBCs alternative, where four criteria were present that affect selection of channels. We had MCDM problem we used AHP for selection of best EBC. We could also use regression analysis with AHP, if we want to categorize criteria. Priya and Venkatesh (2012) used to regression for determining decision criteria for selecting best locations for expansion of market. With support of literature work in this field, we have criteria and sub-criteria, then we are only focussed on AHP application.\nV. Results and discussion: From Table II, it has been analyzed that the factors, age and occupation have more priority than other demographic factors. The reason behind it was that a specific age group has more inclination toward using electronic banking. According to respondents, occupation is the second most important factor in selecting EBC, because the different job profile gives variation in environment, time and facilities which might create need and wish to use EBC. For example IT manager and banking professional have more probability of using EBC, whereas probability is very low in case of small-level self-businessman or persons working as drivers, carpenters, etc. On the other hand salary and gender were not found to be of much importance. As ATM has more penetration in India, it has been given more priority compared to other EBC with reference to factor age, gender and salary. While considering occupation mobile banking has highest priority because its base is smartphones with supporting software.\nVI. Conclusion: The paper discusses the four major criteria involved in the selection of best EBCs namely demographics, technological factors, service components and intention to use. As far as demographics are concerned, our result explains age and occupation contributes more than gender and salary. From Table II, it can be seen that the priority of factor age is highest followed by occupation.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the important criteria responsible for the adoption of different electronic banking channel (EBC) alternatives in Indian perspective. It also evaluates the intensity of the various criteria, namely demographics, technological factors, service components and intention to use in determining the overall priorities of EBC.\nDesign/methodology/approach: Data were collected from 110 respondents using pairwise comparison-based survey instrument. The collective data was analyzed using analytical hierarchy process and priorities of different EBC were computed.\nFindings: ATM was found out to be the most preferred EBC alternative, followed by internet banking (IB) and mobile banking. Awareness, self-efficacy, ease of use, usefulness and security were the criteria responsible for making ATM, the most preferred channel.\nPractical implications: This paper imparts a better understanding of customers' preferences among different EBC alternatives. It also explores the relationship between different criteria and sub-criteria based on the data collected from both the parties involved in electronic banking, i.e. customers and bankers. This research would be helpful to guide the strategy maker of electronic banking, so that penetration rate may improve in India. This would also give an idea which factor should be given more emphases, while offering new electronic banking services. Considering same method, further detailed research on specific EBC is required. Similarly, studies of different demographic area may consider for robustness of concept.\nOriginality/value: In the past research work, many researchers talked about the specific EBC and now it is needed to find the gap between all channels selection attitude, considering same parameters. Through pairwise comparison, priorities of different criteria and sub-criteria have been evaluated. This also leads to the selection of most preferred EBC. The results would also be useful in enhancing the customer base, thus reducing the barriers in the adoption of IB and mobile banking.\n",
"role": "assistant"
}
] |
Automatic identification of structural process weaknesses in flow chart diagrams
|
[
"Business process reengineering",
"Automatic business process analysis",
"Structural business process analysis",
"Process weaknesses",
"Banks"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Process models have been established as a broadly applied instrument in business process management. Therefore, researchers have developed many modeling languages for the formal representation of business processes since the arrival of the first business information systems (Dumas et al., 2005; Zairi, 1997). Popular examples range from Petri nets (Petri, 1962 for the original source, for a detailed explanation see Petri and Reisig, 2008) over event-driven process chains (Keller et al., 1992; Scheer, 2000) and the UML activity diagram (Object Management Group, 2004) to the business process modeling notation (Object Management Group, 2006). With the help of these modeling languages it is possible to construct a formal representation of real world processes. These models allow a documentation and communication of as-is business processes as well as to-be definitions of future business processes in software development and business process reorganization projects (Hammer and Champy, 1992). They describe the logical sequence of activities, the resulting products and services, the required resources and data, as well as the involved organizational units (Lindsay et al., 2003). These process models can be used, e.g. as a basis for decisions on IT investments, reorganizations or the selection and implementation of information systems.Languages for representing business processes try to avoid the fuzziness of natural language descriptions by semi-formal or formal process representations. However, the inherent impracticability of mathematical formulations is represented in semi-formal, graphic forms of representation (Thomas and Fellmann, 2007). Fundamental work is done in the field of graph theory (Gori et al., 2005). Based on a given graph, these approaches discuss the identification of structurally equivalent (homomorphism) or synonymous (isomorphism) parts of the given graph in other graphs. However, with a semi-formal specification of business process models (e.g. with the help of the event-driven process chains) an automated model analysis of model elements and models is very difficult in terms of semantic similarity. It may be possible to identify patterns in models on a syntactical level in order to analyze the occurrence of a particular collection of model elements (e.g. the number of media breaks as an indicator for the introduction of a document management system) (Pfeiffer, 2008). The automated analysis of business process models would allow significant cost saving potential in contrary to manual evaluations. However, current broadly distributed, commercial modeling tools provide only limited support for the automation of analyses (Blechar, 2007; van der Aalst et al., 2003). As a result, researchers come to the conclusion that banks do not fully exploit the potential of process analysis compared to the conducted effort (Becker et al., 2010a, b).Business process analysis is a highly relevant area in business process management research (Mayer et al., 1998). van der Aalst et al. (2003) see business process analysis as an \"emerging area\" as research still indicates problems in conducting automatic analyses (Becker et al., 2009c; Yu and Wright, 1997). Formal analysis techniques can deliver important support during business process reengineering (BPR) efforts (van Hee and Reijers, 2000) but also for benchmarking. Owing to the size of process models and their complexity, companies strive for a solution that allows an automatic business process analysis (Desel and Erwin, 2000). The value of process modeling can only be uncovered when time-consuming analysis is performed at best automatically in various ways. Therefore, we propose an approach for the automatic analysis and detection of potential weaknesses in structurally analyzable business process models. In systems analysis and design, so-called design patterns are used to describe best practice solutions for common recurring problems. Common design situations are identified, which can be modeled in various ways. The most desirable solution is identified as a pattern and recommended for further usage. The general idea originates from (Alexander, 1977), who argued about patterns in the field of architecture. In IS, patterns are commonly used in system design or workflow modeling. However, in most cases patterns are not used for matching but for the manual implementation of best practices (for a detailed discussion cf. Becker et al., 2009b). Hence, the underlying research question of this article is:How can flow chart patterns be defined that allow for the automatic identification of potential structural process weaknesses and for process path benchmarking?In order to achieve this research aim, we undertook a comprehensive case analysis. As the need for extensively analyzing business processes for multiple purposes is currently of major relevance in the banking sector (Harmon and Wolf, 2008; Drake et al., 2009), we selected a case from the banking sector in order to evaluate our structural weakness patterns. We base our findings on the flow chart notation that is a very simple and mature process modeling language. However, our findings are neither limited to the modeling language nor to the finance sector.The paper is structured as follows: Section 2 builds the theoretical foundation of our work. Subsequently, the discussion of the research methodology applied takes into account issues of method selection, case selection, and data collection and analysis. Following the development and demonstration discussion in Sections 4 and 5, we seek to reflect on the implications for theory and try to open up new vistas for BPM practice in terms of process improvement and benchmarking. The final section contains conclusions.\n2 Theoretical foundation: Currently business process models are mainly analyzed manually (Vergidis et al., 2007). Especially in smaller organizations, the methodical knowledge of how to collect data about the business processes and how to benchmark process models is often not available (Benamou, 2005). Therefore, external consultants are hired to construct and evaluate models (Davenport and Short, 1990; Rosemann, 2007). These consultants, coming from outside of the organization, use their methodical skills to acquire the relevant domain knowledge. By modeling the processes they gain an understanding of the structures, products, and services of the organization. Subsequently, they manually analyze the process models with the objective of identifying potential weaknesses (Baacke et al., 2009, Becker et al., 2006; Kusiak et al., 1994) or evaluating the compliance of corporate rules and processes (Namiri and Stojanovic, 2007). Furthermore, they try to identify possible risks (Herrmann and Herrmann, 2006; Jallow et al., 2006), to assess the overall performance in areas of business objects, material and organizational resources of an organization (Kueng, 2000; Becker et al., 2009a), or to reorganize processes, e.g. through implementing ICT-concepts (Arendsen et al., 2008; Becker et al., 2008).In recent years four different approaches for the automated analysis of business process models have emerged that are uncoupled with each other (Pfeiffer, 2008):* The formal structural approach for analyzing business process models considers models as graphs. Similarity metrics for graphs have been suggested based on the maximal common subgraph (Bunke and Shearer, 1998) or the graph edit distance (Bunke, 1997). Recent research suggests to apply formal patterns to compare and analyze the formal structure of process models (Delfmann et al., 2009). In the structural approach two business process models are equivalent when they have the same formal structure.* The formal behavioral approach examines the dynamic aspects of process models. The approach comprises multiple, varyingly strong equivalence notions which rely on the formal execution semantics of the underlying models (e.g. Arnold, 1993; de Medeiros et al., 2008; Hidders et al., 2005; Hirshfeld, 1993; Pomello et al., 1992). In general, two business process models are considered equivalent in this approach when both models show an identical behavior during a simulation.* The semantic annotation-based approach has its roots in the ontological research on the foundations of conceptual modeling (Guizzardi et al., 2002; Wand and Weber, 1990). It addresses the analysis of business process models by offering a common terminological reference point in the form of a domain ontology (Gromoff, 2008; Hofferer, 2007; Thomas and Fellmann, 2007). Two model elements are identical when they refer to the same ontology element.* The modeling language-based approach is concerned with specifically designed business process modeling grammars that avoid semantic conflicts in the first place (Pfeiffer, 2007). It addresses the problem of deviations by offering language constructs that limit the choices of the model creator. For this purpose, the set of constructs is carefully selected, and restrictive metamodels or grammars are defined. In this approach two model elements are the same when they have been constructed from the same real world fact.In this paper, we especially address the formal structural approach as we propose structural patterns for an identification of potential process weaknesses and hence a comparison and benchmarking of processes and process path alternatives.\n3 Research design: 3.1 Research propositions\n4 Artifact development: 4.1 Activity-based analysis of structural and semantic business process weaknesses\n5 Artifact demonstration: 5.1 Background information on underlying banking case\n6 Artifact evaluation - discussion of findings and limitations: The example proves evidence that it is possible to automatically identify structural process weaknesses or compare process paths in terms of benchmarking. Within our research project with external partners from consulting and banking, we have identified various structural weakness patterns based on flow chart element occurrences and were able to establish a holistic approach for a pattern-based analysis of processes. As such, we were able to automatically identify potential process weaknesses (P1). We were able to define different weakness types with different complexity and structural depth (P2). The article provides general evidence that it is possible to define such structural weakness patterns and will offer a list of patterns. However, we are not able to define an exhaustive list of patterns in this article, as there will always be new analysis contexts. Furthermore, it is possible to transfer the introduced concept of structural process weakness patterns to other modeling languages that may offer additional possibilities for weakness patterns and quantitative key indicators.The possibility of identifying and formalizing structural weakness patterns are necessary for effectively automating the identification of weak process structures and for benchmarking. Automatically identifying potential weaknesses with the help of quantitative key indicators offers an effective possibility of analysis (P3). However, a manual crosscheck is necessary in a second step in order to ensure correct results. Nevertheless, the automatic (pre-)analysis can significantly unburden modeling experts and process owners in their struggle to improve processes.Reflecting the approach of pattern-based business process analysis, at least two main limitations should be discussed for further research.The pattern-based approach depends upon how well structural weakness patterns are defined and formalized. Identified problems remain \"potential\" weaknesses until a manual analysis reveals that the identified potential weaknesses are actually real weaknesses or not, e.g. due to law regulations. Although the approach can be refined iteratively through empirical evaluation, this depends on the given input for the algorithm. We do not want to discuss the potentials of the area of artificial intelligence and the possibility of automatically learning from/within the pattern-based approach in this context but leave this limitation as is for further examination. Generally speaking it is best to characterize weaknesses with as much detail as possible and also to formalize as many of these characteristics. Following our actual set of structural weakness patterns this also means defining more complex weakness patterns in a next evaluation step compared to using simple patterns to increase effectiveness of our approach. This will help us to find more complex and so more hidden potentials automatically through defining process-spanning weakness patterns in combination with more complex attribute combinations.So far, we have only concentrated on structural weakness patterns. Hence, we only look at syntactic elements of the process flow diagrams. To some extent it may also be possible to automatically analyze semantic weakness although many researchers doubt this possibility. The identification of such patterns especially depends upon how well the processes have been documented. However, this is future research and not part of this research contribution.Going into a more detail, there are further limitations of the presented automatic structural weakness identification approach. For example, we only analyzed one complex and large business process model with various subprocesses on different granularity level and a limited set of weakness patterns, so that our results can just be seen as a first indicator of the potential of this approach for weakness analysis.\n7 Conclusion and further research: With respect to our contribution to the body of knowledge, we have conducted design research according to the design research guidelines, defined by Hevner et al. (2004), by creating an innovative and purposeful artifact for a pattern-based automated analysis of structural weaknesses in business process models. By developing, applying and evaluating our approach, we have provided a research artifact in a rigorous design science research cycle. Applying our approach in practice, it turned out that our modeling and especially our automated analysis approach is highly relevant to the domain of banking and offers much potential for the identification of structural weaknesses and hence for improvements in banking processes. The approach allows a flexible, fast and automatic evaluation of flow chart models based on identified weakness patterns not only by modeling experts but also by decision makers.Thus, it was possible to use the advances in business process modeling languages to reinvent traditional approaches to business process analysis and extend these to in-depth process and activity-based analysis. However, as we argued in our limitations, our methodology for business process analysis is only as good as the people who use it and depend upon the careful definition of structural weakness patterns. In addition, we think that our approach is not limited to the finance sector only but can be applied to process models in different industries.The approach is not limited to flow chart diagrams but can also be used in combination with other process modeling languages that have their roots in the flow chart notation or offer similar capabilities. Furthermore, more complex languages may allow for a more sophisticated analysis as more elements or element combinations can be used for identifying quantitative key indicators. As a result, we believe to have provided a valuable research contribution for benchmarking and weakness identification. Therefore, we suggest future research in the area of how to define the weakness patterns with as much detail as possible. In addition, we recommend research on creating a detailed taxonomy of the different types of structural weaknesses to be found in business processes of banks and even in general in different industries. Finally, we recommend research on applying our enhanced business process analysis methodology in the context of more cases, different industries and even different process modeling languages to prove the generality, we assume in our approach. Giving an outlook on what more potential the idea of automatically identifying structural weaknesses in processes has, we can also imagine that it can be possible to automatically suggest reorganization patterns/alternatives for improving identified weaknesses.\n",
"role": "user"
},
{
"content": "Purpose: - Financial institutions have been engaged in numerous business process reengineering (BPR) projects to make their organizations more efficient. However, the success of BPR projects in banks varies significantly and it remains a challenge to systematically discover weaknesses in business process landscapes. Based on the flow chart notation language this paper seeks to argue for the definition of weakness patterns in order to automatically identify potential process weaknesses.\nDesign/methodology/approach: - The authors developed weakness patterns in the flow chart notation language based on design science principles. To systematically derive process weaknesses that can be formalized, they analyzed each element of the flow chart notation as it was used in a real-life case. They furthermore tested the identified patterns in reality in order to evaluate their validity.\nFindings: - The authors identified various potential weakness patterns that helped in automatically identifying weaknesses in process models. To some extent these findings are generalizable and transferable to other process modeling languages.\nResearch limitations/implications: - The pattern-based approach depends upon how well structural weakness patterns are defined and formalized. Identified problems remain \"potential\" weaknesses until a manual analysis reveals that the identified potential weaknesses are actually real weaknesses or not, e.g. due to law regulations.\nPractical implications: - Using weakness patterns allows for automatically identifying potential process weaknesses in existing flow chart models. This way, this research helps in improving the so far manual analysis of process model landscapes.\nOriginality/value: - The approach is a new way of looking for process weaknesses through process weakness patterns.\n",
"role": "assistant"
}
] |
Does securitization escalate banks' sensitivity to systemic risk?
|
[
"Systemic risk",
"Securitization",
"Bank Holding companies",
"G01",
"G21",
"G32"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The objective of this paper is to explore how balance sheet and off-balance-sheet positions, particularly the securitization activities and the holdings of subprime-related securitized products of the US BHCs have affected the exposure of financial institutions to systemic collapse by transferring risk to third-party investors. This research contributes to the empirical literature on asset securitization and bank risks in several respects. First, the paper focuses on BHCs' individual sensitivity to systemic distress measured by exposure CoVAR and its dependence on both supply (the size of securitization and assets sold activities) and demand (holdings of \"toxic\" securitized products) sides of the securitization. Thus the paper examines whether \"securitization sold\" or securitization bought\" causes the increase of systemic risk. Second, the time horizon under investigation brings focus upon the relationship between securitization and banks' risk exposures including the pre-crisis, crisis and post-crisis periods. Taking into account a series of reforms undertaken by the Basel Committee on Banking Supervision in response to the crises to improve the resilience of banks and banking systems, it remains a remarkable question where the Great Recession has led the market of financial institutions.\n2. Data and data statistics: 2.1 Exposure to systemic risk\n3. Cross-sectional analysis: models specification: This section provides the empirical analysis to test whether the size of outstanding securitization balance of BHCs escalates the firm's sensitivity to systemic risk. The empirical model includes a number of variables to control for bank characteristics and activities which may influence bank risk-taking propensity (see for example Dionne and Harchaoui, 2003; Uzun and Webb, 2007; and Stiroh, 2006). In addition, following Casu et al. (2011), GDP is used to control for macroeconomic effects, while time effects are captured by introducing quarter dummies. The basic regression model is given below:\n4. Estimation results: The parameter estimate of most interest in terms of this study is securitization. The coefficient on securitization is found to be positive and significant at the 1 per cent level as it is evidenced in Table AVI. In other words, a greater outstanding balance of securitization is associated with banks higher level of exposure to financial distress. Signs of coefficients are consistent for two different levels of systemic risk. Looking at the economic significance, the magnitude of the estimated coefficient suggests that 1 per cent increase in securitization balance leads to increase in VaR of a financial institution conditional on market performance by 0.109(0.057) or by approximately 1.3(1.3) per cent relative to the average annual systemic risk measure calculated at 1(5) per cent critical value. Similar conclusions can be applied to the impact of assets sales activities. Further examination of the results reported in Table AV reveals that control variables included in the model are statistically significant, in the line with this study expectation and previous literature. In particular, the ratio of total MBS and ABS securities to total assets of BHC suggests that not only the supply of securitized loans, but also the demand for securitized bonds leads to increased systemic risk, although, the economic effect of the latter is smaller. To test the hypothesis that the supply side (b = 0.109) and the demand side (b = 0.034) standardize beta weights and are statistically different from each other, their corresponding 95 per cent intervals are estimated via bias corrected bootstrap. The results suggest that confidence intervals overlap by less than 50 per cent, the beta weights are statistically significant from each other at 5 per cent significance level (Cumming, 2009). At the same time, examining the magnitude of the estimated coefficients, supply side of the securitization activities reveals more pronounced economic impact. In addition, larger banks tend to have higher exposure to systemic changes. This is consistent with prior empirical studies and particularly could be linked to the size-related diversification effect described by Demsetz and Strahan (1997) and Casu et al. (2011). The parameter estimate on the loan ratio reflects a significant negative impact of the size of the loan portfolio on BHC's systemic risk. Not surprisingly, a greater amount of loans and therefore higher on-balance-sheet credit risk exposure makes banks more vulnerable to market distress. Overall, the regression results show that the current performance of a bank influences its high receptiveness to systemic crisis.\n5. Concluding remarks: This paper examines the impact of different securitzation products on banks' exposure to systemic risks using a sample of U.S. BHCs over the period between 2004 and 2016. Prior literature either investigates the function and impact of securitization activities as whole in banking system or emphasizes on the evaluation of systemic risks for banks. This research adopts an alternative perspective to measure whether the size of securitization balance or the size of holdings of securitized bonds influences the institutional level risk for BHCs, evident not only in the years during the financial crisis, where banks have dramatically off balance changes for their securitization activities, but during the Great Recession as well. While previous research examines mainly the impact of the supply side of securitization activities on systemic risk measures, this study tests whether it is the demand for securitized bonds or the supply of securitized loans that leads to increased sensitivity to systemic risk.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to test empirically the impact of asset securitization and sale activities as well as the holdings of sub-prime related securitized products on the US bank holding companies' (BHC) exposure to systemic risk.\nDesign/methodology/approach: This paper adopts a robust econometric method to estimate the conditional value-at-risk as a measure of BHCs' institutional sensitivity to market crushes. Using the data over the period of 2004-2016, the study also uses OLS with robust standard errors and panel estimation with random effects as two alternative estimation techniques to assess the impact of securitization activities on the sensitivity of BHCs to systemic risk.\nFindings: Residential mortgage and other forms of securitization activities are positively related to an increase in the US BHCs' sensitivity to systemic distress. The significant cross effects of both securitized loans and holdings of securitized products play a crucial role in determining risks in financial sector.\nOriginality/value: This study contributes to the empirical literature on the effects of securitization on BHCs' risk exposures in several ways. First, the paper considers the complexity of the bank's risk profile; it focuses on BHCs' individual sensitivity to systemic distress and its dependence on the size of securitization and assets sold activities considering both supply and demand sides of securitization. Second, the time horizon under investigation sheds a light on the relationship between securitization and banks' risk exposures including the pre-crisis, crisis and post-crisis periods.\n",
"role": "assistant"
}
] |
Digital preservation practices in university libraries: An investigation of institutional repositories in Africa
|
[
"Africa",
"University libraries",
"Librarians",
"Digital preservation",
"Institutional repositories",
"Digital preservation softwares"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Digital preservation remains one of the most critical challenges facing scholarly communities today. Libraries are struggling with how to preserve the scholarly and cultural record now that this information is increasingly being produced in digital formats. From e-journals, e-books, electronic theses and dissertations to e-mails, blogs and more, electronic contents are proliferating fast and libraries worldwide are racing to preserve information for the next generations before technology obsolescence, or even data loss, creep in (Li and Banach, 2011). Digital information is fragile and faces many threats including technological obsolescence and the deterioration of digital storage media (Li and Banach, 2011).\nResearch questions: RQ1. What is the software used to preserve content in the IRs in Africa? RQ2. What are the digital preservation strategies used by librarians in Africa? RQ3. What are the digital preservation policies available in the IRs in Africa? RQ4. Are the necessary rights and agreements in place to preserve the content in IRs? RQ5. Are there sustainable funding and sufficient staff to carryout digital preservation activities in university libraries in Africa?\nLiterature review: Digital preservation efforts in Africa\nMethodology: Data were collected from university libraries in Africa that have developed IRs. The study was in two phases: website investigation and online questionnaire. Because not all university libraries in Africa have developed IRs, so preliminary website investigation of 189 university libraries of English speaking countries in Africa was done to first identify university libraries in Africa that have developed IRs, and the researchers found 52 university libraries with IRs in Africa. Second, online questionnaire was designed by the researchers to collect qualitative and quantitative data from the librarian in-charge of the IR in each of the 52 university libraries. E-mail addresses were collected from the various university websites, and the link to the survey (Appendix) was forwarded to the librarians' in-charge of the IRs.\nResults/Discussions: Software used in IRs in Africa\nConclusion: The study found that the majority of IRs in Africa used DSpace software to manage their contents in the IR and that more than half of the IRs engage in information migration. The study also revealed that the majority of the responding institutions provide long-term digital preservation in their IR. Findings from the study showed that majority of the IRs have digital preservation policies to guide the implementation of digital preservation for the IR content, while others are in the drafting process.\nRecommendations: Based on the findings, the following recommendations are made:\n",
"role": "user"
},
{
"content": "Purpose: The purpose of the study is to investigate the digital preservation practices in institutional repositories (IRs) in Africa.\nDesign/methodology/approach: Data were collected from the IRs developed in university libraries in Africa, and it was done in two phases. The phases are website investigation to identify the university libraries in Africa that have developed IR and online questionnaire.\nFindings: Results from the study showed that the majority of IRs in Africa used DSpace software to manage their digital contents, and more than half of the IRs engage in information migration. The study also revealed that the majority of the responding institutions provide long-term digital preservation in their IR. Interestingly, the majority of the IRs has developed digital preservation policy to guide the implementation of digital preservation for IR contents. Finally, the majority of the respondents indicated that they do not have long-term funding and lack the necessary technical staff with required skills to handle and manage the IR.\nResearch limitations/implications: Because of language barriers, data were collected from only universities in English speaking countries in Africa.\nPractical implications: The findings of this study will make librarians in universities in Africa and other developing countries understand the key issues relating to digital preservation and longevity.\nOriginality/value: The findings of this study will inform information professionals, librarians in developing countries that are planning to create IRs and provide long-term digital preservation of electronic resources in their institution.\n",
"role": "assistant"
}
] |
Building a developmental culture of feedback
|
[
"Trust",
"Leadership",
"Culture",
"Feedback",
"Adult development",
"Individual and organizational capacity building"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: At every level of the system, educators are working to generate new solutions to complex challenges, address pressing issues of equity and diversity, and better support students' well-being and academic achievement. Educational leaders, in particular, have been tasked with leading change with and for students and communities, supporting the diverse adults in their care, and growing their own leadership (Bogotch and Shields, 2014; Grogan, 2013; Jean-Marie et al., 2009; Khalifa et al., 2016). Within this contemporary milieu, professional feedback has emerged as a key lever for school improvement - and for building professional capital and internal capacity (Drago-Severson and Blum-DeStefano, 2016; MacDonald, 2011; Marshall, 2013; Stone and Heen, 2014). Likewise, policy and cultural shifts toward professional learning and collaboration permeate our \"new normal\" world, and underscore the critical importance of effective communication (Davis, 2009; DuFour and Marzano, 2011; Gates, 2013; Goddard et al., 2007; Hargreaves and O'Connor, 2017; Leithwood and Louis, 2012; Ronfeldt et al., 2015).\nTheoretical framework: This paper is informed by two main bodies of literature: current feedback challenges in education, and constructive-developmental theory. As mentioned, and as we will describe in greater detail in our discussion of our research base, it is also informed by our longitudinal research and experiences supporting adult development in universities and K-12 professional learning initiatives.\nResearch base: While in many ways a practice-oriented, conceptual paper, this paper also draws from a rich data pool, including in-depth (open-ended) survey research with 14 educational leaders about their experiences giving and receiving feedback, and more than 25 years of research (ie., surveys, interviews, and evaluations) and work with school leaders around the world. Below, we describe our methods of data collection and analysis, as well as some limitations and possible extensions.\nBuilding a culture of feedback: This section highlights key lessons synthesized from our research and practice, including strategies for building a culture of feedback. This is essential because, as many educators have shared with us, meaningful, effective feedback is largely dependent on the context in which it is offered. When, for example, \"real\" feedback falls outside of organizational norms, educators share that it can be hard to offer one's authentic thinking with colleagues and directors, for fear of retribution or \"rocking the boat.\" In a similar way, when feedback is presented or shared in ways that heighten anxiety (or feel like \"gotcha moments,\" as many participants described), it can be hard to reframe feedback as a tool for growth, or an expression of trust and respect. As one principal from our survey research explained (and this is a common sentiment leaders voice):most feedback to which we're initially introduced is negative: parents/teachers/ coaches/directors telling us what we did wrong, police telling us we broke the law, employers telling us we will be fired [...]. These are not great introductions to feedback.\nConclusion and significance: For many years, John Hattie (2012a) has argued that teacher feedback is one of the most powerful influences on student achievement - although this influence can be either positive or negative. Research about professional feedback in both the education and business sectors similarly points to the banes and boons of the feedback adults give and receive (e.g. Hargreaves and O'Connor, 2017; MacDonald, 2011; Marshall, 2013; Stone and Heen, 2014), yet a developmental perspective has been largely missing from this literature. In this paper, we described the importance of building a culture of feedback that supports meaningful collaboration and connection as the foundation for any meaningful change. The fact is, we all need to feel well held (Drago-Severson, 2012) in our work and learning to lean into the risk and exhilaration of growth - and to share our honest thinking, feeling, and feedback with others. Toward this end, this paper outlined six promising, developmental strategies for nurturing the preconditions of trust, safety, and respect that set the stage for what educators in our research describe as \"real\" feedback: recognizing the value of mistakes, modeling vulnerability and a learning stance, honoring the (inter)personal, making expectations clear from both sides, sharing developmental theory and practices, and employing and strengthening collaborative structures (Drago-Severson and Blum-DeStefano, 2016).\n",
"role": "user"
},
{
"content": "Purpose: This paper draws from more than 25 years of research with aspiring and practicing educational leaders to present six strategies for building a culture of feedback in schools, teams, districts, professional learning opportunities, and other educational settings. These strategies reflect key elements of the authors' new, developmental approach to feedback. The paper aims to discuss these issues.\nDesign/methodology/approach: Through the lens of adult developmental theory, the authors highlight foundational learnings from open-ended survey research with 14 educational leaders about their experiences giving and receiving feedback, and prior qualitative, mixed-method, and longitudinal research with principals, assistant principals, teachers, superintendents, and other educational leaders.\nFindings: The authors share six developmentally oriented strategies for establishing trust and building conditions for authentic, generative feedback: finding value in mistakes, modeling vulnerability, caring for the (inter)personal, clarifying expectations, sharing developmental ideas, and building an infrastructure for collaboration.\nPractical implications: This work has implications for leadership and leadership preparation, especially given contemporary emphases on collaboration and high-stakes evaluations as tools for ongoing improvement, enhancing professional capital, and internal, individual, and system-wide capacity building.\nOriginality/value: Because a developmental perspective has been noticeably missing from the wider feedback literature and leadership preparation curricula, this work extends and enhances tenets from different fields (e.g. business, developmental psychology, educational leadership and educational leadership preparation), while also addressing urgent calls for educational reform; leadership preparation, development, and practice; and professional capital building.\n",
"role": "assistant"
}
] |
Organizational culture and knowledge creation capability
|
[
"Organizational culture",
"Knowledge creation",
"China"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Knowledge creation capability, which refers to an organization's ability to exchange and combine knowledge to create new knowledge, plays a critical role in competitive advantage (Nahapiet and Ghoshal, 1998). But, why are there significant differences between firms' knowledge creation capabilities and how can firms improve their knowledge creation capabilities? Existing research clearly indicates that organizational routines are critical antecedents of knowledge creation capability (Smith et al., 2005). Thus, as an important element of organizational routines, organizational culture has a significant effect on knowledge creation capability (Bhagat et al., 2002; Park et al., 2004).Owing to the importance of knowledge creation capability, its antecedents become a crucial research issue in that it can provide a more complete picture on how firms create new knowledge and advance knowledge management research (Smith et al., 2005). Investigating the impact of organizational culture on knowledge creation capability can identify a key antecedent of knowledge creation capability, thus it has a distinct contribution to knowledge management research. Unfortunately, there is minimal research on it, which is a serious research gap.The purpose of this study, therefore, is to address this research gap. Specifically, the authors employ three dimensions of organizational culture - individualism-collectivism, uncertainty avoidance, and power distance - to examine the impact of organizational culture on knowledge creation capability. These three dimensions have been validated and widely used in existing organizational culture research (e.g. Wuyts and Geyskens, 2005).Unlike previous research, the authors use the survey data of Chinese firms to empirically test the hypotheses. China is deemed appropriate as the research setting for three compelling reasons. First, there is a high variation in knowledge creation capability among Chinese firms. Some Chinese firms belong to technological leaders in the world, such as ZTE and Huawei, while a larger part of Chinese firms have relatively less sophisticated capabilities and knowledge (Hitt et al., 2000). Such a high variation in knowledge creation capability provides an ideal context to test its antecedents. Second, as the research horizon of knowledge management is now increasingly expanded to China, it is critical to know more about \"what is going on there\" (Meyer, 2006). Finally, improved knowledge about China has enormous implications for Western firms to compete and/or collaborate with Chinese firms (Su et al., 2009).\n2. Literature review and hypotheses development: 2.1 Knowledge creation capability\n3. Methods: 3.1 Sample and data collection\n4. Discussion and conclusion: 4.1 Contributions\n",
"role": "user"
},
{
"content": "Purpose: - Organizational culture comprises a firm's climate that informally and tacitly defines how the firm develops and uses knowledge, thus it has a significant effect on knowledge creation capability. The purpose of this study is to investigate the impact of organizational culture on knowledge creation capability.\nDesign/methodology/approach: - The data of 212 Chinese firms collected through face-to-face interview is used to empirically test the hypotheses.\nFindings: - This study finds that organizational culture plays a critical role in knowledge creation capability. Specially, collectivism has a positive impact on knowledge creation capability, while power distance and uncertainty avoidance have negative effects.\nOriginality/value: - This study not only contributes to knowledge management research by identifying a key antecedent of knowledge creation capability - organizational culture - but also is of importance to organizational culture literature by demonstrating the proper organizational culture for knowledge creation capability.\n",
"role": "assistant"
}
] |
The development of supply chain risk management over time: revisiting Ericsson
|
[
"Supply chain risk management",
"Ericsson",
"Resilience",
"Capabilities",
"Longitudinal",
"Risk monitoring",
"Reactive",
"Proactive"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: When entering the Ericsson headquarters in Stockholm, Sweden early in the morning of June 18, 2018, for the purpose of conducting interviews, we, the authors of this article coincidentally experienced the effectiveness of Ericsson's supply chain risk management (SCRM) approach in real time. One of the interviewees asked us whether we had heard about the 5.5Mw earthquake that had hit Osaka, Japan, the same morning at 7.58 a.m. Japan Standard Time. While we had not, Ericsson had already leveraged its SCRM processes. Shortly after the first alerts and before our arrival, Ericsson had analyzed the number of suppliers and sub-suppliers in the affected geographical area; investigated manufacturing sites; analyzed the potential financial influence on sold products (business interruption value) and whether there would be any supply disruption; discussed alternative suppliers; contacted suppliers in the affected area by local sourcing offices to garner information about people hurt, damages and affected flows that could prevent deliveries to Ericsson; evaluated whether Ericsson personnel were in the area and/or hurt; and received information from Japanese suppliers that they were not impacted. Ericsson's conclusion from this quick initial analysis process, in place since 2011, was that there was no need to \"press the red button.\" What we observed was not a major event for Ericsson, but it gave us a taste of the effectiveness and practical applicability of Ericsson's contemporary SCRM approach.As Ho et al. (2015) found that most SCRM research is of a theoretical nature, they advise scholars to use primary data to investigate the practical applicability of SCRM models. They found the original article about the Ericsson case (Norrman and Jansson, 2004) to be one of very few articles that investigates SCRM with the aid of a real-life case. In addition, in their literature review, Fan and Stevenson (2018) identify a crucial research gap in the holistic approach that considers all four stages (identification, assessment, treatment and monitoring) of the SCRM process being almost absent in the academic literature. They identified only six articles, including Norrman and Jansson (2004), that have taken a holistic approach that covers all stages, with only two of them (Tummala and Schoenherr, 2011; Lavastre et al., 2012) having been published less than 10 years ago. Concentrating on only a few process stages comes with the downside of missing out on what practitioners have repeatedly highlighted as being the key to SCRM in an effective manner: the understanding of the interwoven connections between the identification, assessment, treatment and monitoring stages. What is missing in the literature, therefore, is an updated real-life approach with a holistic perspective.A distinction can be made between proactive and reactive SCRM. The proactive approach requires decision-makers to be able to forecast possible future changes and resist these forecasted changes (Wieland and Wallenburg, 2013). Although ongoing technological developments might increasingly enable decision-makers to get closer to \"total control\" of the end-to-end supply chain, the expectation that this can at some point be possible is certainly too optimistic (Hoberg et al., 2020). Therefore, while necessary, proactivity alone is not sufficient for SCRM. Reactivity is also needed, i.e. the actions which are required after a risk has already been detected. Organizations need to be able to (1) recognize a risk and initiate a response; (2) put in place a disruption management team; (3) develop an initial plan; (4) review and revise the plan in light of new information; and (5) evaluate the reactive work, learning to improve for future risks (Hopp et al., 2012). An essential part of reactive practices is the protection of the reputation of the company, highlighting the importance of communication after a risk occurrence (Bland, 2013; Ponis and Ntalla, 2016). The resilience literature assumes that systems, like supply chains, constantly evolve, suggesting that systems should be able to adapt by coping with changes (Holling, 1996). Within SCRM, there needs to be the proper balance between proactive and reactive approaches.To help to fill the research gap described regarding holistic real-life cases, the purpose of this invited article is to assess Ericsson's contemporary SCRM practices and how both its proactive and reactive SCRM procedures have developed over time. Thus, we returned to Ericsson to collect new data for a longitudinal case study and complement the original findings. Recommended steps for conducting rigorous case study research have been followed (e.g. Gibbert et al., 2008; da Mota Pedrosa et al., 2012). To study the focal unit of analysis (the SCRM practices of Ericsson) and the developments since 2004, a case study protocol including a semi-structured interview guide based on an initial research framework has been used (available on request). Different data sources were combined during the data collection process: archival data as well as interview data from multiple senior informants representing different involved functions and perspectives (Table 1). The interviews were recorded at the Ericsson head office on June 18 and 19, 2018, and took place as group interviews. The transcribed and summarized data were later shared and discussed with the main informants during multiple rounds to validate and complement the data. Also, our analysis was shared with Ericsson for the same purpose.The article proceeds by presenting how Ericsson's SCRM practices have developed over time and across different eras. One incident, Fukushima in 2011, is highlighted to provide empirical details. Thereafter the analysis is presented, followed by a concluding discussion of implications and suggestions for future research.\nEricsson's SCRM practices throughout the course of time: Strategic profile of Ericsson's supply chains\nAnalysis of the developments: We now turn to an analysis of the general developments over time, this time not by distinguishing between eras, but by taking different management perspectives on SCRM, starting with Ericsson's governance mechanisms and organizational structure. This is followed by their processes and tools; systems; assessment and learning; and finally triggering events and enablers.Corporate governance and organizational structure\nConclusions, contributions and future research: This article contributes to the need for more holistic perspectives on the SCRM cycle (Fan and Stevenson, 2018) and incorporates both proactive and reactive practices in rather stable times as well as when incidents are emerging and ongoing. By studying the case of Ericsson as a real-life example and in continuation of the approach outlined 15 years ago in Norrman and Jansson (2004), the article also responds to the need for more investigation of how different SCRM models are used in practice (Ho et al., 2015), with many tools being described that have the potential to be useful in practice. Another contribution lies in the detailed empirical illustration of the consequences of the 2011 earthquake and tsunami in Japan for Ericsson, a situation that exemplifies Hopp et al.'s (2012) five parts of reactive SCRM practices (recognition; get a disruption management team in place; initial planning; reviewing and revising the reactive practices; and evaluating and using learning from the reactive work for improvement). The study also contributes to supply chain risk monitoring, an area that has almost certainly received the least attention in the SCRM literature (Berg et al., 2008; Fan and Stevenson, 2018), for instance, by describing how Ericsson assesses the implementations of SCRM projects, mitigation strategies and tools.However, the main contributions are insights regarding the development of SCRM practices and capabilities over time, a matter that, to our knowledge, has not been investigated before. A longitudinal view is taken regarding the development of a large global corporation over 15 years. By extracting different eras, overall patterns have been found, including primary development activities, enablers and trigger points perceived as important for both the continuous development of SCRM and ensuring that an SCRM culture would stick in the organization.In summary, over time, Ericsson has developed its SCRM practices by first formulating and improving functional processes and tools (technical as well as knowledge-based capabilities) within the Supply, Sourcing and Security functions, but then also combining them cross-functionally. This has been facilitated by improved technical information systems, but also through a more fine-grained and cross-functionally coordinated organization for SCRM. Another enabler is a strengthened focus on monitoring and compliance, which has supported implementation and served as a platform for a learning organization. In general, processes have been more formalized throughout time, bolstered by corporate governance wherein during the studied time period, many corporate projects were implemented that are aligned with, and hence enable, the development of SCRM.Future research could further address the development of SCRM capabilities, for example, by investigating the dynamic capabilities (Teece et al., 1997) that are needed to modify ordinary SCRM capabilities, or the relationship between (inter-)organizational learning and generating resilient supply chains.Future research could investigate the following eight enablers, which this study suggest to have a positive impact on SCRM outcomes. Four enablers relate to how relationships should be established within and across business functions and organizations: (1) Top management support: Ensure that the corporate senior management gives sufficient priority to SCRM practices and keeps being focused on SCRM also over longer periods when no major incidents occur; (2) Intra-functional processes: Simultaneously develop and fine-tune functional SCRM processes and tools, and integrate them across functional silos to reduce risks; (3) Cross-functional relationships: Implement and organize SCRM across business functions in a formalized and fine-grained fashion to make the risk culture \"stick\" to the whole organization; (4) Inter-organizational relationships: Collaborate with supply chain partners and other external actors across organizational boundaries to increase the visibility and speed of SCRM processes. The next four enablers support the implementation of SCRM internally: (5) Adaptability: Focus on continuous development and formalized learning from experiences, thereby offering training possibilities that can help to turn experiences from reactive actions into proactive processes; (6) Monitoring risk: Focus on monitoring and assessing the level of compliance with tools and the implementation of projects, and develop mitigation strategies to catch up with ongoing developments; (7) Risk-management scope: Extend the scope of SCRM by incorporating more types of risks - both downstream and upstream; (8) Technological capabilities: Develop and leverage technological capabilities (e.g. business analytics) to establish visualized, formalized and rapid SCRM practices that enable both proactive and reactive decisions. Regarded in isolation, each of these enablers might not be novel. However, it is the way successful companies like Ericsson integrate them that can turn them into powerful practices.This study points to the need to formalize SCRM processes, integrate resources, and enforce compliance and measurement practices. A potential downside of this could be rigidity, bureaucracy and, in turn, even more vulnerability. Future research could therefore investigate how management practices could simultaneously be adaptive and formalized, thereby ensuring that SCRM is still open enough to cope with the \"new normal.\" It could be analyzed whether more formalized SCRM processes, compliance and \"learning from history\" can improve the capabilities to observe and understand new risk sources or whether this could create blinders, even impeding the manager's attention towards new risk scenarios. Overly rigid processes could even demotivate managers and prevent them from being experimental and creative.Ericsson increased visibility by developing and combining IS (e.g. ERP, planning and analytics) within and between different business functions. An alternative approach is to buy and implement commercial \"off-the-shelf\" systems. Future research could focus on the best SCRM practices for different types of organizations, potential difficulties when establishing new IS, and potential challenges when leveraging already installed IS (internally and between supply chain partners).This study highlights the importance of cross-functional SCRM approaches, but it also demonstrates the difficulty to maintain them and that they are often triggered by major incidents. More research could complement previous research on functional silos in SCM (e.g. Richey et al., 2010) by concentrating on what establishes and conserves such silos in the context of SCRM. Particularly worthwhile could be research on how organizations could, in both stable and turbulent times, proactively tear down functional silos that could impair successful SCRM.Given that external stakeholders and corporate leaders turn out to be important triggers, more research could explore their interdependent influences on SCRM. Especially in supply chains with many independent decision makers and fragmented SCRM practices (e.g. critical infrastructure supply chains operated as public-private partnerships) holistic governance approaches can be crucial. Theories about risk governance and supply chain risk governance (Ahlqvist et al., 2020) could be further developed, especially related to what kind of governance mechanisms and incentives will align with different situations to increase the inter-organizational coordination of different stakeholders' and actors' SCRM behaviors.The Ericsson case shows that organizations need to ensure that their proactive SCRM efforts do not loose pace. While incidents and other triggers can help to alert an organization, this should not be the main mechanism for creating a proactive SCRM culture. While this article points out that formalization, an increasingly fine-grained organization, monitoring, and a use of structured scenario-based training can keep an organization awake, more research could focus on investigating how different types of organizations can become, and stay, proactive also in relatively stable periods.Finally, this case shows how perceptions of supply chain risks and the appropriate organizational structure change over time. An avenue for future research would be to deepen our understanding of the drivers of managers' risk perception to change, learning patterns regarding what solutions fits when, and ways of how lessons from one incident can be transferred to novel risk situations.We hope that both the proposed practices and suggested avenues for future research will inspire SCRM in both practice and research, leading to the advancement of SCRM knowledge.\n",
"role": "user"
},
{
"content": "Purpose: This invited article explores current developments in supply chain risk management (SCRM) practices by revisiting the classical case of Ericsson (Norrman and Jansson, 2004) after 15 years, and updating its case description and analysis of its organizational structure, processes and tools for SCRM.\nDesign/methodology/approach: An exploratory case study is conducted with a longitudinal focus, aiming to understand both proactive and reactive SCRM practices using a holistic perspective of a real-life example.\nFindings: The study demonstrates how Ericsson's SCRM practices have developed, indicating that improved functional capabilities are increasingly combined across silos and leveraged by formalized learning processes. Important enablers are IT capabilities, a fine-grained and cross-functional organization, and a focus on monitoring and compliance. Major developments in SCRM are often triggered by incidents, but also by requirements from external stakeholders and new corporate leaders actively focusing on SCRM and related activities.\nResearch limitations/implications: Relevant areas for future research are proposed, thereby increasing the knowledge of how companies can develop SCRM practices and capabilities further.\nPractical implications: Being one of few in-depth holistic case studies of SCRM, decision-makers can learn about many practices and tools. Of special interest is the detailed description of how Ericsson reactively responded to the Fukushima incident (2011), and how it proactively engaged in monitoring and assessment activities. It is also exemplified how SCRM practices could continuously be developed to make them \"stick\" to the organization, even in stable times.\nOriginality/value: This is one of the first case studies to delve deeper into the development of SCRM practices through taking a longitudinal approach.\n",
"role": "assistant"
}
] |
Partial mediation of the political skill-reputation relationship
|
[
"Political skill",
"Relationship quality",
"Personal reputation",
"Workplace",
"Employees attitudes",
"Interpersonal skills"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: It has been suggested that organizations are inherently political arenas (Mintzberg, 1985). Although organizational politics traditionally have been viewed negatively, the most current perspective is more neutral and objective, and even positive in nature (e.g. Hochwarter, 2012). Sederberg (1984, p. 7) characterized organizational politics as attempts to \"create, maintain, modify, or abandon shared meanings.\" Similarly, scholars have argued that the objective of political behavior \"is to manage the meaning of situations in such a way as to produce desired, self-serving responses or outcomes\" (Ferris et al., 1994, p. 4).\nTheoretical foundations and hypothesis development: This study intends to expand theory regarding how political skill helps individuals make positive or favorable personal reputations at work through their interactions with others. Personal reputation development often involves deliberate and purposeful action, whereby individuals are motivated to manage impressions to maximize goals of identity creation and maintenance, which is instrumental in the acquisition of rewards and prevention of punishment (Bozeman and Kacmar, 1997). Also, individuals may engage in political manoeuvres to manipulate reputational signalling to their advantage (e.g. Ferris and Judge, 1991).Politically skilled individuals are effective at building their personal reputations because they transmit signals that establish favorable images to the public through their proactive network building activities, and the use of situationally appropriate influence tactics. With their social astuteness and flexible, adaptive approach to interpersonal influence, politically skilled individuals know precisely how to demonstrate their behavioral repertoire in order to develop the desired image (e.g. Ferris et al., 2007). Furthermore, the personal reputation signals transmitted by politically skilled individuals are likely to spread quickly because of the established connections with influential people in their social networks.More specifically, it is argued that politically skilled individuals develop strong relationships with their immediate supervisors in order to secure the outcomes associated with positive personal reputations. These politically skilled individuals also perform citizenship behaviors that benefit their coworkers, thus signaling their worthiness of a positive personal reputation (Spence, 1974). Consequently, decision makers who observe these politically skilled individuals' relationship-building behavior and OCBs should reciprocate their efforts, from a social exchange perspective (e.g. Blau, 1964), and evaluate them favorably, assigning them positive personal reputations. The proposed model is presented in Figure 1, and in the next sections, the theoretical foundations of the model are discussed, the model linkages are specifically articulated, and hypotheses are developed.Theoretical foundations of the model\nMethod: Sample and procedure\nData analyses and results: Prior to testing the model via SmartPLS, an exploratory factor analysis (EFA) was performed on the personal reputation scale. It has been suggested that personal reputation at work has two dimensions of performance and character (Ferris et al., 2003; Zinko et al., 2007). SPSS 17.0 was used to conduct a principal axis factor analysis with an oblique, direct oblimin factor rotation scheme for the personal reputation scale. The exploratory factor analysis indicated that the personal reputation scale is composed of the two dimensions of character and performance. The first five items loaded on the character dimension, whereas the last seven items loaded on the performance dimension.The hypothesized model (presented in Figure 1) was tested with an item-level structural equation model (SEM) using variance-based SmartPLS Graph Version 2.0M3 (Ringle et al., 2005). SmartPLS simultaneously analyzes the reliability and validity of the measurement model and estimates hypothesized structural model parameters. The primary rationale for using SmartPLS was our relatively small sample size, which was an outcome of the triadic nature of the data collected. PLS is robust to small sample sizes. In particular, it avoids the non-convergence problems and improper solutions that are common to covariance-based techniques using sample sizes of less than 200 (Boomsma and Hoogland, 2001).Additionally, in line with the theory-building nature of this study, PLS is an appropriate method in contexts of rich data and weak theory (Wold, 1985). Like regression analysis, PLS follows a rule-of-thumb of ten cases per predictor. That is, the sample size must be equal to ten times the largest of the variables with the most indicators or the endogenous variable with the most independent variables impacting it (Bentler and Chou, 1988). Following this heuristic, we would have required at least a sample of 110, which we exceeded by a fair margin with our 145 triads.Tests of the measures\nDiscussion: Ferris et al. (2007) proposed a theory of social/political influence in organizations that focused on the critical role of political skill. Although the development of personal reputation was an important consequence of political skill in organizations, the intermediate linkages between political skill and personal reputation had not been considered to date. The present study attempted to extend theory in this area by testing the political skill-personal reputation relationship as well as hypotheses that politically-skilled individuals develop high-quality work relationships with their immediate supervisors, and perform citizenship behaviors that benefit their coworkers, in ways that build personal reputation; thus, that work relationship quality and citizenship behavior serve as multiple partial mediators of the political skill-personal reputation relationships.Contributions to theory and research\n",
"role": "user"
},
{
"content": "Purpose: - Although reputation is important to career success, little is known about how individuals develop their personal reputation at work. This study seeks to investigate the role of work relationship quality and citizenship behavior as partial mediators of the political skill-personal reputation relationship.\nDesign/methodology/approach: - In total, 145 triads from a manufacturer in the Midwestern USA provided data for the study. Employees rated their political skill and citizenship behavior, supervisors rated their relationship quality, and coworkers rated the employees' personal reputation. Based on the complementary theories of signaling and social exchange, the relationships between the constructs were analyzed with structural equation modeling.\nFindings: - Political skill demonstrated both direct and indirect effects on the development of personal reputation. In particular, work relationship quality and citizenship behavior partially mediated the relationship between political skill and personal reputation.\nResearch limitations/implications: - Personal reputation was evaluated by a randomly selected coworker, but a collection of perceptions would be helpful.\nPractical implications: - Political skill training and/or mentoring relationships may help individuals manage their personal reputation at work, thus benefiting their careers.\nSocial implications: - This study focused on personal reputation in a work environment. However, the results also may be useful to individuals in different types of organizations.\nOriginality/value: - This is one of the first studies to investigate how individuals develop their personal reputation at work. Unlike previous research that used self-evaluations of personal reputation, this study used peer evaluations, which is more appropriate for the construct.\n",
"role": "assistant"
}
] |
Master the three phases of business continuity planning
|
[
"Business planning",
"Disasters",
"Emergency measures"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Suppose your company is in an office building housing multiple firms. You arrive at the office one morning and discover that the entrance to your suite has been shattered. There is glass everywhere. You call the police and assess the situation. Computer and conference room equipment has been stolen.\nThe three continuity phases: The typical corporate approach of writing a disaster response manual covering all the phases and disruptions simply doesn't work. The results are lost time, lost revenue and lost opportunity. The following guidelines will keep your company on the growth path and out of bankruptcy court when the inevitable happens.Table I summarizes the major activities in each of the continuity phases (i.e. resolve, respond and rebuild). Let's look at them in chronological order.The resolve phase\nThe three disruption levels: Now that we've covered the phases of business continuity, it's time to look at the disruption levels - minimal, moderate and major - in more detail. These also require unique skills and approaches.Of course, the level of disruption is relative to the preparedness of the organization. For example, imagine what would happen if Raleigh, North Carolina experienced a winter like an average one in Buffalo, New York. The city of Raleigh would be seriously incapacitated because they are simply not prepared or equipped for such conditions. Meanwhile, Buffalo carries on normally, and without fanfare, year after year.So, what constitutes a disaster? There are many definitions, mostly by government agencies. They often refer to casualties and the loss of life. But disasters are defined by more than just their magnitude.In simple terms, a disaster is any adverse situation requiring people to act urgently, do things they don't normally do, and face circumstances they may never have seen before.By this definition, responding to a disaster is more than just following procedures and deploying additional personnel. It is also about having strategies for dealing with the unexpected, the uncommon and the unknown.Table II shows the three disruption levels and provides guidelines for differentiating them. Of course, there will always be situations that straddle the boundaries. When in doubt, use the next higher level. It is better to over-respond.\nA minimal disruption: Most disruptions are minor and should stay that way. By responding quickly and decisively, minor disruptions can be prevented from mushrooming into full-blown major disasters.Small fires, water damage, human error and security violations are examples of adverse events that are frequently encountered. Staff members should be identified to respond to these types of emergency situations. They should be trained in what to do and who to call so that they can be expected to react quickly and without hesitation.Training exercises should be conducted on an irregular basis to keep everyone sharp. Knowing what to do and where critical items are stored can make a big difference as the following story sadly shows.This story was reported by the DeKalb, Illinois, Daily Chronicle in 2004. A small storage facility was severely damaged when a fire started in one of the storage units. The fire was most likely started carelessly by a spent cigarette or improper disposal of cigarette ashes.Such fires are easily extinguished if controlled quickly with a handheld fire extinguisher. In this case, the lack of a fire extinguisher sent employees scrambling to find one and allowed the fire to get out of hand. The result was $125,000 in damages plus substantial lost revenue.Precious time was wasted looking for fire extinguisher and a small fire was allowed to do extensive damage. Proper training and regular safety inspections could have limited this disaster event to the nuisance category.A moderate disruption\nThe isolated disaster: The most common disasters are fairly isolated. An isolated disaster is one that affects only a single system, product or group, for example:*a critical information technology system such as e-mail or the customer database is damaged;*a manufacturing line is disabled; or*a field office is hit by avian flu.These are the most likely situations you will encounter, not tsunamis, hurricanes or terrorist attacks. With proper preparations, isolated disasters can be handled quickly and effectively. Operations can be moved to another area, replacement systems can be brought online, repairs can be expedited, and offices can cover for each other. It just takes planning!Business continuity plans must evolve with the company and the threat landscape. Whenever new systems are installed, new offices are opened, or new threats appear, make business continuity planning a priority. Treat this planning as an ongoing effort, not a one-time event.Rapid growth requires more than just finding new customers and markets. You must manage small disruptions and prevent them from becoming major meltdowns.\n",
"role": "user"
},
{
"content": "Purpose: - One of the biggest hindrances to corporate growth lurks in the shadows - emergency disruptions. How companies deal with them will govern how rapidly they are able to grow. There is too much emphasis on \"major disasters\" like floods, earthquakes and terrorist attacks. Meanwhile, there is not enough emphasis on \"daily disruptions\" like security violations, supply problems and equipment breakdowns. It's the little things that can destroy productivity and leave a company scrambling for survival.\nDesign/methodology/approach: - A comprehensive review of the literature related to business continuity and disaster recovery planning served as a foundation for this article, though the real-world stories of how companies handled disasters was more valuable.\nFindings: - Most companies create lengthy manuals full of disaster response information. This material is difficult to understand and maintain. Worst of all, there is little emphasis on prevention.\nPractical implications: - Companies must separate business continuity planning into three phases: planning and prevention (resolve phase), disaster response (respond phase) and, return to normal (rebuild phase). They must also think about three levels of disruption because disasters vary widely in scope and complexity.\nOriginality/value: - Breaking down business continuity planning into phases and levels is a new approach. Senior executives, middle managers and business owners should embrace the concepts outlined in this article.\n",
"role": "assistant"
}
] |
The impact of experiential augmented reality applications on fashion purchase intention
|
[
"Mobile applications",
"Purchase intention",
"Augmented reality",
"Hedonic motivation",
"Experiential retail"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Retail has experienced seismic shifts over recent years due to the growth of digitalization and online channels (Verhoef et al., 2015). Grewal et al. (2017) argued that retailing is evolving at an accelerated rate due to changes made possible by new technologies. One such technology, with \"the potential to transform the shopping experience\" (Duncan et al., 2013, p. 6) is augmented reality (AR). AR can enhance sensory perceptions for consumers by superimposing virtual elements directly into the real-time environment (Yaoyuneyong et al., 2016). With the increasing ubiquity of smartphones and tablets, AR applications (apps) are increasingly being embraced by retailers as a tool for creating immersive customer experiences. For example, Burberry, Topshop, Sephora and Panasonic are all examples of retailers which have recently launched AR mobile apps. However, whilst it has been argued that AR will play an important role in the future of retail (Grewal et al., 2017; Javornik, 2016a), our understanding of how it impacts consumer behavior is still relatively under researched.\n2. Theoretical background: 2.1 AR in experiential retailing\n3. Methodology: To test the hypotheses, an online experiment was conducted with 162 participants, where the experiential AR makeup retail app of a leading cosmetics brand served as the stimulus. Two conditions existed within the experiment, using an approach in keeping with Javornik (2016a). In the condition \"augmentation,\" participants interacted with the AR makeup app (having downloaded it), while in the condition \"non-augmentation,\" they interacted with the mobile makeup shopping site from the same brand. Thus, the effect of the stimulus \"augmentation,\" could be measured in relation to purchase behavior.\n4. Results: Table II presents the means, standard deviations and correlations of the constructs.\n5. Discussion: This study drew on the SOR model (Mehrabian and Russell, 1974) to understand the potential effect of augmentation on purchase intention, where augmentation formed part of the stimulus. The SOR model is well established in the broader experiential retailing literature, but its application to AR is limited. Whilst Javornik (2016a) did consider both affective and behavioral responses to AR technologies, she did not consider the relationship between affective and behavioral responses (rather they are treated as different dimensions of brand related responses). The findings here suggest that augmentation creates a more positive emotional response than interactions without augmentation. This finding is in keeping with Javornik (2016a), who also found that augmentation creates stronger positive emotional responses than other forms of interactivity. However, our results further show that it is this enhanced emotional response that creates greater purchase intention for those experiencing augmentation, i.e. the effect of augmentation on behavior (in this case purchase intention) is mediated by the positive affective response it evokes. To understand why this enhanced emotional response is achieved, it is useful to consider some of the key features of augmentation. Augmentation provides modality richness, the \"intensity with which a mediated environment is able to present information to the senses\" (Li et al., 2002, p. 45), presumably by creating greater sensory depth. Thus, as suggested by Javornik (2016a), AR does not necessarily create greater interactivity than apps without AR; rather, it creates a rich sensory experience, resulting in stronger emotional (and therefore behavioral) responses.\n",
"role": "user"
},
{
"content": "Purpose: Utilizing the stimulus-organism-response model, the purpose of this paper is to examine the effects of augmented reality (AR) (specifically augmentation) on consumers' affective and behavioral response and to assess whether consumers' hedonic motivation for shopping moderates this relationship.\nDesign/methodology/approach: An experiment using the manipulation of AR and no AR was conducted with 162 participants aged between 18 and 35. Participants were recruited through snowball sampling and randomly assigned to the control or stimulus group. The hypothesized associations were analyzed using linear regression with bootstrapping.\nFindings: The paper demonstrates the benefit of using an experiential AR retail application (app) to positively impact purchase intention. The results show that this effect is mediated by positive affective response. Furthermore, hedonic shopping motivation moderates the relationship between augmentation and the positive affective response.\nResearch limitations/implications: Because of the chosen research approach, the results may lack generalizability to other forms of augmentation. Therefore, researchers are encouraged to test the proposed model using different types of AR stimuli. Furthermore, replication of the study with other populations would increase the generalizability of the findings.\nPractical implications: Results of this study provide a valuable reference for retailers of the benefits of using AR when attempting to optimize experiential value in online environments.\nOriginality/value: The study contributes to experiential retail and consumer purchase behavior research by deepening the conceptualization of the impact of experiential technologies, more specifically AR apps, by considering the role of hedonic shopping motivations.\n",
"role": "assistant"
}
] |
The bright side and dark side of group heterogeneity within online brand community
|
[
"Online brand community",
"Brand community",
"Brand commitment",
"Tenure",
"Brand community commitment",
"Group heterogeneity",
"Value heterogeneity",
"Visible heterogeneity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Recent decades have witnessed the growing popularity of the online brand community (OBC). Empowered by the growth of social media, an OBC is usually understood as an online social group composed of a brand's admirers (Hanson et al., 2019). OBCs enable members to seek brand knowledge, share brand experience, find socialization and participate in activities initiated by a brand (Raies et al., 2015; Kumar and Nayak, 2019). Traditionally, brand community literature has been dominated by the homogeneity view that brand community formation and engagement are based on members' homogeneity. For instance, Muniz and O'guinn (2001) maintain that brand community is characterized by consciousness of kind, rituals and traditions, and moral responsibility. Stokburger-Sauer (2010) propose that members in a brand community are connected by collective goals and mutual commitment. Using this view, scholars have contributed ample evidence of the positive role of brand community in creating brand loyalty (Manchanda et al., 2015).\nLiterature review: Online brand community\nHypothesis development: The effect of perceived visible heterogeneity\nMethod: Sample\nDiscussion: The popularity of OBCs has been continually growing in the past decades. A small but burgeoning body of literature reveals that member heterogeneity within OBCs is extensive (Ozboluk and Dursun, 2017; Pan et al., 2014). Accordingly, understanding the consequences of brand community heterogeneity is becoming essential. Drawing from organization research, we examined the effect of community heterogeneity on brand community commitment and brand commitment. Using a survey, we found evidence that perceived visible heterogeneity fosters brand community commitment. This finding suggests that members may regard visible heterogeneity as an advantage of brand communities because it offers them good opportunities to interact with people of diverse background and thus creates a positive community experience. It is consistent with prior research that suggests that heterogeneous online communities can produce more informational and social benefits (Butler, 2001; Krebs et al., 2006). In contrast to the positive effect of perceived visible heterogeneity, the results demonstrated that perceived value heterogeneity impedes brand community commitment. Probably, the negative effect arises because value heterogeneity undermines the mutual understanding of brand meaning and shared consciousness among members, the cornerstones of a brand community (Muniz and O'Guinn, 2001). The negative effect indicates that OBCs are a double-edged sword that can backfire, opposed to the unbridled optimism notion suggested by main stream of brand community research (Manchanda et al., 2015; Kumar and Nayak, 2019). In fact, this research falls into an increasing body of studies suggesting that brand community poses increased risk in managing consumers as a group (Weijo et al., 2019; Herhausen et al., 2019; Thomas et al., 2013). Further, the results show that brand community commitment positively relates to brand commitment and plays a positive mediating role between group heterogeneity and brand commitment. How brand community engagement generates a positive consumer-brand relationship has consistently been an important research question. This finding suggests that brand community commitment is one of the important underlying mechanisms through which brand community creates brand loyalty. Finally, consistent with prior studies showing that members of varying tenure differ in many ways, such as having different participation focus (Tsai and Bagozzi, 2014), brand knowledge (Adjei et al., 2010) and membership legitimacy (Thompson et al., 2016), we find the impact of heterogeneity appears to be different for short-tenure members and long-tenure members. Specifically, the positive relationship between visible heterogeneity and brand community commitment exists only for short-tenure members, whereas the negative effect of value heterogeneity on brand community commitment is significant for all members but it appears to be stronger for long-tenure members compared to short-tenure members. This difference arises because short-tenure members' participation is at the surface-level. In other word, they care more about addressing problems and seeking information; and it takes time for them to integrate into a brand community and appreciate the community's deep-level aspects such as culture, value and norms. Long-tenure members are more experienced in product usage and community engagement as well. Rather than seeking informational benefits, they may desire a value-cohesiveness community. However, due to the open nature of OBCs, they can easily detect heterogeneity in value and thus be influenced by it.\n",
"role": "user"
},
{
"content": "Purpose: Despite the increasingly common view that online brand community (OBC) members are heterogeneous, knowledge concerning the impact of group heterogeneity on community and brand level outcomes is lacking. In response and drawing from organization research, this paper aims to study the consequences of two types of group heterogeneity (i.e. visible heterogeneity and value heterogeneity) on brand community commitment and brand commitment. The moderating role of tenure in a community is also examined.\nDesign/methodology/approach: A survey of 467 members of OBCs was conducted, and structural equation modeling was used to test hypotheses.\nFindings: The results show that perceived visible heterogeneity positively affects brand community commitment, whereas perceived value heterogeneity has a negative effect on it. Brand community commitment positively relates to brand commitment; it also mediates the effect of perceived visible heterogeneity and perceived value heterogeneity on brand commitment. Further, the positive effect of visible heterogeneity on brand community commitment is stronger for short-tenure members, but the negative effect of value heterogeneity is stronger for long-tenure members.\nPractical implications: The findings suggest that managers should make efforts to foster visible heterogeneity and reduce value heterogeneity. In addition, managers are advised to emphasize the characteristics that carry different appeal for members of different tenure.\nOriginality/value: This research is one of the first few quantitative studies to examine the influence of brand community heterogeneity on community, and especially brand level outcomes. It extends the literature on the effect of brand community on brands and adds to the emerging heterogeneity view of OBCs.\n",
"role": "assistant"
}
] |
Innovation types and performance in growing UK SMEs
|
[
"Innovation",
"Performance management",
"Small to medium‐sized enterprises",
"United Kingdom"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The importance of SMEs has been noted in many studies. For example, Birch (1989) demonstrated that two-thirds of the increase in employment in the USA between 1969 and 1976 had been in firms with fewer than 20 workers. Wolff and Pett (2006) argued that SMEs and entrepreneurial firms are a key segment and driver for most national economies. As such, many governments have been attempting to develop programmes to stimulate entrepreneurship. The governments of most developed economies see new and smaller enterprises as the well-spring of economic growth and job and wealth creation. Several Organization for Economic Cooperation and Development (OECD) member nations are discussing fiscal measures that promote innovation and research and development. Entrepreneurship and innovation have also been closely linked and are high on the policy agenda.The UK Government, in particular, is specifically attempting to encourage innovation in SMEs. A simple search of the web site of the UK Government's Department of Trade and Industry (www.dti.gov.uk) for the word \"innovation\" returned 500 entries. Within this government department, there is a specific innovation unit which is currently running a very large initiative called \"living innovation\". In support of this initiative they indicate that:There is no doubt that competitive pressures on UK businesses are increasing, most significantly from international, low cost nations. To encourage UK companies to develop new and improved added value products and services, processes and ways of doing business, the Innovation Group at DTI puts resources, help and guidance at their disposal (DTI, 2004).In the UK at least, it is clear that policy makers see an important link between innovation and business and economic growth. Innovation activities in SMEs have also been of interest to academics (Motwani et al., 1999; Cosh and Hughes, 2000; Oliver et al., 2000). However, studies about innovation management in SMEs are few compared with similar studies on large firms (Cagliano and Spina, 2002; De Toni and Nassimbeni, 2003; Mosey, 2005). Previous studies have argued that because of their nature, SMEs - particularly those operating in manufacturing, electronics, engineering and general high technology industries - are able to undertake radical innovation more easily than large firms and that introducing pioneering products is an important entrepreneurial activity and the lifeblood of SMEs (Kanter, 1985; Simon et al., 2002). However, this is not a fact without refute. Some argue that while a small firm in any of these industries may have high research and development intensity (R&D/sales), a larger firm with more slack can actually devote more resources to R&D. The assumption that SMEs focus more on radical innovations than incremental innovations has not been empirically researched. In general, studies on the types of innovation that SMEs pursue have been largely ignored in the literature. Therefore, the objective of this study is to empirically identify the types of innovation that are predominant in UK SMEs, whether these innovations are radical or incremental and to investigate their impact on innovation performance and business performance.The paper is structured as follows. First, we review the literature on innovation in SMEs and define the research questions and hypotheses. Second, we discuss the research methodology employed to carry out the empirical work. Third, the analysis is presented and discussed. Finally, the managerial implications of the results are discussed and the paper is then concluded with suggestions for future work.\nLiterature review: In order to provide a theoretical basis for the study, we review previous studies on SMEs in the operations management literature. We note a dearth of studies on innovation in SMEs. We then extend our review of studies on SMEs to other fields. We find that generally most innovation studies in SMEs fall within two categories - firstly, innovation in SMEs linked with performance and economic growth, and secondly, SMEs and innovation types. We find that the former stream of literature has attracted more studies. Unlike innovation studies in large firms, there is a dearth of studies relating to the type of innovations that SMEs pursue. We develop our research questions and hypotheses to contribute to this aspect of the literature.SME studies in the operations management literature\nMethodology: A quantitative empirical research approach was chosen to investigate the research questions. This is necessary to be able to statistically test some of the hypotheses, and be consistent with previous research for comparison. The design of the questionnaire used for this study was based on questionnaire items used in previous studies (Griffin, 1997; Goffin and Pfeiffer, 1999; Avlonitis et al., 2001; Oke, 2002, 2004). Specifically, the questionnaire investigated the following:1. Innovation types and innovative focus: three types of innovation were identified (product, process and service). For each type, respondents were asked to state the actual number of radical (described as introduction of a completely new product/process/service) and incremental (described as minor or major improvements to existing products/process/service) innovations that their businesses have introduced in the previous two years.2. Innovation performance. Respondents were asked to indicate the degree to which they agree or disagree (on a seven-point scale) with statements that seek to measure their innovation performance compared to competitors. For example, - our business tends to:* be better at time-to-market for new ideas;* be one of the first to market with innovative new products and services;* be perceived by our customers to be more innovative; and* have more new products/services in our portfolio at any one time, etc.3. Business performance. The most common measure of business performance as it relates to innovation in the literature is the amount of sales (or sales turnover) generated from innovations or new products (Griffin, 1997). Business measures such as return on assets, return on equity and the like have been largely ignored because of the difficulty in linking such measures with innovation activities. Therefore, we adopted two quantitative measures of business performance namely sales turnover and net profit before tax. Respondents were asked for the following information for each of the last three years:* sales turnover;* net profit before tax;Choice of survey method\nAnalysis and results: In total, 108 SMEs completed the survey on-line during an eight week period - the length of time the survey ran for. The sectoral distribution of the responding firms is similar to that of the sample and, as an extension, to the BGP database. It can, therefore, be argued that the dataset is representative of the population that was invited to take part in the survey and that the results are generalizable to the database population that is being looked at. Fifty firms responded in the first four weeks and, following a reminder, a further 58 firms responded in weeks five to eight. Overall, this represents a response rate of approximately 13.8 percent. Analysis has been undertaken for where data is available and a clear response has been given by the respondent. Analysis is therefore based on valid data.In order to ascertain the representativeness of the data and to test for non-response bias, analysis was undertaken of the 50 firms responding in the first time period. This dataset was then compared with the remaining 58 firms that completed in the second time period (Armstrong and Overton, 1977). This latter group would have been classified as non-respondents if the cut-off date had been at the end of the first time period. Analysis shows that there was no significant difference between the two sets of data on the key variables utilized in this article. Generalizability to the population of the database that is being looked at is thus further ensured.The respondents are very senior, with over two-thirds (73.1 percent) indicating that they are either a managing director or a chairman. In terms of the businesses themselves, firms responding are generally quite mature. 64.8 percent have been established for over 10 years. Of the firms that responded, 81.6 percent have up to 50 full-time equivalents working in the business.Performance outcomes\nAnalysis by firm characteristics: It could be argued that the results of this study may have been influenced by the characteristics of the SME, such as the age of the firm. It has been suggested that the development of radical innovations is one of the main drivers for bringing SMEs into existence. We would thus expect relatively newer SMEs to focus more on radical innovations than the older firms. Therefore, we carried out further analysis using the kh2 test to explore whether the innovative focus (radical or incremental) of innovations varies (Table XIII) with the ages of the responding firms (that is, how long the SMEs have been in business). Interestingly, we find no significant differences (kh2 value =3.799, p>0.05). In other words, whether the SMEs in our sample choose to focus on incremental or radical innovations is not significantly related to the length of time that the firm has been in business.Discussion and conclusion\n",
"role": "user"
},
{
"content": "Purpose: - The objective of this study is to explore the types of innovation that are predominant in UK SMEs, whether they are predominantly radical or incremental, and to investigate the impact of these innovations on performance.\nDesign/methodology/approach: - A web-based survey instrument was used to administer survey questionnaires to a sample of UK SMEs in manufacturing, engineering, electronics, information technology and telecommunications industries. The response rate was 13.8 percent. Relevant statistical analytical techniques including regression for analysis was then used.\nFindings: - It is found that the SMEs tend to focus more on incremental than radical innovations and that this focus is related to growth in sales turnover.\nPractical implications: - It is not such a bad idea for SMEs, particularly those operating in high technological industries, to focus on incremental innovations as these are actually related positively to sales turnover growth.\nOriginality/value: - An investigation of the types of innovation that SMEs pursue is pioneering in the field as previous studies of this type have been based on large firms. The contention that the raison d'etre of SMEs is to develop radical innovations is not supported empirically, at least for the sample of SMEs. In addition, it is found that the age of an SME is not related to its focus on either incremental or radical innovation. These are the contributions of this paper.\n",
"role": "assistant"
}
] |
Creating a sustainable university and community through a Common Experience
|
[
"Common Experience",
"Sustainability",
"University outreach",
"Community involvement",
"Instructional innovation in sustainability",
"Sustainability education",
"Dispositional thinking",
"Teaching",
"Planning",
"Academic lifecycle",
"Education",
"Universities"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: College and university leaders recognize that progress toward finding solutions to sustainability problems requires engagement and input from all stakeholders within their campus and surrounding community through shared vision and face-to-face dialogue (Pollock et al., 2009). This article provides an overview of Texas State University's Common Experience, an innovative initiative that engaged tens of thousands of people in shared consideration of sustainability as a single topic during academic year 2010-2011.Texas State University is a public institution of higher education with over 34,000 students pursuing 97 bachelors, 87 masters, and 12 doctoral degree programs offered through its nine colleges. The main campus is located in the city of San Marcos, a growing community of 50,000 people about halfway between Austin and San Antonio, along the Interstate Highway 35. The area sits on the edge of the Texas Hill Country, where black land prairies roll into beautiful hills. Stately cypress and pecan trees on main campus provide a picturesque setting, and its location on the banks of the crystal-clear San Marcos River affords recreational activities for students throughout the year. The university's students represent people from around the world and are highly diverse, in that 35 percent are ethnic minorities. In 2011, the institution achieved federal recognition as a \"Hispanic Serving Institution\" based on its student demographics reaching 25 percent Hispanic origin. Texas State ranks 14 in the nation for the number of bachelor's degrees awarded to Hispanic students.In 2004, Dr Denise M. Trauth, president of Texas State University, announced the \"Common Experience\" as an initiative intended to connect first-time entering students to a common intellectual conversation across the campus. The initiative grew quickly, however, to include upper-level classes, student support services, campus activities, performing arts, and the San Marcos schools and community. The Common Experience thus became a yearlong initiative designed with the objectives to cultivate a common intellectual conversation across the campus, to enhance students' participation in the intellectual life of the campus, and to foster a sense of community across the campus and extended community.The Common Experience theme for academic year 2010-2011 was \"Sustainability: Science, Policy, and Opportunity,\" which examined the integrative nature of economic, environmental, and social problems faced by humanity today at local, regional, and global levels. Through interdisciplinary dialogue, outreach, and education, the Common Experience provided the Texas State community with opportunities to embrace and actively participate in the science and practice of sustainability. The following is an overview of the Common Experience in sustainability that inspired people on campus and in the community to change behavior and social policy, build scientific understanding, connect local environmental issues to global themes, and reshape values - components crucial to sustainability education.\n2 The common experience - theoretical framework: From a pedagogical perspective, the Common Experience is an instructional methodology based on five principles of effective instructional practice identified by the National Research Council (NRC) in their synthesis of the research from the fields of cognitive, developmental and educational psychology, and brain research on how people learn.2.1 Learning builds on people's previous experiences\n3 Method: 3.1 Case study research design, data collection, and analysis\n4 The common experience - sustainability: The NRC (2000) principles provide a theoretical framework for understanding why the Common Experience has the potential to create a sustainable university and community. This section describes the content of the yearlong program intended to inspire people on campus and in the community to change behavior and social policy, build scientific understanding, connect local environmental issues to global themes, and reshape values - components crucial to sustainability education.4.1 Course-based common experience activities\n5 Common experience - location, time, and duration of non-course activities: 5.1 Location of non-course common experience activities\n6 Common experience - planning for non-course activities: 6.1 Collaboration, coordination, and attention to details\n7 Conclusion: This article provided an overview of how a university engaged tens of thousands of people in shared consideration of sustainability as a single topic during academic year 2010-2011. Based on five principles of instructional practice, the Common Experience is a pedagogical innovation designed to cultivate a common intellectual conversation across the campus, to enhance students' participation in the intellectual life of the campus, and to foster a sense of community across the campus and extended community. Thus, the Common Experience is more likely to yield progress toward finding solutions to sustainability problems because it promotes engagement and input from all stakeholders within the campus and surrounding community through shared vision and face-to-face dialogue in a variety of contexts. The implication is that the Common Experience has the potential to empower institutions with the ability to engage with its community in meaningful ways that contribute to the quality of life, and the economic and social development of the region.The Common Experience brought students together, for example, to read and engage with a powerful text related to sustainability, explore related themes in university seminar classes, and write responses and reactions to the text in writing courses. Faculty also integrated where feasible the Common Experience theme of sustainability into their curriculum, while academic advisors encouraged students to take these courses during the regular fall and spring semesters. In comparison, the hallmark of the Common Experience was non-course activities. While course-based Common Experience activities were limited to enrolled students in the university, non-course Common Experience activities significantly expanded the opportunity for engagement in intellectual consideration about sustainability that reached across the campus into the community and beyond. The implication is that the Common Experience provides institutions with a wide range of venues for educating people about sustainability - on campus and far beyond into the wider community.There are also implications for future research. From developing this case study, the researcher formulated recommendations intended to advance the implementation of similar programs at other institutions. When planning for a non-course Common Experience activity, for example, organizers must consider its place, time, and duration, as well as the intended audience. They should also consider the institution's academic lifecycle defined by its own schedule of events (e.g. academic start and end time, breaks, and holidays) and context (e.g. location and climate factors). Lastly, non-course Common Experience activities do not happen without the collaborative and coordinated support of people from various levels of the institution and the community. Their primary goal is to identify, design, and develop common experience activities that would be of interest to a wide range of participants. In seeking this goal, the following question guides organizers: \"What experiences do we intend to give participants, relevant to the Common Experience theme of sustainability?\" Once selected, the activities must then undergo a public relations campaign to motivate people to attend and participate. Organizers should consider their institution's public relations department or mass communication program to help identify suitable titles and descriptions for common experience activities that result in high attendance among faculty, staff, students, and people from the community. The implication of this result for researchers is the opportunity for scholarship that examines and compares how different activities change people with regard to their cognitive and affective learning relevant to sustainability thereby, producing research that contributes significantly to the field.The Common Experience in sustainability ended in April 2011, but its effect on the Texas State campus and in the San Marcos community endures. Texas State continues to offer its students programs and courses in sustainability, and many non-course Common Experience activities, especially those that were project-based, continue with the funding support of a student environmental fee. Texas State's web site (www.fss.txstate.edu/sustainability/) provides detailed information on these projects and other new sustainability initiatives on campus and in the community.\n",
"role": "user"
},
{
"content": "Purpose: - This article aims to provide an overview of Texas State University's Common Experience, an innovative initiative that engaged tens of thousands of people in shared consideration of sustainability as a single topic during academic year 2010-2011.\nDesign/methodology/approach: - The discourse begins with an overview of the Common Experience from a pedagogical perspective based on five principles of effective instructional practice identified by the National Research Council (NRC) in their synthesis of the research from the fields of cognitive, developmental and educational psychology, and brain research on how people learn. Based on a case study research design, the article next provides detailed descriptions of course-based and non-course Common Experience activities, including examples from the year-long initiative. The discourse then focuses on important elements organizers should consider when planning non-course activities, followed by a description of the planning process needed to replicate the initiative at other institutions.\nFindings: - Based on five principles of instructional practice, the Common Experience is a pedagogical innovation designed to cultivate a common intellectual conversation across the campus, to enhance students' participation in the intellectual life of the campus, and to foster a sense of community across the campus and extended community. In so doing, the Common Experience in sustainability inspired people on campus and in the community to change behavior and social policy, build scientific understanding, connect local environmental issues to global themes, and reshape values - components crucial to sustainability education.\nOriginality/value: - The Common Experience is more likely to yield progress toward finding solutions to sustainability problems because it promotes engagement and input from all stakeholders within the campus and surrounding community through shared vision and face-to-face dialogue in a variety of contexts.\n",
"role": "assistant"
}
] |
Implementation model for cellular manufacturing system using AHP and ANP approach
|
[
"Analytical hierarchy process",
"Analytic network process",
"Cellular manufacturing system",
"Implementation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Cellular manufacturing (CM) is an approach for enhancing operations of job shop and batch shop production. CM is used to achieve the advantages of a product-oriented production system for medium volume, medium variety condition by preparing a group of parts on a group (cell) of machines. CM, which clusters machines, is dedicated to the part family of similar components (Dekkers, 2018). Although CM does not have the flexibility of job shops in making an extensive variety of items yet, it has a high production rate and efficient flow (Liu et al., 2018). In the present batch production situation, demand for the product is considered by endless varieties concerning volumes, product mix and new product development. CM increases productivity, delivery performance by reducing work-in-progress inventory and lead times, hence helping manufacturing industries to be more competitive. There are significant benefits that can be accomplished as a result of implementing a cellular manufacturing system (CMS). These include reduced handling of materials, reduced tooling and equipment, reduced setup time, reduced work-in-process inventory, reduced part makespan and enhanced operator capability (Sakhaii et al., 2013). Changing over the process, from batch production to CM system, is crucial to address the challenges of a worldwide economic crisis (Garbie, 2011). CMS shows the enhanced performance than the continuous shop and job shop manufacturing system, in fulfilling the demand for mid-variety and mid-volume of products.\n2. Literature review: The idea of CM was first proposed by Burbidge (1979). A manufacturing cell can be well defined as a group of functionally different machines, placed together on the floor, devoted to the manufacture of a family of similar parts (Ham et al., 1985). CM has been seen as a standout amongst the most inventive approaches to enhance flexibility and productivity for today's manufacturer, for diverse product range and low volume production, since it can reasonably change batch-type production into line-type production. Evidence shows that firms on their journey to transform to CM often struggle with implementation (Wemmerlov and Johnson, 2000; Yauch and Steudel, 2002). Few researchers have proposed models for the implementation of CM that emphasis on a particular area of the implementation process (Hyer, 1982; Sambasivarao and Deshmukh, 1995; Badham and Couchman, 1996; Olorunniwo, 1997). Well-designed cells can yield productivity (Irani, 1999). Afzulpurkar et al. (1993) emphasized on technical issues that could critically impact the successful implementation of CM. In an investigation of performance obstacles in CM implementation in a manufacturing plant, Park and Han (2002) found that critical factors were employee training, teamwork, supervision and scheduling. Fraser et al. (2007) suggested a sequential model of six-phase CM implementation.\n3. Methodology: The research methodology includes two stages. The first stage involves the selection of success factors by literature survey and discussion with experts. The second stage involves assigning weights and ranking factors using the AHP and ANP methods. The framework for this research work is illustrated in Figure 1.\n4. Model development: The entire problem of modeling the factors into hierarchy and network, pairwise comparison, inconsistency check and ranking of factors were carried out using the \"Super Decision\" software developed by Creative Decisions Foundation.\n5. Sensitivity analysis: Sensitivity analysis was used to explore how sensitive the rankings of the criteria and their related sub-criteria are to change, if the weights of the criteria and its related sub-criteria are subjected to change. Toward this end, the percentage of each criterion was increased by 10 percent and after that, for the sub-criteria, it was disseminated equally so that the total sum is kept balanced, and then figuring out the changes in ranking (Abd El-Naby, 2015). It may be noted that, in the sensitivity analysis, one factor is modified at a time, while other factors are maintained unaltered to perceive what the effects or enhancements happen to the modified factor and its related sub-factors.\n6. Results and discussions: The study attempts to identify critical factors to implement CM successfully and analyze the importance of factors by applying AHP and ANP techniques. The study is based on interaction with five experts from academia and industry. Essential factors have been identified from the literature and validated by experts. A total of 17 factors of CM implementation have been categorized in 4 dimensions.\n7. Conclusions and implications of the research: CM is presently getting significant practical consideration. The organizations need to pay careful attention to choose CM implementation factors, as implementing all of them simultaneously can become a massive task. This study utilized AHP and ANP methodologies to prioritize essential CM factors which are further validated through sensitivity analysis. Initially, all possible factors were identified from the literature and discussed with five experts from academics and industry.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to visualize the prioritization among essential factors of cellular manufacturing system (CMS) implementation using the analytic hierarchy process (AHP) and analytic network process (ANP) methods.\nDesign/methodology/approach: Based on literature review, 4 enabler dimensions and 17 CM factors were identified which were validated by experts from academia and industry. Then, AHP and ANP models are proposed in evaluating CMS implementation dimensions and factors. The results are validated using sensitivity analysis.\nFindings: These models give firms a straightforward and simple to utilize way to deal with CMS efficiently. The two strategies were appeared to be powerful in choosing a strategy for CMS implementation. The two strategies brought about nearly similar outcomes. Both methods consider the particular necessities of the organization through its own accessible ability.\nPractical implications: The techniques exhibited in this paper can be utilized by a wide range of organizations for adopting CMS that have a higher impact on performance and thus overall productivity. The two techniques are explained in a step-by-step approach for easier adoption by practitioners.\nOriginality/value: The strength of the present study is that it is one of the first few to be conducted in perspective for CM implementation factors analysis.\n",
"role": "assistant"
}
] |
The relationship between continuous improvement and maintenance performance
|
[
"Maintenance",
"Quality management",
"Continuous improvement",
"Maintenance performance"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The contemporary global marketplace has been putting enormous pressures on the manufacturing organizations to continuously adapt proactive, innovative strategies for enhancing their manufacturing capabilities (Ahuja and Khamba, 2008). To meet the challenges posed by the contemporary competitive environment, the manufacturing organizations must infuse quality and performance improvement initiatives in all aspects of their processes to improve their competitiveness. Furthermore, in contemporary highly challenging environment, a reliable production system has been considered as a crucial factor for competitiveness (Pintelon et al., 2006). Thus, for achieving effectiveness, maintenance has to be treated as a strategic issue in manufacturing organizations and to make its proper contribution to profit, productivity, and quality, it must be recognized as an integral part of the plant production strategy (Kumar et al., 2004). Al-Najjar (2007) stated that applying effective maintenance aims to enhance organization's profitability and competitiveness through continuous cost-effective improvement of production process efficiency, effectiveness and productivity, which can be achieved with maintaining and improving the quality of all the elements contribute in the production process continuously and cost-effectively. One of the existing quality initiatives for achieving competitive excellence is, therefore, continuous improvement (Oakland, 1999). Given this increasing role as a driver of competitive advantage, there is a growing body of research on this subject. Terziovski (2006) found out that the most important predictor of productivity improvement was pursuit of continuous improvement (CI). In a review of the literature covering the relationship between quality management practices and organizational performance Lakhal et al. (2006) identified a positive relationship between them. Furthermore, several studies also identify the importance of CI, as a part of quality management practices, in the sense of its contribution to the organizational performance (Kaynak, 2003; Sila, 2007; Terziovski and Samson, 1999; Cua et al., 2001; Douglas and Judge, 2001). Likewise, continuous improvement contributes to theorganizational efficiency and effectiveness considering the total productive maintenance (TPM) initiatives (Cooke, 2000). While several studies look at CI in the light of organizational performance, there is still lack of studies that would emphasize the role of CI on maintenance performance. Therefore, using empirical data collected from Slovenian organizations, this study attempts to address the relationship between CI and maintenance performance. Thus, the overall purpose of this paper is to contribute to a better understanding of the integration of continuous improvement into maintenance processes, especially from the maintenance performance point-of-view.\n2. Literature review: 2.1 Continuous improvement in maintenance\n3. Methodology: 3.1 Sample\n4. Results: 4.1 Construct validity and reliability\n5. Discussion: In this study we have provided empirical evidence that CI has a positive impact on maintenance performance. As shown by the regression results (Table III), CI is an important predictor of maintenance performance (Beta=0.819, p=0.000). From theoretical perspectives, this finding could be interpreted in the light of the organizational performance. One explanation of this could the argument, that an effective maintenance could influence the profitability of a manufacturing process through its direct impact on quality, efficiency and effectiveness of operation (Alsyouf, 2007). Having this in mind, we can support our result by the findings of studies (Cua et al., 2001; Douglas and Judge, 2001; Kaynak, 2003) which have shown the positive relationship between quality management practices and organizational performance. CI is in organizations often introduced as part of quality management systems, such as ISO 9001. Since its advent, ISO 9000 has also been an important research topic within management literature. In particular, previous literature has evaluated the relationship between ISO 9000 certification and organizational performance and reported mixed results (Prajogo, 2011). For example, studies of Terziovski et al. (1997) and Singels et al. (2001) found no evidence of a link between ISO 9000 and organizational performance. On the other hand, Feng et al. (2008) found a positive and significant relationship between ISO 9000 and operational performance. However, based on the results of this study (Table IV), we did not find the evidence that organizations with ISO 9001 compliance, achieve better results in terms of maintenance performance. Nonetheless of this finding, we believe that if organization focuses on CI solely as systematic approach for striving to achieve improvements in maintenance processes, these results in better maintenance performance.The findings of correlation analysis provide further understanding of the variables, with respect to their correlation with maintenance performance. As evidenced by the correlation analysis presented in Table II, tools for CI and information concerning the quality of processes are the most significant activities regarding maintenance performance. This is in line with study of McKone et al. (2001), in which they found a positive relationship between TPM measures, among which were also teams and information tracking, and manufacturing performance. The finding about information flow could also be supported by the study of Seth and Tripathi (2005) in which they revealed a positive relationship between information architecture (information and data needed to support key processes and performance improvement) and performance factors. The findings, therefore suggest that information provides necessary infrastructure to facilitate decisions in right direction. Furthermore, positive result on the relationship between the tools for CI and maintenance performance support the study of Duffuaa and Ben-Daya (1995) in which authors highlight the idea of applications quality tools in different maintenance activities in order to improve maintenance quality.As evidenced by the correlation analysis presented in Table II, teams are shown to have a significant and positive relationship with maintenance performance. The results is consistent with the finding of the literature stream which considers team work as an encouragement of participation of all employees, including the operatives who are ideally placed to make a significant contribution to improving performance (Bond, 1999). Furthermore, the result also support the discussion in TPM literature concerning the role of TPM in achieving manufacturing performance, where teams are considered as an important component in achieving high availability, production efficiency and better quality (Sharma et al., 2006). As such, organizations with team based maintenance strategy will perform better. This could be also argued for a management commitment, which also appears to have a significant role in maintenance performance, according to our results from the correlation analysis. The finding, therefore, implies the importance of management commitment to CI in achieving high maintenance performance, which can be interpreted by the argument, that management commitment is one of the key factor in implementation of the TPM, as well as TQM (Hansson et al., 2003). The results presented above are also consistent with finding that principles of TPM are positively correlated to maintenance performance. In particular, the finding support the notion that TPM could benefit to the maintenance performance. This is in line with the assertion that TPM describes a synergistic relationship among all organizational functions, particularly between production and maintenance, for the continuous improvement of product quality, operational efficiency, productivity and safety (Labib, 1999).The results of PCA correspond to the findings of Factor Analysis, indicating that proposed variables correctly reflect the measure of CI. Furthermore, finding of PCA, support the results of the correlation analysis. As it is shown by results of PCA (Figure 2), all variables of CI are strongly correlated. By comparing the score and loading plot, we can also identify the relationships between samples and variables. Based on the position of the observed cases (marked with numbers), one can infer at which cases the variables (CI) were correlated. Moreover, we can also link the findings of PCA with maintenance performance measure. It appears that organizations with a high level of maintenance performance (see number 5, 15, 22, 32, 34, 37, 39, 41, 42, 50) show significantly higher scores on the first principal component in comparison to organizations with low maintenance performance (i.e. 10, 13, 16, 44). This leads us to conclusion that integrating CI practices in maintenance processes is considered to be an important element in achieving higher maintenance performance. However, one can also easily identify outliers, i.e. organizations that have different characteristics in comparison to other organizations in the same group (i.e. 19, 37).From a managerial perspective, the study emphasizes the need to recognize the importance of CI activities, particularly their roles in in?uencing the outcomes of the implementation of CI into maintenance processes. Two practical insights can be drawn here. First, it is important for managers to recognize the role of CI activities in the field of maintenance and understand their effect in enhancing the maintenance performance. Second, from the results we can infer that positively correlated CI activities could reflect in better maintenance performance and could therefore also represent a benefit to manufacturing performance. This could lead to competitive success, and from that perspective we acknowledge our findings also as a practical contribution for the organizations.\n7. Conclusion and future directions: The results supported empirically the theoretical assertions made in the study. Therefore, in response to the purpose of this paper, the findings provide empirical evidence that CI significantly and positively contributes to maintenance performance, in terms of maintenance processes efficiency and effectiveness. CI is shown to be an effective way of improving maintenance performance. Although continuous improvement is widely accepted approach in the field of quality management, it is still in the development in the evolution of maintenance management theory (mainly in the context of maintenance performance) and practice. As such, the emphasis is currently on the role of CI as a part of TPM practices and its correlation with manufacturing performance, rather than on the creation of direct relationship between CI and maintenance performance.There are a few limitations of the study that present directions for future investigations in this domain. Despite the fact that our study results in valuable insights into CI - in particular in the maintenance performance context - there are some limitations. First, our study focused on one specific factor that may influence maintenance performance. Future research could broaden the investigation to identify more complex measures of CI. It could also be focused on moderating effects of various contextual factors (such as quality orientation) on the relationship between CI and maintenance performance. Second, our sample size should be expanded to a larger group in order to increase the generalizability of the results.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to present an empirical study, which examines the relationship between continuous improvement (CI) and maintenance performance. By examining CI in relation to maintenance performance, this study seeks to increase the understanding of quality management practices in the field of maintenance.\nDesign/methodology/approach: - The empirical data for this study were drawn from a survey of Slovenian organizations in order to address the research problem. Several statistical methods including correlation analysis, regression analysis as well as principal component analysis (PCA) are utilized to accomplish the objective of the study.\nFindings: - The findings suggest that CI significantly and positively relates to maintenance performance. In addition, findings advocate the importance of incorporation of quality management practices into maintenance processes.\nResearch limitations/implications: - Based on the findings of this study, it is concluded that future research could broaden the investigation to identify more complex measures of CI. Moreover, sample size should be expanded to a larger group in order to increase the generalizability of the results.\nPractical implications: - The study contributes to a better understanding of CI activities in the field of maintenance and therefore provides insights for managers to recognize the role of CI activities and understand their effect in enhancing the maintenance performance.\nOriginality/value: - The findings provide empirical evidence that CI is shown to be an effective way of improving maintenance performance.\n",
"role": "assistant"
}
] |
Has the UK court created a lease guarantee nightmare?
|
[
"United Kingdom",
"Leasing",
"Legislation",
"Higher courts",
"Landlord",
"Tenant",
"Guarantee",
"Assignment",
"Liabilities",
"Release"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nHas the UK court created a lease guarantee nightmare?: In an extraordinary move, the Court of Appeal in K/S Victoria Street v. House of Fraser (Stores Management) [2011] EWCA Civ 904 elected to go beyond the issues required to reach judgment. The court decided to provide guidance on the effect of Landlord and Tenant (Covenants) Act 1995 on the liability of former tenants and guarantors. That guidance certainly provided clarity, but may have created significant difficulties and adverse valuation effects for commercial landlords.The Court of Appeal ruled that where a lease granted on or after 1 January 1996 created a \"new\" tenancy for the purposes of the 1995 Act, the tenant is entitled to a release from liability when it assigns the lease and that its guarantor is entitled to a consequential release from its liability. To ensure that the act is not excluded or frustrated, the court ruled that a guarantor cannot be required to provide a guarantee in respect of the assignee. There was absolutely no surprise in that conclusion.The Court of Appeal provided useful confirmation that the guarantor of an outgoing tenant is able to give a \"sub-guarantee\" (AGA), standing behind the new obligations taken on by the outgoing tenant in an \"authorised guarantee agreement\". That provides landlords with the means to retain the covenant strength of the guarantor in circumstances where the outgoing tenant is an insubstantial covenant - a common situation in retail leases.The real sting in the guidance was the Court of Appeal's view that a guarantor cannot itself take a valid assignment of the lease. That conclusion potentially affects a large number of transactions effected since the 1995 Act came into force, in some cases arguably creating a void lease, in others causing significant doubt about the enforceability of covenants.Unless and until the relevant issues reach the Supreme Court the guidance provided by the Court of Appeal, though obiter, must be regarded as conclusive.\nWhat is an AGA: The legislative background is the compromise secured by commercial landlords during the parliamentary process leading to the 1995 Act. In its original form, the 1995 Act provided an automatic release from liability for a tenant on assignment, and a consequential release for its guarantor. Concerned that this would deprive them of covenant strength, crucial for valuation, landlords obtained the ability to call for an AGA requiring the outgoing tenant to guarantee performance by the assignee it has introduced into the relationship.An AGA may be a separate document or included in a licence to assign. In either case, it must be in writing and signed by the guarantor to be enforceable (Statute of Frauds, 1677, s 4).An AGA must:* be an agreement where the tenant guarantees the performance by the assignee of the covenants from which the tenant has been released. Where a tenant assigns in breach of covenant or by operation of law, the tenant has not been released; if this is the case, the tenant and the assignee will be jointly and severally liable and an AGA will not have come into existence; and* be entered into where the lease contains a covenant against assignment without the landlord's consent and consent will only be given on the condition (which must be lawfully imposed) that the outgoing tenant guarantees the performance of the tenant covenants by the assignee. This condition is lawfully imposed if set out in the lease.Where a lease is silent, a reasonableness test applies, dependent upon the assignee's covenant strength. The Code for Leasing Business Premises in England and Wales 2007 states that AGAs should only be required where the proposed assignee is of lower financial standing than the outgoing tenant (and its guarantor) at the date of the assignment.\nThe contents of an AGA: An AGA may:* require the tenant to become liable under the AGA as a sole or principal debtor. This means that its liability is separate from and in addition to the assignee's liability under the lease. If there is a breach of covenant, the landlord can take immediate action against either the tenant or the assignee;* require the tenant to assume liabilities as guarantor, which are no more onerous than those to which he would be subject if he were liable as sole or principal debtor under the lease. Tenants should resist any attempts to widen the indemnity given under the lease;* require the tenant, in the event of disclaimer, to enter into a new lease of the premises. The new lease cannot expire any later than the disclaimed lease and must be no more onerous. A guarantor's liability to a landlord under an AGA continues despite the insolvent tenant's liquidator disclaiming the lease. Following Shaw v. Doleman [2009] All ER (D) 34 (Apr) this is even the case where the guarantor's liability is limited to the period during which the assignee was \"bound by the tenant's covenants of the lease\". Parties intending for the AGA to determine upon disclaimer must expressly provide for it to do so; and* include provisions which are \"incidental or supplementary\" to the above three points (e.g. provisions to prevent release on variation following assignment of the lease).Most guarantee covenants in leases (non-AGA ones) oblige the guarantor to take a new lease in the event of not only disclaimer but also forfeiture and liquidation. LT(C)A 1995 only refers to disclaimer so attempts to widen this under an AGA may fall foul of the anti-avoidance provisions contained in section 25. Section 25 states that any agreement relating to a tenancy will be void to the extent that it attempts to exclude, modify or otherwise frustrate the operation of any provision of LT(C)A 1995. If an AGA fails to follow section 16, it may be held to be void. It is good practice to include a non-severance provision in an AGA; this allows the offending term to be removed and the rest of the agreement remains valid.\nCan an outgoing tenant's guarantor be forced to give an AGA in place of the outgoing tenant?: In some instances, a landlord might prefer to take an AGA from the tenant's guarantor as opposed to the tenant. However, s 16(2) provides that the outgoing tenant must give the AGA. Consequently, an agreement requiring the guarantor (as well as or instead of the tenant) to give an AGA would fall foul of the anti-avoidance provisions in s 25.In K/S Victoria Street v. House of Fraser (Stores Management) [2011] EWCA Civ 904, the Court of Appeal confirmed that a tenant's guarantor cannot agree to give, nor give, a guarantee for the tenant's assignee other than by guaranteeing the tenant's obligations under an authorised guarantee agreement. It acknowledged that this approach meant that a lease could not be assigned to a guarantor, even where both tenant and guarantor want it. This closes off another approach which was not considered to be conclusively barred by Good Harvest v. Centaur Services [2010] EWHC 330 (Ch); assignment to a guarantor who then gives an AGA on assignment to the third-party assignee.\nA void lease?: It is worth emphasising the point: under no circumstances can a guarantor take an assignment of the lease. Intention is irrelevant. The parties may have perfectly good commercial reasons for entering into such an agreement, but the 1995 Act prevents them from doing so.The logic is that if the guarantor could take an assignment of the lease then it would be possible to circumvent the release provisions of the 1995 Act. Indeed, many commercial leases entered into since 1 January included provisions allowing the landlord to withhold consent to an arms-length assignment unless the lease was first assigned to the outgoing tenant's guarantor. The rationale was the guarantor would become, albeit for only a moment, the tenant and so would be able to give an AGA in respect of the ultimate assignee.The Court of Appeal guidance creates an alarming possibility. Where a lease has been assigned to a guarantor the provisions of the 1995 Act are breached. The guarantor does not get its release from liability. However, the breach cannot be cured by simply avoiding the assignment. That would push the lease back to the tenant, preventing its release from liability. The only way to arrive at a release for both is to avoid the lease.The result adds an extremely significant new layer to due diligence in commercial property matters. Before the Court of Appeal's guidance, there was a widespread view that the 1995 Act affected executory agreements (i.e. agreements that remain to be performed) and not executed agreements. It was thought that if a guarantor actually entered into a new guarantee it would be valid, though the guarantor could not have been held to an agreement to provide it. Similarly, if a guarantor accepted an assignment of the lease then there seemed to be no reason to doubt the validity of the assignment once completed.Judgments are declaratory. The Court of Appeal's guidance does not change the law. It states the law as it has been since 1996. Consequently, it is entirely possible that guarantors have taken assignments of leases, whether as a stage in an arms-length assignment or for a commercial reason wholly unconnected with the 1995 Act, such as a group restructuring. There is no reason for them to have been aware at the time of the assignment that the result could be to avoid the lease.Equally, an arms-length assignee receiving a lease that was first assigned from outgoing tenant to guarantor now faces the possibility that they received nothing.For landlords who consented to an assignment to, or through, a guarantor there is now the live risk that they consented to a transaction that rendered the lease void.On due diligence, it is now essential to check that no tenant in a chain where the lease has been successively assigned was guarantor for its immediate predecessor.\nSubleases: The implications go further. If the lease that was assigned was itself subject to one or more subleases, the status and survival of those subleases must be in doubt. Law of Property Act 1925, s 139 preserves a sublease and makes its covenants directly enforceable between the landlord and subtenant where the lease is surrendered or merged into the reversion. If a lease becomes void because of conflict with the anti-avoidance provisions of the 1995 Act, it certainly cannot be said to have been surrendered. There is no consensual act of surrender. Consequently, survival of the sublease and enforceability of its covenants depends on the result being analysed as a merger.The alternative would be to adopt a fiction similar to that applied when a lease is disclaimed by a liquidator. The lease ends unless there are other parties whose rights depend on its continued existence in some form - for example, a subtenant or mortgagee. The result, unless and until a vesting order is made, is the continuation of a \"phantom\" lease, stripped of liabilities but kept alive simply to preserve others' legal relationships. While that might be a pragmatic solution, it complicates due diligence and injects a significant measure of uncertainty into the valuation and pricing of a property investment.Continued occupation - trespass or new tenancy?\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to highlight the practical difficulties resulting from the UK Court of Appeal's recent guidance on the effect of anti-avoidance provisions in the Landlord and Tenant (Covenants) Act 1995.\nDesign/methodology/approach: - The paper analyses the Court's guidance, focusing on the inability of a guarantor to take a valid assignment of a lease from the outgoing tenant.\nFindings: - The paper identifies significant implications for due diligence, valuation and pricing of property investments, arguing that the result could be to deprive landlords of access to covenant strength that is crucial to its valuation or to any sale price.\nOriginality/value: - The paper reflects practitioner discussions since the Court of Appeal gave its ruling in K/S Victoria Street v. House of Fraser. It highlights significant new issues that must be factored into property due diligence.\n",
"role": "assistant"
}
] |
Towards quality consumption: Segmentation of the Hungarian wine market based on wine consumption and purchasing habits
|
[
"Market segmentation",
"Consumer behaviour",
"Survey research",
"Wine consumption",
"Wine purchasing",
"Wine-related preferences and behaviour",
"Hungary"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Hungary was 23rd among the world's wine producers as measured by vine area, with 67,000 hectares; it was 15th in wine production, with 2.5 million hectolitres in 2016 (just after New Zealand, OIV, 2018). The country was 24th in wine consumption, with 2.1 million hectolitres in 2016 (OIV, 2018), indicating a 26-litre per capita consumption annually (OIV, 2017). These data show that Hungary is self-sufficient in terms of wine production, with a low volume of both wine exports and imports.\n2. Literature review: 2.1 Segmentation of the wine market\n3. Research methodology: We conducted an online survey using convenience sampling, measuring wine consumption, buying habits and wine-related preferences of Hungarian consumers. The questionnaire was divided into four sections: wine consumption habits (frequency, place of consumption and consumption of wine with soda), wine-related preferences (favourite wine types and selection aspects), wine buying habits (stripping, wine territories, regions, buying frequency, value, number of pieces, purpose and location) and demographics questions (gender, year of birth, qualification and type of settlement). Data collection took place from April to July 2016. The questionnaire drew from two previous domestic surveys held in 2011 (Harsanyi, 2012) and 2014 (Harsanyi and Galne, 2015) and a Hungarian-German research (Harsanyi and Szolnoki, 2014). Some adjustments were made in the online questionnaire after pilot testing to correct minor mistakes and make the questions as clear, accurate and obvious as possible. Responding to survey questions was motivated by potential small value incentives.\n4. Research results: The demographic characteristics of the sample were presented in Table II. People with higher education, those living in larger cities and those with higher incomes are slightly overrepresented in the sample compared to the Hungarian population over 18 years of age. This can partly be explained with the online data collection method.\n5. Conclusions: Our study identified four wine consumer segments based on the consumption, purchasing habits and preferences using a large sample. Ordinary wine consumers (51 per cent) prefer domestic wine, drink wine more often with soda water and most frequently consume dry wine. This is the largest segment and its members spend an average amount of money per bottle.\n6. Practical implications, limitations and recommendations for future research: Despite the relatively small market size, Hungary has a long tradition in wine production. It has 22 wine regions and more than a thousand wineries. It is very difficult to succeed on the market; therefore, marketing and segmentation are essential. The results provided by our research deepens the understanding of the segmentation of the Hungarian wine market and promotes a more target-oriented design for local and foreign wine producers' marketing strategies. As Kotler and Keller (2011: 213) state: \"Identifying and satisfying the right market segments is often the key to marketing success.\" Brunner and Siegrist (2011, p. 369) add:\n",
"role": "user"
},
{
"content": "Purpose: Hungarian wine culture has been developing steadily over the past decades. The number of wineries offering quality wine is growing as consumer interest in quality consumption increases. The purpose of this paper is to study the segmentation of the Hungarian wine consumers based on their purchase habits and preferences to identify how this shift towards quality consumption is represented in these segments.\nDesign/methodology/approach: An online survey was conducted in Hungary. Nearly 28,000 consumers filled in the questionnaire. The TwoStep Cluster analysis could handle the large database and variables measured on different measurement scales.\nFindings: The growing interest towards quality wine in the market is reflected in the study's segmentation results. The large sample size made it possible to identify a special Hungarian wine consumer group, which has different buying habits than traditional wine consumer segments. Four segments were evolved: ordinary wine consumers, unsophisticated wine consumers, wealthy wine-experts and open-minded consumers. The last two segments seem to be the most valuable groups for high-quality wine producers.\nOriginality/value: The sample size allowed a relatively novel segmentation, whereby the preferences and purchasing habits of smaller, high-quality wine consumer groups became measurable.\n",
"role": "assistant"
}
] |
Students' perceptions and behavior toward on-campus foodservice operations
|
[
"Satisfaction",
"Student perceptions",
"HBCU",
"Historically black colleges and universities",
"On-campus food service"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: University foodservice is one of the largest segments of the foodservice industry globally, with an increasing number of the captive college-student market using the foodservice outlets on campus during their studies (Garg, 2014). In 2014, there were 11.8m college and university students under age 25 and 8.1m students 25 years old and over - a significant increase over previous decades (US Department of Education, National Center for Education Statistics, 2016). Consequently, university students' perceptions of their on-campus foodservice operation are crucial to influencing their levels of satisfaction, dining frequency and overall college experience (Hall, 2014). With the reported enrollment growth, economic conditions and the captive nature of the college-student foodservice market, it has been increasingly challenging to achieve student satisfaction with on-campus foodservices among diverse student groups, particularly on campuses with restricted foodservice options. Moreover, foodservice providers face many challenges serving the Millennial students that are culturally diverse, fickle, sophisticated, experienced in their dining habits and often confined to campus, thus making their needs more complex (Choi et al., 2013; Joung et al., 2014). The College and University Consumer Trend Report (Technomic, 2011) indicates that only 28 percent of American students surveyed were satisfied with the healthy menu options on campus, and only 34 percent expressed overall satisfaction with their on-campus dining facilities, ultimately impacting their dining frequency. Therefore, on-campus foodservice providers will need to recognize their consumers' behavior and work toward satisfying them to successfully attract users to the dining services.\nLiterature review: Globally, empirical studies have explored levels of satisfaction with on-campus foodservice operations including services provided by contracted companies, restricted foodservice operations and foodservices that include multiple locations and options on site (Joung et al., 2016; Tudin et al., 2010). In these studies, several attributes including food quality, service quality, atmosphere, cleanliness, location, food and beverage options, convenience, price and value were examined to determine levels of satisfaction. Hu et al. (2009) indicated that customer satisfaction evolves over a series of different experiences and defined it as an intellectual or emotional reaction that develops in response to a single or extended set of service encounters. How individuals perceive quality as it pertains to food and service reflect their opinions of the overall satisfaction with the food or service and the discrepancy between their expectations and their perceptions of the food or service (Joung et al., 2016). While perceptions of quality are subject to individual experiences, they are not necessarily evaluated objectively, because of the complex and intangible characteristics associated with the concept of perception (Kwun, 2011). Thus, it is necessary to review previous research that discussed these critical characteristics that impact customers' perceptions and ultimately their levels of satisfaction with foodservice operations within higher education institutions.\nHypothesis development: Generally, students' perceptions of their on-campus foodservice experience reflect their overall satisfaction with their dining experiences (Ali and Ryu's, 2015; Hall, 2014; Othman et al., 2013). Considering students' perceptions can influence their satisfaction, which in turn could impact the institution's reputation, student enrollment, student retention and the marketability of the institution to future students, it is critical to examine the students' perceptions of the on-campus food and beverage options, service quality, quality of food, value for money and ambience in relationship to overall satisfaction. Accordingly, in the context of the HBCU students' perceptions, we hypothesized the following:\nMethods: Sample and data collection\nResults: Measurement model\nConclusion and implications: The purpose of this study was to determine the HBCU students' perceptions of an on-campus foodservice facility, specifically in relation to the operation attributes of quality of food, ambience, value, food and beverage options and service and how it impacts their dining frequency. Results indicate that the six hypotheses tested were supported. All service and operations attributes investigated have a positively significant impact on the overall satisfaction of students using the foodservice facility and that the overall satisfaction significantly impacts their dining frequency at the foodservice facility. Consequently, there is compelling evidence indicating negative perceptions of the foodservice facility based on the statistically significant relationship found between overall satisfaction and quality of food, ambience, value for money, food and beverage options and service quality. Therefore, any positive adjustments to these attributes will positively influence the overall satisfaction of the end users, the students, which will ultimately shape their favorable behavior. The results further support previous studies (Ali and Ryu, 2015; Lam and Heung, 1998; Yen-Soon et al., 2004; Joung et al., 2016) which indicated a positively significant relationship between student satisfaction and behavior and the various elements of the foodservice facility and operations that were examined. Additionally, it corroborates Williams' (2013) five trends that foodservice providers should adopt to yield customer satisfaction. As well, the research validates previous research (Chen et al., 1994; Joung et al., 2016; Kwun, 2011; Lim, 2010) which purported that perceived value, influenced by price and quality, inspires positive behavioral intentions. This would explain the difference indicated between the commuter and residential student groups on the perceived value, as they would be comparing the value to that of competitors off campus. Thus, the management of the facility should consider reviewing its opening hours, food and beverage options, ambience and the culture of the institution and its student body, as areas for improvement in its customer satisfaction efforts and increasing students' dining frequency.\nLimitations and suggestion for future research: The major limitation to this study is generalization. Data for this study were collected at an HBCU in the Southern USA and may therefore not reflect the perceptions of students at other HBCUs in other parts of the USA. Subsequently, future research should include students at other HBCU campuses. As well, the study cannot be generalized to student bodies at predominantly white institutions (PWIs) due to the minimal representation of Caucasian respondents. Thus, future research could include student populations at PWIs. Even though this study looked into the impact of students' perception on their behavior related to on-campus food operations, future studies should also analyze if students' perception of foodservice plays in their decision to attend one institution vs another. Another interesting research idea will be to examine the role of negative perception of the on-campus foodservice operations on the institutional image.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to determine the students' perceptions of an on-campus foodservice operation at an identified historically black college and university (HBCU) and its effect on their satisfaction and dining frequency.\nDesign/methodology/approach: A survey of 685 students was conducted to collect data. Partial least squares based structural equation modeling is used to test the proposed structural model with SmartPLS 3.0.\nFindings: Results confirm that quality of food, ambience, value for money, food and beverage options and service quality have a positively significant impact on students' overall satisfaction with the on-campus foodservice operation and dining frequency. As such, all the hypotheses are supported.\nResearch limitations/implications: These findings indicate that on-campus foodservice operators should focus on quality of food, ambience, value for money, food and beverage options and service quality to achieve student satisfaction. This in turn could positively impact the institution's reputation, student retention and the marketability of the institution to future students.\nOriginality/value: This study would help on-campus foodservice operators to better understand the impact of the various elements of foodservice experience which will lead to students' overall satisfaction and dining frequency, particularly in a HBCU setting.\n",
"role": "assistant"
}
] |
Designing a performance measurement system for collaborative network
|
[
"Collaboration",
"Performance measurement system",
"Performance management",
"Collaborative network"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In the modern world, companies operate in globalised, turbulent and complex markets (Cocca and Alberti, 2010; Barrows and Neely, 2011; Nudurupati et al., 2011; Pekkola, 2013). It is commonly accepted that in order to survive in such a competitive environment, companies have to collaborate with one another, focussing on meeting customers' needs more effectively and efficiently (Bititci et al., 2007, 2012). If companies therefore aim to create and sustain a competitive advantage through collaboration, their network structures should be fully understood and managed; otherwise, the objectives will not be achieved and the network will fail (Verdecho et al., 2009). Therefore, a network-level performance measurement system (PMS) is needed by collaborating organisations. Unfortunately, only limited empirical research has been conducted on PMSs and their design in small- and medium-sized enterprise networks (Yin et al., 2011; Bititci et al., 2012). It is important to apply system-level control mechanisms and performance measurement (PM) tools to a network when attempting to manage it. A network itself has no intrinsic value; rather, it is a tool for organising operations amongst companies. Therefore, recently the networks' ability to succeed in their tasks has been receiving considerable attention (Bititci et al., 2007; Varamaki et al., 2008; Laihonen et al., 2014). Companies in a collaborative network are interested in the benefits and costs of networking, investors are keen on the revenue opportunities involved and customers in the value chain are concerned with the network's ability to manage production tasks as well as or better than a single, integrated company. This situation leads to the following question: how are networks considered in measurement systems?\n2. Literature review: 2.1 Classification of networks\n3. Methodology: 3.1 Case research\n4. Findings: 4.1 How can a PMS for a collaborative network be designed?\n5. Conclusions: Researchers have emphasised the need for a structured design process for a PMS for a collaborative network (e.g. Busi and Bititci, 2006; Yin et al., 2011; Bititci et al., 2012). To address this demand, a conceptual framework for developing such a process was presented and empirically investigated. This paper concludes by illustrating how this goal can be achieved and its influencing factors. The study also highlights the most essential practical experiences related to this process.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine how a performance measurement system (PMS) can be designed for a collaborative network and to identify which factors affect such a design.\nDesign/methodology/approach: This is a single-case study of a collaborative network. The data have been collected from semi-structured interviews conducted during 2008-2009 and after the design process in 2010 and 2012, respectively.\nFindings: The research results present a five-step process model for designing a PMS for a collaborative network. The findings show that a participatory development style that enhances socialisation, the positive development of a network culture and an outside facilitator all have beneficial effects on the design process.\nPractical implications: The practical contribution of this study is related to knowledge about the PMS design process for a collaborative network to support its measurement-related development projects. This knowledge involves the phases of such a process as well as the various factors supporting or hindering it.\nOriginality/value: The study presents a PMS design process for the case network, which can be utilised in other collaborative networks in a similar context. It also highlights the most essential practical experiences related to this process.\n",
"role": "assistant"
}
] |
The regulatory considerations and ethical dilemmas of location-based services (LBS) : A literature review
|
[
"Ethics",
"Location-based services",
"Literature review",
"Legal aspects of IT",
"Regulation/deregulation",
"LBS"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The essential aim of this paper is to define the ethical dilemmas of location-based services (LBS) that influence regulatory choices. LBS are commonly termed location services, location-aware services and/or location-related services due to their application in a variety of industries, and even as context-aware services (Kupper, 2005, pp. 1-2; Zhu et al., 2010, p. 410). As Bauer et al. (2005, p. 212) note, defining features of LBS include mobility, ubiquity and personalisation. Various definitions of LBS are available, given the multiple means associated with classifying LBS, and the myriad of application areas serving specific types of end-user groups (Australian Communications and Media Authority 2010; Kupper and Treu, 2010, p. 214; Spiekermann, 2004, p. 10; Jacobsen, 2004, p. 84; Lopez, 2004, p. 171; Shiode et al., 2004, p. 350; Astroth, 2003, p. 269; Giaglis et al., 2003, p. 68). The consolidated definition that has been adopted for this paper is: an application that combines the location or position of a mobile device associated with a given entity (individual or object) together with contextual information, to offer a value-added service and/or fulfil a particular need for the user across a wireless network. An important consideration with LBS with respect to the value-added component in particular is that LBS are not solely based on knowledge of an entity's location, but rather rely on the coupling of location information with further data of interest to the user according to Giaglis et al. (2003, p. 69).\n2. The qualitative approach to reviewing literature: The purpose of a literature review is to evaluate and describe gaps in existing research, thus situating and providing context for a given investigation. According to Aveyard (2010, pp. 9-19), a literature review should be conducted in a comprehensive and systematic manner, adopting a meticulous process for reviewing literature. Notably, the process involves defining the research topic, determining the methods to be employed for reviewing, analysing and examining relevant scholarship, and discussing research gaps.\n3. Background: the LBS value chain/network: Satisfying user needs through the provision of LBS requires a complex set of interactions amongst a series of stakeholders. This is traditionally represented as a value chain, which refers to \"a map of the entire set of competencies, investments and activities required to produce, deliver, maintain and reap the proceeds from a product or service\" (Rulke et al., 2003, p. 122). The value chain representation is common in the literature and in industry (Paavilainen, 2001; Barnes, 2002; Rulke et al., 2003; ELBA Consortium, 2004; DCITA, 2006), with the term chain implying sequential and linear interactions and processes. However, Alanen and Autio (2003, p. 176) argue that a value net or network is preferable, given that mobile business services are in fact non-sequential. Importantly, this net is a mesh of three distinct value chains, specifically the service development, service provisioning and devices value chains (Alanen and Autio, 2003, p. 176). Li and Whalley (2002, pp. 464-466) concur, in an article focusing on the deconstruction of the telecommunications value chain and sub-markets such as the mobile market, that value network better encapsulates industry complexities.\n4. Ethical dilemmas of LBS: Studies addressing the ethical implications of emerging technologies, such as LBS, generally reflect on the risks resulting from the implementation of a particular technology within a given social context. While numerous approaches to ethics exist, all are inextricably linked to ideas of morality, and an ability to distinguish good conduct from bad. Ethics can be considered as the \"study of morality\" (Quinn, 2006, p. 55; Tavani, 2007, p. 32), where morality refers to a \"system of rules for guiding human conduct and principles for evaluating those rules\" (Tavani, 2007, p. 32). This definition is shared by Elliot and Phillips (2004, p. 465), who regard ethics as \"a set of rules, or a decision procedure, or both, intended to provide the conditions under which the greatest number of human beings can succeed in 'flourishing', where 'flourishing' is defined as living a fully human life\" (O'Connor and Godar, 2003, p. 248).\n5. Discussion: regulating LBS: 5.1 Approaches to ICT regulation\n6. Findings: major considerations in LBS regulatory discussions: This paper has demonstrated that investigating the need for LBS regulation requires an inclusive approach, in which the ethical dilemmas pertaining to LBS must be identified and the multiple approaches to regulation be explored. The thematic review of literature has furthermore revealed that regulatory discussions should in fact focus on three broad considerations:\n7. Conclusion: This paper has offered a qualitative examination of LBS regulation scholarship through a thematic review of the LBS value chain, the ethical dilemmas of LBS and the associated regulatory implications. The LBS value chain presented throughout this paper highlighted the potentially varying interests and viewpoints that must be accounted for in the delivery of an LBS solution. Appreciation of divergent stakeholder interests entails a certain degree of sensitivity when examining the ethical dilemmas associated with LBS. As this paper has indicated, this also involves the recognition that a range of regulatory approaches may be necessary when regulating LBS, in order to incorporate differing stakeholder perspectives. Additionally, this paper has emphasised the value of stakeholder involvement in the LBS regulatory process and the importance of simultaneously assessing technical, social and environmental considerations in future regulatory discussions and related research.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to comprehensively review existing literature regarding the ethical dilemmas posed by location-based services (LBS) and their impact upon the adoption of a regulatory framework.\nDesign/methodology/approach: This paper employs a qualitative approach for reviewing LBS scholarship, in which existing knowledge is presented in narrative form and is critiqued thematically.\nFindings: In reviewing contemporary scholarship, the value of technical, social and environmental considerations is demonstrated. This encourages an understanding of the complexities, multiple interests and contextual factors that must be incorporated into the examination of LBS regulation in any social context.\nPractical implications: Approximately 85 per cent of handsets now have a global positioning system chipset on board. LBS affect a great number of mobile users. This research will create awareness among users of the potential benefits and harms that can come from the (mis)use of the technology. It will also promote an awareness of the complexities surrounding LBS regulation, drawing attention to the importance of collaboration and involvement of LBS stakeholders in the regulatory process.\nOriginality/value: Defines the ethical dilemmas of LBS that influence regulatory choices through a review of applicable literature and proposes that future research simultaneously address technical, social and environmental factors relevant to LBS.\n",
"role": "assistant"
}
] |
Developing inclusive and collaborative entrepreneuring spaces
|
[
"Women’s entrepreneurship",
"Feminism",
"Policy",
"Inclusion",
"Collaborative spaces"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Significant barriers for women starting or growing their businesses are still prevalent in the UK despite nearly 30 years of attention and research in this area (Ahl and Marlow, 2019). In 2018, the UK Government commissioned \"The Alison Rose Review of Female Entrepreneur'ship\" in recognition of this issue as a continued response to the prevailing discourse of more women starting and growing businesses being the solution to economic vibrancy. The review cites lower access to finance, attitude to risk, perceived lack of skills and experience, disproportionate childcare responsibilities and lack of sponsorship and role models (Rose, 2019) as evidence of little progress given three decades of focus. The UK Government responded to the Rose Review (HM Treasury, 2019) by setting the target of increasing the number of women entrepreneurs to 50 per cent by 2030, equating to nearly 600,000 more women entrepreneurs. They have committed to develop Gender Equality and Economic Empowerment Strategy, a new code for investing in women, supporting the expansion of networks which connect women entrepreneurs with one another, with the Chancellor committing an additional PS20m to strengthen local peer-to-peer business networks across England (HM Treasury, 2019). While all of the above is supportive in an effort for a more positive and inclusive outlook for women in entrepreneuring in the UK, the concern is how can greater progress be ensured in the next 30 years. This paper proposes the development of inclusive collaborative spaces whose focus is to influence and shape processes and structures beyond the individual and community level to the macro level of enterprise as one way to ensure more substantial progress is made.\nA feminist framing: learning from inclusion developments: The feminist pragmatic realist ontology of this study is important in understanding how extant knowledge and understandings of the diversity and minority networks within the broader women in leadership, women in management, diversity and inclusion fields are framed. Feminism is understood as exploring and challenging the gendered social order (Knights and Kerfoot, 2004) to create new knowledge towards cultivating social and political change (Crasnow, 2008; Brooks, 2007; Fonow and Cook, 2005; Hurley, 1999). This focus emphasises the macro level. The pragmatic realist ontological position within a feminist framing, draws upon Watson's (2011) understanding, acknowledging that there are no absolute truths, but relative truths with the realities of social structures understood to emerge from human practices. The value of pragmatic realism for this study is its concern with how people can develop spaces, which create new practices, discourses and cultures to reshape structures for inclusive entrepreneuring.\nCollaborative entrepreneuring communities: There is a lack of understanding of the co-ordination between researchers, policymakers and practitioners (entrepreneurs and business support providers) to develop a collaborative entrepreneurship landscape, which both advances issues of gender (Ahl and Marlow, 2019; Arshed et al., 2019; Benschop, 2009; Jones, 2014; Jones et al., 2018; Pecis and Priola, 2019; Vayreda et al., 2019; Henry et al., 2017) and other categories of social difference (Essers et al., 2010; Essers and Tedmanson, 2014) . The paper turns to explore how each of these communities: women's enterprise policy; academic research and entrepreneurial practice-based communities are challenging gendered structures.\nDiscussion: Through a feminist pragmatic realist lens, this paper has explored recent developments in the broader diversity and inclusion fields to appreciate the processes and practices that enable inclusive and collaborative entrepreneuring spaces. Exploration of issues and criticisms has been framed within three entrepreneurship communities: policy; research and practice with respect to gender, and specifically for women. In doing so, the need for inclusive and collaborative spaces whose focus is to influence and shape structures and processes beyond the individual or community, which Dennissen et al. (2019) frame as the organisation level, but this paper extends for the purpose of entrepreneurship to a macro level of entrepreneurship policy to enact significant change.\nConclusion and implications for future research: The need for engagement of stakeholders across the entrepreneurial ecosystem (Arshed et al., 2019; Rouse and Woolnough, 2018) is critical for the real change to be seen. To do so, we need to work with existing collectives within entrepreneuring (Arshed et al., 2019) to develop a movement whereby women are empowered by deeper political critiques that enable the proposal of policy solutions (Rouse and Woolnough, 2018) which ignite a larger social movement for greater equity within the entrepreneurship ecosystem. The need for greater collaboration through engagement and collective activism across research, policy and practice will ensure postfeminist sensibilities (Ahl and Marlow, 2019) are not perpetuated but rather become more progressive through enactment of change. This study offers a framework for developing inclusive and collaborative spaces which values feminist knowledge, opening up the possibility for greater intersectional appreciation without denying the importance of collective action which is not solely driven by economic measures in order to reveal to enable a reordering of gendered social relations within entrepreneuring (Ahl and Marlow, 2019). The framework supports the process of dismantling and rummaging through current inequalities in one particular context. By focusing at the macro level to dismantle and resist gendered regimes and change processes and practices (Jones et al., 2018) the focus can be on developing systems, structures, processes and practices, which are more inclusive. In creating such inclusive and collaborative entrepreneuring spaces, which are more equitable, fair and just, will result in economic and social benefits for all who chose to engage in the entrepreneurial ecosystem because it is right for them in a particular context.\n",
"role": "user"
},
{
"content": "Purpose: The call for more women to start up and grow businesses as a vehicle for economic vibrancy is a prevailing discourse in the UK. There have been calls for greater co-ordination between research, policy and practice to create collaborative spaces whose focus is to influence and shape structures and processes beyond the individual or community level to a macro level of enterprise policy. However, calls have not specifically focussed on the issues of gender or other categories of social difference. This study aims to understand how such co-ordinations can be established to enable progress within the women's entrepreneurship space through the development of collaborative spaces fusing research, policy and practice and how they should be structured to ensure inclusion through the process as well as enabling greater inclusion as part of the collaborative space outcome.\nDesign/methodology/approach: Taking a critical feminist perspective, the study draws from extant literature on women and minority networks research from the women in leadership, diversity and inclusion fields as a lens through which to frame the analysis of women's enterprise policy in the UK, research and practice.\nFindings: The study highlights the importance of collective feminist action drawing upon post-feminist sensibilities and an Engaged-Activist Scholarship approach. Such collective feminist action appreciates the importance of the micro as an enabler to progressive action at the macro level to enact structural and system change within the entrepreneurial ecosystem. A framework for inclusive and collaborative entrepreneuring space development is offered.\nPractical implications: This paper offers policymakers, researchers and practitioners a framework as a practical way forward to ensure efforts are progressive and enable structural and systemic change.\nOriginality/value: The paper offers a framework for developing inclusive and collaborative entrepreneuring spaces to ensure progression by lifting the focus to a macro level of change to enable inclusion as part of the process and outcome of such collaborative spaces.\n",
"role": "assistant"
}
] |
Workload factors impacting managers
|
[
"Managers",
"Lean production",
"Labour efficiency"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: A typical manager's work day ... to beat the traffic, our typical manager starts for work early. Unfortunately, an accident on the way means our manager arrives with the crowd at the normal time. The plan for a proactive \"jump\" on the work to be done today goes out the window. Even though our manager checked e-mail until 11 pm the night before, there is still new e-mail to be reviewed. After deleting the garbage and insignificant messages, this recent batch of e-mail creates more assignments for the new day.Possibly because of the traffic, this day provides relief to the norm because our manager had time to actually dock her laptop and check e-mail before the first unscheduled guest arrives at the office door. Invariably, there are always some direct reports that consistently arrive early to work. Well past their e-mail check-up, these employees have begun their work day and are already to the point where managerial direction is required before they can proceed to complete tasks.It would be a luxury for this manager if all daily interruptions came from expected sources, like direct reports and supervisors. But, just like those numerous e-mails, office visitors and phone callers seem to come from all directions, both within and outside the organization. A printout from the manager's planner shows five hours of scheduled meetings today and a business lunch. What is certain to be at least a ten-hour day is already more than half-consumed in dead time. Remaining time must be efficiently deployed just to stay on schedule for impending deliverables. Any thinking about the context of the work queue must be dealt with outside of that ten-hour day. Even with all the new gadgets the company has provided our manager, a laptop, a personal digital assistant, and a Blackberry, our manager feels no more efficient or in control.We paint this picture of a typical manager from hundreds of informal conversations with them. Those conversations were part of a more formal study of how managers allocate their time today and their perceptions about the level of workload and the level of control. Neither subject is new but studying the relationship between them given contemporary workplace shifts in technology and globalization merits attention. This research goes beyond a basic sampling of managers' work. The objectives are to tie in that quantifiable transaction data with perceptions about a manager's control of their assigned tasks. The authors hypothesize that relationships exist between perception of control and how time is spent. This research attempts to model those relationships for further testing.Background\nResearch questions: Our guiding hypothesis was that the workload of lower and middle-level managers in service industries could be alleviated by applying some of the techniques of lean management from industrial production. We administered a survey to determine what factors relevant to lean production could be reliably measured among lower and middle-level managers. We focused on the following questions:* What is a reliable measurement of a lower and middle-manager's workload?* Is it possible to measure how lower and middle-managers allocate their time at work?* What is a reliable measurement of \"work in progress\" for lower and middle-managers?* What is a reliable measurement of \"lead time\" for lower and middle-managers?* What indicators of \"lean production\" can be reliably measured among lower and middle managers?* Does reduced queue time increase individual productivity among lower and middle managers?* Does reduced queue time increase group productivity among lower and middle managers?This is an exploratory study and our primary goal is to identify which, if any, measures of lean production can be transferred usefully to lower and middle-management in service industries.We consider our subjects to be knowledge workers. Drucker (1959) first used the term \"knowledge worker\" to describe employees that work with intangible resources. More recently, Drucker (1988) has stated that \"knowledge tends to be concentrated in service staffs, perched rather insecurely between top management and the operating people\". Dove (1998) and Coates (1986) take an even broader view of what constitutes knowledge work but many prominent researchers agree that for knowledge work, \"workers' brains comprise the means of production\" (Ramirez and Nembhard, 2004). Based on these definitions, the authors strongly believe this sample to be representative of knowledge workers, but our assessment of workload is not concerned with how these subjects manage knowledge for future dissemination.\nMethod: Subjects were 125 lower and middle managers enrolled in part-time MBA classes at a Northeastern university. Participation was voluntary and responses kept anonymous, with no monetary or grade compensation. The surveys were handed out in class, with instructions to take them home and return them at the next class. Over 90 percent of surveys were returned and useable. Subjects were employed in lower and middle management at a wide range of organizations. The survey consisted of 42 items covering the subject's workload, daily tasks, projects, and attitudes. Sixteen items measured subjects' attitudes toward their workload and work control on a five-point Likert scale. The remaining 26 items covered the subjects' workload, daily tasks, and pending projects. Items queried the types of communications subjects received, the queue sizes in numbers of items and in time to complete, the time spent in various office activities, and the number and type of interruptions experienced throughout the day.Descriptive statistics\nResults: To explore the relationship between measures of managers' workloads and measures of their attitudes, principal components analysis (SPSS, 2002) was performed using Varimax rotation (Kaiser, 1958; Norusis, 2004) to identify primary factors. Cattell's (1978) scree plot was used to select the number of primary factors to rotate. Five primary factors were identified which accounted for over 56 percent of total variance (Figure 1). The rotated factor matrix is shown in Table I.Description of the five factors\nConclusion: In this exploratory analysis of the perceptions and attitudes of middle-managers regarding their workloads, we have identified five primary factors that together account for approximately 56 percent of total variance. Based on face validity, the first four of these factors, labeled \"WIP\", \"empowerment\", \"non-value time\", \"facilitation\", appear to correspond to commonly accepted lean production constructs. Validation of these factors and elucidation of the relationships between them awaits further research. Four of the five primary factors demonstrate content validity, since the questions relate to the construct (Kline, 1994; Spector, 1992). They also demonstrate face validity since the relationship between the items and the construct is not hidden (Kline, 1994; Spector, 1992). However, much more research is required to demonstrate construct validity (Cronbach and Meehl, 1955; Kline, 1994), to prove empirically that the factors actually measure the characteristics we hypothesize. In testing for that construct validity, we hope to also create a valid measure of leanness for knowledge-based work.\n",
"role": "user"
},
{
"content": "Purpose: - The paper seeks to identify elements of lean production that may be applicable to managers.\nDesign/methodology/approach: - A survey was administered to 125 middle and lower level managers to assess perceptions about level of workload and amount of workload control. Principal components analysis was employed to identify primary factors.\nFindings: - Five primary factors accounted for 56 percent of total variance. The principal factors corresponded to the lean production concepts of work in process level, empowerment, and non value-added time.\nOriginality/value: - These middle and lower level managers identified five primary factors that accounted for 56 percent of the variance in their self-reported levels of workload and control. Three of these factors correspond to existing metrics that have been utilized in lean production in industry.\n",
"role": "assistant"
}
] |
Ranking the strategies for Indian medical tourism sector through the integration of SWOT analysis and TOPSIS method
|
[
"TOPSIS",
"SWOT analysis",
"Indian medical tourism",
"MADM technique",
"Strategic decisions"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Organizations have to evaluate their internal and external environments in this highly competitive world. Strengths, weaknesses, opportunities and threats (SWOT) analysis is a very good and useful technique which analyzes the strengths, weaknesses, opportunities and threats of an organization for taking strategic decisions and it also provides a foundation for formulation of strategies. SWOT analysis may be integrated with multicriteria decision-making techniques like the technique for order preference by similarity to ideal solution (TOPSIS), analytical hierarchy process (AHP), etc., to quantify the importance of the factors affecting the organization.\nSWOT analysis: SWOT analysis is the most widely used approach of an organization in the strategic planning and decision-making process. An organization can utilize the opportunities and face the threats by diagnosing its external environment. On the other hand, an organization can frame new strategies taking into account its strengths and weaknesses by diagnosing its internal environment. SWOT analysis identifies the most important internal and external strategic factors affecting an organization's future. It is a tool which is used for interpretation and decision making under a variety of circumstances in different organizations and firms. This tool offers a structured framework to analyze the strategies and decisions of an organization. SWOT analysis can be used for product development, business planning and marketing. Higginbottom and Hurst (2001) developed a therapy quality assurance strategy by using the SWOT analysis in the national health sector in the UK. Helms and Judy (2010) conducted a comprehensive literature review of the use of SWOT analysis in a decade and framed a classification of its levels and applications. With the help of the SWOT matrix, four different types of strategies are developed which are strengths-opportunities, weaknesses-opportunities, strengths-threats and weaknesses-threats. SO strategies are developed by taking advantage of external opportunities and using a firm's internal strengths. In WO strategies internal weaknesses are improved by taking advantage of available external opportunities. ST strategies are developed by using a firm's strengths to reduce the influence of external threats. WT strategies are defensive strategies which are aimed at decreasing the internal weaknesses and avoiding the environmental threats.\nBackground of study: Hill and Westbrook (1997) and Kurttila et al. (2000) pointed out that SWOT analysis has a lot of analytical deficiencies, though it is commonly used in various strategic planning processes like how to calculate SWOT group and factor weights in a unidirectional hierarchical structure. The second issue is how to analyze and detect the best strategy on the basis of SWOT factors. So some studies have been conducted integrating SWOT analysis with multicriteria decision-making techniques to remove these systematic and analytical deficiencies in order to improve the effectiveness of SWOT analysis (Kurttila et al., 2000; Kangas et al, 2003; Shrestha et al, 2004; Stewart et al, 2002; Ygksel and Dagdeviren, 2007; Masozera et al, 2006; Zaerpour et al, 2008). Ghazinoory et al. (2007) conducted the SWOT analysis in an Iranian food corporation by using fuzzy theory concept and in their research, the best strategy area was detected based on the internal and external factors of the corporation. Strengths and weaknesses were determined as internal factors and threats and opportunities as the external factors. But the study did not quantify the importance of strategies and only a strategy area was determined. SWOT analysis has been integrated with some multiattribute decision-making techniques like AHP. Azimi et al. (2011) used SWOT analysis to derive various strategies for Iranian mining sector. The weights of SWOT factors were obtained by using a multicriteria decision-making technique called analytical network process and finally TOPSIS technique was used to rank the strategies. Ajmera et al. (2015) integrated SWOT analysis with AHP to prioritize different strengths, weaknesses, opportunities and threats of Indian medical tourism sector. In another study conducted by Nesrin (2013), SWOT was integrated with TOPSIS in strategic decision making about a furniture firm. In total, 12 evaluation criteria and 4 strategies were formulated in this research. The study identified ST strategy to be the best strategy which matches significantly with the four elements S, W, O and T of SWOT matrix and 12 strategic indicators.\nMaterials and methods: The SWOT analysis of Indian medical tourism sector has been carried out by interviewing five managers of the international wings of three hospitals of Delhi NCR in India and a SWOT matrix along with four different strategies has been developed. SWOT analysis is then integrated with TOPSIS to rank different strategies of Indian medical tourism in the order of priority.\nMedical tourism in India: Medical tourism industry is one of the rapidly growing healthcare industries in the world economy and is faced with new issues and challenges. Currently, the medical tourism industry is considered to be one of the best ways to boost the healthcare industry especially in the developing countries of the world (Siddiqi et al., 2009). There are several factors that make India an attractive destination for people seeking healthcare services. These include its qualified and well-trained health professionals, a large population of the English-speaking hospital and medical staff, clinical excellence, availability of allopathic and alternative systems of medicine, the presence of super-specialty centers, good preoperative counseling and postoperative care and use of technologically advanced diagnostic and clinical equipment. But the medical tourism sector in India is facing competition from the countries like Thailand, Malaysia, Singapore, etc., which have established themselves as credible and favorable medical tourism destinations (Chanda, 2002). Because of this growing competition, it is necessary to determine suitable strategies for the Indian medical tourism sector and to further prioritize the implementation of these strategies.\nAim of the study: To prioritize the strategies for the Indian medical tourism sector so that the policy makers can use this ranking in their strategic decision-making process and can allocate their resources in the best possible manner.\nResults: Results of the study show that SO strategy (providing best facilitation and care to the medical tourists at par to developed countries) is the best strategy which matches with the four elements of S, W, O and T of SWOT matrix and 35 strategic indicators. The managers of the hospitals where the study was conducted confirmed that the internal and external factors are covered appropriately by the proposed framework, and the strategies based on these factors, accurately represented the environment influencing the Indian medical tourism sector. It was stated that the strategies which have been developed in the present study are acceptable and appropriate for the vision, mission and objectives of the medical tourism sector in India.\nConclusion: Proper coordination between factors in the internal and external environment is required for good organizational performance. Selection of strategy is the most complicated task for the organizations.\n",
"role": "user"
},
{
"content": "Purpose: Organizations have to evaluate their internal and external environments in this highly competitive world. Strengths, weaknesses, opportunities and threats (SWOT) analysis is a very useful technique which analyzes the strengths, weaknesses, opportunities and threats of an organization for taking strategic decisions and it also provides a foundation for the formulation of strategies. But the drawback of SWOT analysis is that it does not quantify the importance of individual factors affecting the organization and the individual factors are described in brief without weighing them. Because of this reason, SWOT analysis can be integrated with any multiple attribute decision-making (MADM) technique like the technique for order preference by similarity to ideal solution (TOPSIS), analytical hierarchy process, etc., to evaluate the best alternative among the available strategic alternatives. The paper aims to discuss these issues.\nDesign/methodology/approach: In this study, SWOT analysis is integrated with a multicriteria decision-making technique called TOPSIS to rank different strategies for Indian medical tourism in order of priority.\nFindings: SO strategy (providing best facilitation and care to the medical tourists at par to developed countries) is the best strategy which matches with the four elements of S, W, O and T of SWOT matrix and 35 strategic indicators.\nPractical implications: This paper proposes a solution based on a combined SWOT analysis and TOPSIS approach to help the organizations to evaluate and select strategies.\nOriginality/value: Creating a new technology or administering a new strategy always has some degree of resistance by employees. To minimize resistance, the author has used TOPSIS as it involves group thinking, requiring every manager of the organization to analyze and evaluate different alternatives and average measure of each parameter in final decision matrix.\n",
"role": "assistant"
}
] |
The meaning of job performance in collectivistic and high power distance cultures: Evidence from three Latin American countries
|
[
"Employee behaviour",
"Performance management",
"Venezuela",
"Colombia",
"Ecuador"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: As a central criterion for I/O psychologists, performance has attracted significant research attention over the last years (Austin and Villanova, 1992). This body of research has placed special emphasis on construct specification, leading to isolate employees' behaviors - as a core component - from other factors that appraisals often address (i.e. traits, outputs; Wexley and Klimoski, 1984). Accordingly, Campbell's (1990) conceptualization of performance as behaviors (a) under the employee control and (b) causally related to organizational goals has gained broad acceptance in the field.\nPerformance categories: Organizational citizenship behavior (OCB)\nImpact of cross-cultural values on models of work behaviors: Collectivism\nMethod: Research instrument\nAnalysis: Measurement invariance\nDiscussion: We have maintained here that our knowledge of the cultural variations of job performance is deficient. Because the majority of investigations have sampled employees from individualistic and relatively LPD settings, the extent to which work-behavior models transcend cultures is unclear. This study addresses this knowledge gap by observing the pattern of co-variances among three major behavioral categories in a sample of 1,022 employees from collectivistic and HPD countries. Results indicate that values account for culture-specific variations, thus challenging the multicultural robustness of existing theories. Theoretical and practical implications stemming from results are discussed next.Theoretical implications\n",
"role": "user"
},
{
"content": "Purpose: - Three broad behavioral categories have been related to organizational goals: task (in-role), citizenship (extra-role), and counterproductive behaviors. Because most studies modeling these behaviors have been conducted in culturally similar contexts (individualistic and relatively low power distance settings), the purpose of this paper is to test the invariance of such a triad categorization of performance in collectivistic and high power distance contexts.\nDesign/methodology/approach: - Data of employees' proficiencies (n = 1,022) in 34 work activities representing the three behavioral performance categories were factor analyzed. Data were collected by adapting existing behavioral-based instruments exhibiting strong psychometric properties.\nFindings: - Although results corroborate the existence of a triad categorization of employee behaviors, culture-specific variations attesting to the partitioning of in-role behaviors according to the distribution of power in organizations were found. Results also suggest that collectivistic individuals narrowly conceptualize extra-role behaviors by excluding discretionary interpersonal actions.\nPractical implications: - The paper's findings contribute to our understanding of how job performance varies in a global economy. These variations must be considered in appraisal instruments, especially in organizations operating across cultural contexts.\nOriginality/value: - This paper is believed to be the first to test the cultural invariance of a triad categorization of relevant employee behaviors.\n",
"role": "assistant"
}
] |
Passion, performance and concordance in rock "n" roll entrepreneurship
|
[
"Passion",
"Entrepreneur",
"Performance",
"Concordance",
"Creative economy",
"Music industry"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Passion is a source of motivation and inspiration that drives individual development and entrepreneurial performance. The theory of entrepreneurial passion (EP) has made a major contribution to explaining the constitution of passion and its effects (Cardon et al., 2013, 2017a). However, entrepreneurial action can arise from pursuing interesting ideas or activities driven by a passion that is initially non-entrepreneurial. Researchers have documented this phenomenon in diverse contexts, such as those of academic, cultural, social and user entrepreneurship (Huyghe et al., 2016; Ratten et al., 2019; Ruskin et al., 2016; Shah and Tripsas, 2007). Creative economy studies in particular have shown that a pre-existing passion for a creative domain often generates conflicts with the demands of business venturing (Schulte-Holthaus, 2018). Thus, the development of passion and its impact on performance may depend on whether passion is in conflict, peacefully coexists with or even extends to entrepreneurship.This paper addresses the research question of how passion, in entrepreneurial contexts, is manifested and relates to performance among individuals whose entrepreneurial journeys are prompted by a non-entrepreneurial passion. Proceeding from the EP theory, Vallerand et al.'s (2003) dualistic model of passion (DMP) is reviewed to develop a theoretical framework that explains developmental paths of passion in the creative economy. For empirical investigation, the methodology of comparative causal mapping (CCM) by Laukkanen and Wang (2015) is applied to capture individuals' passion by means of semi-structured interviews and to create cause maps that visualize its contents and relationships. Rock \"n\" roll entrepreneurs (as defined above) are selected for data collection. The term rock \"n\" roll is here used to encapsulate all diversified styles of handmade, electronically amplified, popular music that have developed since the second half of the twentieth century. Using a sample of 11 successful rock \"n\" roll entrepreneurs provides evidence that the concordance of constituting concept relationships determines the scope and degree of passion and its relationship to performance. Furthermore, the findings suggest that personality traits and life contexts play an important role in the development of passion, and that an optimal unfolding of passion and performance follows a clear pattern.By complementing EP and DMP theory through the introduction of the significance of context and concordance, theoretical and empirical conclusions contribute to the passion literature in entrepreneurship research. In addition, insights into the tensions and resolution mechanisms among the artistic, social and entrepreneurial demands of passion enhance the creative economy literature. Finally, the detailed illustration of passion in causal maps makes an important contribution to CCM research, which allows researchers to investigate sociopsychological phenomena at the boundaries of emotion, cognition and behavior in organizational and management research.\nTheoretical framework: Entrepreneurship in the creative economy\nMethod: Analytical approach\nResults: Constituents and mechanisms of passion\nDiscussion: Contribution and implications\n",
"role": "user"
},
{
"content": "Purpose: Non-entrepreneurial passions may be the beginning of an extensive entrepreneurial journey. However, current passion theories cannot fully capture the essence of such passions and their effects. The purpose of this study is to explore and explain the real-life composition of passion and performance.\nDesign/methodology/approach: The investigation was conducted with comparative causal mapping (CCM) on a qualitative sample of people we designate rock \"n\" roll entrepreneurs (i.e. individuals driven by a passion for music and who are successful both artistically and economically). Aggregated causal maps of passion elicited through semi-structured interviews were analyzed and contrasted with performance indicators.\nFindings: Passion is revealed to be an individual phenomenon, one composed of central and peripheral concepts that include--contrary to prior theories--personality traits and life contexts. Furthermore, the results suggest that the concordance of concepts determines the scope, degree and performance of passion.\nResearch limitations/implications: This study complements prevailing passion theories in psychology and entrepreneurship. As a context-bound study, the generalizability of the results is limited to its context, which, however, paves a clear way for future research.\nPractical implications: Creative economy entrepreneurs and educators can use the mechanism of concordance to consciously reflect passion-driven tensions between artistic, social and entrepreneurial demands and to translate passion into behavioral effectiveness.\nOriginality/value: This study is the first to use a CCM approach to investigate passion. Findings highlight the potential to research entrepreneurial phenomena at the intersection of emotion, cognition and action.\n",
"role": "assistant"
}
] |
Assessing the value of ILL to our users: a comparative study of three US libraries
|
[
"Libraries",
"Survey",
"Library users",
"Interlibrary loan",
"Document supply",
"Resource sharing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: \"Thank you\" are words heard by many interlibrary loan (ILL) staff. These words help staff feel good about services, and it indicates a library is providing a service desired by patrons.\nLiterature review: Inspired to focus on the value of ILL within the larger academic assessment movement, a framework from which to guide our study needed to be built. Looking at the literature, scholars contributing perspectives include Oakleaf (2012), Mak (2011) and Brown (2010), to name but a few. Resource sharing is identified as being within the library collection discussion of improving access and discoverability while shifting collections from print to digital, as well as local institution ownership to consortium and collaborative collections.\nPurpose: This study combines the patron perception of the ILL services offered in three libraries: an ARL campus, a small private liberal arts campus and a large state public library. The goal was to have a vehicle for gathering both satisfaction and value information for each of the services provided by a resource-sharing operation to cross-analyze patron perceptions, satisfaction and value placed on resource sharing at each institution.\nProcess: A survey was created to gather the information desired, and it was designed to be completed in less than 15 minutes (see Appendix 1 for the survey form). The survey wording was adjusted depending on the individual participating organization to provide appropriate text for the local patrons. However, the basic content was similar for each library. The survey included an optional field where the patron could enter their name to allow follow-up as desired. It also included an open question to capture anything else the previous questions did not.\nResults: Private liberal arts college results\nComparisons: Comparing all three organizations, the value of ILL services, according to patron perceptions, ranks overwhelmingly very important with 79 per cent, 93 per cent and 82 per cent for the public library, small liberal arts college and a research university, respectively. The four most important features in all three result sets were: speed, access, people and quality. The difference in results partly reflects the differing expectations of the patrons as well as, of course, the different level of service provided that is described by these features. So while satisfaction rates were high across all three surveys, academic institutions are focused on information-literate scholars and holistic learning experiences; their assessment and value-added services align with goals of the institution, as well as their libraries. Their patrons are attending their institutions for particular purposes and therefore the survey results will be slightly different than the public library survey.\nNext steps: The information gathered during this study was utilized for more than an affirmation of a job well done. Each participant used their patrons' data to tweak benchmarks and workflows within their department, as well as market services less known to their communities. Patron confusion and dissatisfactions were also identified and targeted for improvement. Many respondents in the KU survey commented that they did not know a specific service existed. In response, the University added a step so that when a new patron uses one of their services for the first time they are sent an e-mail with information on additional services available. At SJF, where confusion regarding Document Delivery services was evident, the decision to continue incorporating this service as a seamless aspect of ILL, rather than marketing it as a separate service, allowed further improvements in workflow to take place.\nConclusion: As Oakleaf describes, in multiple ways and with multiple examples, a main focus of assessment efforts and contributions should be on \"where we are helping in areas of institutional relevance and strategic directions,\" as well as \"mobilizing library administrators.\"\n",
"role": "user"
},
{
"content": "Purpose: This study aims to combine the patron perception of the interlibrary loan services offered in three libraries: an ARL campus, a small private liberal arts campus, and a large state public library. The goal was to have a vehicle for gathering both satisfaction and value information for each of the services provided by a resource-sharing operation to cross-analyze patron perceptions, satisfaction and value placed on resource sharing at each institution.\nDesign/methodology/approach: A survey was created to gather the information desired, and it was designed to be completed in less than 15 minutes. The survey wording was adjusted depending on the individual participating organization to provide appropriate text for the local patrons. However, the basic content was similar for each library. The survey included an optional field where the patron could enter their name to allow follow-up as desired. It also included an open question to capture anything else the previous questions did not.\nFindings: This case studies finding include recognition of recurring features found valuable to library patrons utilizing all three institutions: access, speed and people. Each library was able to identify areas of strength and areas to improve upon within their services. Each library also gained insight into their patrons' perceptions of the resource-sharing services offered. This insight is being used for further assessment studies, as well as to implement changes on each local campus.\nOriginality/value: This study provides insight and data into resource-sharing assessment as related to patron value. As the literature review shows, no studies of this kind have been done recently or across the institutions included: small private academic, large research academic or large public library.\n",
"role": "assistant"
}
] |
Resources for a positive perception of work among poor working youth
|
[
"Work values",
"Resilience",
"The Philippines",
"Initiative",
"Working youth"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Resources for a positive perception of work, among poor working youth, the high poverty incidence in the Philippines, has led to an increasing number of underprivileged Filipinos who stop schooling to enter the workforce to augment their families' finances (World Health Organization, 2010). Aged 16 through 30, and involved mostly in agriculture, mining and manufacturing (Philippine Statistics Authority, 2015), poor Filipino working youth typically enter the workforce with limited skills, education and training and, thus, are forced into temporary low-level and low-paying jobs. An estimated 2.3 million of them experience deplorable working conditions, receive insufficient compensation and lack social protection [see the International Labor Organization report of Canlas and Pardalis (2009) for a review].\nTheoretical background and hypotheses: Used as the theoretical basis of the current study, the COR theory examines how individuals protect, recover or gain the resources necessary for one's functioning and well-being (Hobfoll, 1989). Not confined to the material, the resources considered by the COR theory include individuals' social relations and their personal traits and skills (Gorgievski and Hobfoll, 2008). In the current study, the COR theory is used to conceptualize the social-psychological resource that is the workplace and the youth's own internal resource.\nWork values and resilience-building skills as workplace resources: One workplace resource that possibly contributes to youth's positive work perception is their co-workers' endorsement of work values, such as responsibility, professionalism and punctuality (Billett, 2004; Baciu, 2013). Such endorsement provides youth with the standards and guide for performing their work (Ros et al., 1999), enabling them to efficiently deal with difficulties at work (Jocano, 2000; Lantz and Andersson, 2009). In the face of the pressing need to augment the family income, the affirmation that poor working youth receive from co-workers of the material and personal benefits of work (something central to the Filipino work value system; Jocano, 2000) would prod them to work with the proper perception (Yabut et al., 2009).\nPersonal initiative as an inner resource: Examining the contributions of workplace resources to youth development goes hand-in-hand with examining the contributions of youth's internal resources, that is, their qualities and strengths (Benson, 2007; Larson, 2000). The youth are regarded as co-producers, in a manner of speaking, of their developmental outcomes and they do so by determining their actions in environments (Benson, 2007). Thus, the active role that youth should play in their development in the workplace cannot be discounted. For example, just how highly working youth will uphold the work values endorsed in the workplace and just how much they will participate in resilience-building efforts of their co-workers are both markers of their initiative that will determine how extensively they will work with a positive perception (cf. Myers and Sadaghiani, 2010; Smola and Sutton, 2002).\nMethod: Participants\nResults: Shown in Table I are the means, standard deviations and pairwise correlation coefficients between each pair of variables. Gender was not included in the statistical analyses as preliminary analysis showed no significant gender difference in work perception.\nDiscussion: The current research is founded on the premise that acquiring a positive perception of work would help counteract the stress and psychological loss because of non-optimal working conditions (e.g. low-level or low-paying job, lack of job security). Given this premise, the current study aims to establish the external and internal resources through which poor working youth can develop a positive perception of work.\n",
"role": "user"
},
{
"content": "Purpose: A positive perception of work helps counteract the stress and psychological loss because of non-optimal working conditions. This paper aims to hypothesize two pathways through which social-psychological resources in the workplace contribute to positive work perception: one pathway is direct and the other is through the mediating mechanism of youth's internal resource.\nDesign/methodology/approach: Filipino working youth from a government program for out-of-school poor youth, or working students enrolled in a free night high school, completed pertinent scales of the Multicontext Assessment Battery of Youth Development.\nFindings: Co-workers' endorsement of work values and their joint exercise of resilience-building skills build youth's positive work perception. Youth's personal initiative contributes to their positive work perception by its direct influence and also by its mediating mechanism.\nResearch limitations/implications: The cross-sectional survey of this study does not allow for a definitive temporal progression from resources to positive work perception as does a longitudinal study.\nSocial implications: For poor working youth, the threats of a financially bleak future can be tempered by an attitude that recognizes work not only as financially necessary but also as beneficial to one's growth. The social-psychological resources in the workplace and the youth's emerging personal initiative jointly contribute to a positive perception of work.\nOriginality/value: The current research shifts the focus of analysis from disadvantageous employment conditions to the affective and motivational aspects of employment and uses the conservation-of-resources theory to plot the flow of resources from the workplace to the worker.\n",
"role": "assistant"
}
] |
Premature deindustrialization and inequality
|
[
"Development",
"Income distribution",
"Developing countries"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The process of economic development has always been closely linked with industrialization and, more specifically, manufacturing. In this process structural change has always been thought to play a crucial role. One can conceive of a developing country as being dualistic in nature with a low labor productivity sector, often called the traditional sector, and a high labor productivity sector, usually thought to be mainly made up of manufacturing. Thus development involves a shift of labor out of the former and into the latter with the share of manufacturing in GDP and employment growing rapidly. One can combine this with the idea that initially this growth in manufacturing would be labor intensive in nature, given the abundance (and therefore cheapness) of labor in many developing nations, and that an evolving comparative advantage in labor intensive manufacturing would result in a rapid growth of exports of such manufactured goods. The experience of East and Southeast Asia is often used to illustrate this particular process of economic development (Lin, 2003; Francks, 1992; James et al., 1989).\n1. Premature deindustrialization: Relative deindustrialization (whether in terms of employment or output) is part of the typical process of structural change as economic development occurs. Initially most developing nations are dominated by agriculture so that the shares of agriculture in employment and production are both high, with the former exceeding the latter due to low labor productivity in agriculture. As economic development takes place, manufacturing tends to expand as a share of both GDP and employment while that for agriculture declines. Ultimately as development progresses manufacturing begins to decline and services expand in terms of shares of output and employment. Thus it would seem that relative deindustrialization is part of a successful development process.\n2. Demand and inequality: The work of Foellimi and Zweimuller (2011) provides a dualistic framework within which one can begin to think about the impact of inequality on premature deindustrialization. Actually, the model they develop is made up of three sectors: subsistence, mass manufacturing, and exclusive goods. The population is made up of two groups, poor and unskilled and rich and skilled. The poor work in the subsistence sector (producing food and/or traditional services), mass manufacturing, and the exclusive goods sector. The rich work only in exclusive goods production and mass manufacturing and they act as the entrepreneurs in terms of establishing these two types of production. Markets for these two goods are characterized by monopolistic competition. The production technologies for the two goods are identical, requiring only labor to produce goods.\n3. Some empirical evidence: General support for the model is provided in the work of Rydzck (2013). A reduced form estimation of the log of the share of workers in manufacturing is estimated for unbalanced panel of 65 countries for the years 1990-2010. Country and time fixed effects are incorporated for several estimations. The right hand side variable included GDP per capita, population, schooling, openness, and the Freedom House Index. Several measures of inequality are also utilized. All the estimates show that \"more equality has a strong positive impact on the labor share in manufacturing as suggested by the model\" (p. 16). The implication is that when inequality is high a lot of expenditures are directed at commodities which do not have significant employment potential and vice versa. Thus inequality would seem to be related to the extent of employment in manufacturing.\n4. Developed countries: As discussed earlier, the work of Piketty (2014) has shown that inequality in much of the developed world has worsened. The implication then is that as a source of demand for manufactured goods, especially labor intensive manufactured goods imported from developing countries, would be limited. In this short section some evidence supporting this fact will be presented.\n5. Summary and conclusion: Some years ago Ha-Joon Chang (2003) wrote an interesting book entitled Kicking Away the Ladder: Development Strategy in Historical Perspective. He argued that all of today's developed countries followed similar policies in the early stages of economic development. These policies involved using a number of different policy tools to subsidize the development of manufacturing so as to strengthen and speed up the structural change associated with economic development. However, the international institutional structure relating to trade constructed in the late twentieth century sought to reduce and limit the extent to which countries could use domestic policy to protect and subsidize the development of domestic manufacturing. Thus these institutions in effect prevent today's developing countries from utilizing industrial policy to move up the technological ladder, in a sense this ladder has been kicked away.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of the paper is to determine why premature deindustrialization is occurring in many developing countries.\nDesign/methodology/approach: A theoretical structure for explaining premature deindustrialization is utilized. Then the comparative experiences of a number of developing countries are used to illustrate the operation of the theory.\nFindings: The results indicate that increasing inequality among a number of developing countries has reduced the domestic market for labor intensive manufactured goods, resulting in stagnation in manufacturing. Also, the increasing inequality in developed countries has reduced international demand for labor intensive manufacturing. Thus developing countries have fewer opportunities to export labor intensive manufacturing.\nResearch limitations/implications: Data on inequality is limited and it is very difficult to determine causality. However, intuition indicates that causality is most likely bi-directional.\nPractical implications: Strategies of economic development must concern themselves with the effects that increasing inequality will likely have on the development of labor intensive manufacturing.\nSocial implications: Social programs that bolster the purchasing power of poor families are likely to be important (social safety net). Broad-based agricultural growth will provide a basis for labor intensive manufacturing.\nOriginality/value: The originality stems from the linking of deindustrialization with rising inequality.\n",
"role": "assistant"
}
] |
Modeling the standard compliant software processes in the university environment
|
[
"Process management",
"Modelling",
"Educational development",
"Educational policy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Business process modeling is becoming increasingly important activity especially in early phases of software development projects. Latest information technology evolution is heading towards process models as the core drivers of the distributed computing. In addition, the web services technologies stack contains process modeling and synchronization (choreography) languages on its highest level (Juric and Pusnik, 2004).Until now, the business process modeling was often neglected, mainly due to the gap between business requirements for information systems (process models) and ability of automated execution of the process models. The development of the latest XML-based languages for process description and process engines is enabling the execution of the business process models already.Nevertheless, information system developers and business analysts do not have sufficient knowledge about the detailed process modeling, because in the last decades, there was the need only for rough process models, which were performed manually.It is well known, the business process models are used as most valuable input for requirements capture phase of the information systems development. Poor quality process models can cause poor quality software requirements resulting poor information system. Therefore, greater emphasis on the business process-modeling education should be put.At University of Maribor, newer process-modeling techniques such as BPMN (BPMI, 2004) were introduced in the curriculum of information systems course (Rozman et al., 2003).Students at the Faculty of Electrical Engineering and Computer Science, 5th and 8th semester are taught how to model business and software processes. In the recent years, many modeling techniques were used for process modeling and many of them were not appropriate as a teaching method for process modeling. The process-modeling technique, which is appropriate for students (from teaching assistant point of view), should be easy to learn, it should hide the unnecessary details of process model and it should be broadly accepted.The BPMN notation was chosen, because it offers a graphical notation with the support of all important process concepts (process, activity, event, routing, merging, synchronization, messages, roles and so on), which can be used to model various kinds of discrete processes. Another reason for choosing this notation was that the specification of the notation defines mappings to the process execution languages, currently for BPEL4WS (IBM, 2005) and XPDL (WfMC, 2005).At the time of the adaptation of the BPMN notation for educational process (2002), there was no appropriate supporting tool. Therefore, simple Microsoft Visio stencil was created, containing BPMN graphical symbols, to help students at the practical exercises. At the time of writing of this paper, several other Microsoft Visio stencils already emerged. There are also some commercial tools developed, which shows the overall adoption of the notation in the industry is increasing.The goal of the practical lectures of process modeling is, to teach students how to design business and software process models using standards. The following standards are used: CMMI-SW (SEI, 2002) and ISO 9001:2000. As there are no guidelines how to combine standards and the process modeling lessons, our own teaching method was introduced. Experiences of last four years were used to improve the teaching method, which caused the improvement of the quality of students' software process models, too.The student's adaptation of the BPMN notation was followed in several ways: firstly, during the semester, interviews with the students were conducted and at the end of the semester, the survey about usability and usefulness of BPMN was made. Modified technology acceptance model (TAM) (Davies, 1986) questionnaire was used to gather students' opinions. There has been a lot of research work done in the field of adaptation of the new technologies using TAM, for example, see Huang (2005). This model was applied to a wide spectrum of new technologies, including data modeling techniques (Moody, 2002), wireless mobile technologies (Venkatesh and Ramesh, 2006) and so on. A little or no research work was done researching the usage of business process-modeling notation in the industry or education. The main reason is, the notation is new and the community is about to adopt it.The research presented in this paper represents part of the ongoing research about process modeling, which is conducted by our faculty.The rest of the paper is structured as follows. Second section describes the process of teaching the process modeling with this notation in educational process. Third section presents the improved teaching method for process modeling. In fourth section, the survey about BPMN usability and ease of use is described. Fifth section discusses the results of the survey and finally, sixth section concludes.\nThe process of teaching the process modeling: The environment and the limitations\nThe teaching process - improved version: Owing to student's difficulties when transforming CMMI-SW KPA to BPMN process model, the teaching process was improved (Figure 3) according to gathered and analyzed feedbacks.Key differences between first version (Figure 1) and improved version of the teaching method (Figure 3) are:* The new teaching process omits the other modeling techniques and focuses primarily on the BPMN.* Students model the configuration-management process without knowing or studying CMMI-SW standard first. They use their own ideas, limited experiences and brief information about the domain they receive from teacher. In parallel, they are also taught how to use version control system (sub-version) in practice.* Study of CMMI-SW KPA configuration management is introduced in later phases of process modeling. Students audit their process models according to standard and improve them.* Implementation of the document flow using Lotus Notes is skipped and transferred to other practical lectures, to allow more time for process modeling.We noticed also that the improved version of teaching process outputs better process models and improves student's understandability of specific software domains as configuration management, requirements development, process tracking and others.\nThe survey: At the end of last semester, a survey was conducted about the perceptions of usefulness and usability according to the TAM. The survey was used to answer the following two questions:1. Are the BPMN and supporting tool easy of use?2. Are the BPMN and supporting tool useful?According to the original TAM model there is strong correlation between perceived usability, ease of use and intention to use of the notation. If the notation is usable and easy to use, we can say there is a strong possibility the students will use it later at work, whenever there will be the need to model the business processes (for example, if their future work position will be business analyst or software engineer).In developing our measurement instrument for the survey, we followed the generic TAM model questionnaire, adapted it for specific usage. For the answers' scale, the standard seven level Likert scale was used (agree very strongly=-3, agree strongly=-2, agree=-1, undecided = 0, disagree = 1, disagree strongly = 2, disagree very strongly = 3).Questions were grouped into five groups:1. usability (Table I) where system X represents BPMN;2. usability (Table I) where system X represents supporting tool;3. ease of use (Table II), where system X represents BPMN;4. ease of use (Table II) where system X represents supporting tool; and5. the demographic questions.For demographic purposes, we asked the population for their education, sex, course, and time elapsed from the lectures.The survey was conducted at the end of the course and before the exam.\nThe results: The final sample consisted of 39 individuals (10 percent female and 90 percent male students, half of them from 5th semester and other half from 9th semester).This study has delivered the following findings. There is a positive attitude towards usability (Figure 4) and ease of use (Figure 5) of the modeling notation and the supporting tool (where level 3 means strong positive attitude, 0 means undecided and -3 means strong negative attitude). The average value of the answers for perceived ease of use and usability is around value 1 (agree).Results for the BPMN ad supporting tool (Visio plug-in) show that average value of answers are very similar showing (see paired columns, (Figure 4) and (Figure 5)), which suggest, students do not perceive the difference between them.Interestingly, the lowest positive attitude can be seen at the 5th question of the ease of use category (it is easy to become skilful in BPMN and the tool). Although the students perceive the notation and the tool useful and easy to use, they are undecided when asked about becoming skilful. Reasons for such results are currently unknown and will be analyzed in the future.It can be also speculated, if the research is conducted after the exam, this result (becoming skilful at using notation) could be better due to greater knowledge obtained when preparing for the exam.To sum up the results, our research suggests that there is positive attitude towards perceived BPMN usability and ease of use.\nConclusions: Usage of BPMN for process modeling practical lectures for undergraduate students showed itself as a good choice. It proved that notation is very expressive and easy to learn, which is mainly due to previous experience of the authors and students with process modeling in EEPC, Petri nets and UML activity diagrams notation.Problems, which arose from software process modeling, were not due to BPMN complexity, but complexity of the domain and modeling approach. We have found out that transforming CMMI-SW's KPAs to process model diagrams caused many problems to students which resulted in improvement of the teaching approach. Instead of introducing the standard SEI CMMI-SW in the early phases of process modeling, the standard was introduced when the draft of process model was already finished. The direct implication of teaching process change was improved student's satisfaction with their process-modeling knowledge.The results of the survey showed that there is a positive attitude towards usability and easiness of use of the BPMN notation and the modeling tool, which suggests the notation has a good chance to be used by students at their future work.We believe that teaching the students using BPMN is a right way to teach them business and software process modeling and help the notation to become better known and widely used.\n",
"role": "user"
},
{
"content": "Purpose: - One way to improve the quality of business and software processes is to make them compliant with standard. This paper seeks to explore the suitability of the BPMN and the teaching method for process modeling, which uses SEI CMMI-SW standard as input and produces software process models as output.\nDesign/methodology/approach: - The research approach consists of two parts. First, the students' responses during practical lectures of process modeling were gathered and studied. Then, the teaching process was modified and students' responses were studied again. The second part involves the survey about students' attitude towards BPMN usability and ease of use.\nFindings: - There are two key findings of the study. First, using the SEI CMMI-SW standard in early phases of software process modeling is not very effective, because it overloads the students with abstract requirements, which are included in the standard. The more effective way is to use the standard in later phases of process modeling, in the phase of assessment and improvement of the process model. The second key finding is that there is a positive attitude among students towards BPMN usage for process modeling.\nResearch limitations/implications: - There is no evidence that this teaching approach can be generalized beyond the population of information science students and academic environment.\nPractical implications: - The improved teaching method for business and software process modeling can be used in similar environments (practical lectures of process modeling). Although there is no evidence, quality managers from organizations can also benefit from the lesson learned and avoid a less suitable process-modeling approach.\nOriginality/value: - The paper is useful to all roles in education and businesses, who deal with process models and standards (teachers, business and software analysts, quality managers, process designers).\n",
"role": "assistant"
}
] |
Food safety practices in European TV cooking shows
|
[
"Education",
"Hygiene",
"Food safety",
"Consumer",
"Kitchen",
"TV cooking shows"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Even though the contamination of food can occur at any stage of food production and major responsibility lies with the producers, a large proportion of foodborne diseases are caused by foods that are improperly cooked or mishandled either in domestic kitchens or in food service establishments (Fotannaz, 2011). Taking into consideration that home prepared food has been suggested to cause a proportion of 9-90 per cent of food borne illnesses, the proportion varying among countries (FSA, 2000; Redmond and Griffith, 2003), it becomes obvious that it is necessary to find means of improving consumer understanding of food safety; one approach could be TV cooking shows.\n2. Materials and methods: In all, 19 TV cooking shows (Table I), which were broadcast during November 2012 to January 2013 in six European countries (Austria, Ireland, the Netherlands, Norway, Romania and UK) were analysed regarding the hygiene and food safety practices they promoted. The shows were randomly chosen, but covered all the days of a week and different hourly segments.\n3. Data analysis: The reliability of the data regarding food safety practices collected by the evaluators of the above mentioned TV cooking shows was checked using the Cronbach's a coefficient for multiple items (Clayton and Griffith, 2008; Saba and Messina, 2002). Cronbach's a assessed how reliable the results of the TV evaluation reports were to measure the same construct (food safety practices). This coefficient was calculated using the MINITAB 16 software for the data set after standardization (the mean values for each variable were subtracted from each variable value and the result was divided by the standard deviation of the values for each variable). Cronbach's a values approaching 1 suggest high internal consistency, while a benchmark value of 0.7 is commonly used to indicate that the items measure the same construct (Saba and Messina, 2002).\n4. Results and discussions: 4.1 Descriptive data\n5. Conclusions: Evaluation of the cooking shows by food safety professionals, using questionnaires combined with PCA analysis, enabled identification of similarities and differences in the cooking practices in TV shows.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to determine how well cooking shows promote safe food handling via TV and to suggest their use for providing good hygiene and good cooking practices examples for consumers.\nDesign/methodology/approach: Principal component analysis was applied for the multivariate statistical analysis of the cooking shows, the components being: personal hygiene, cross-contamination, cooking and storing practices and risk communication. Data were collected via a questionnaire special designed for the purpose of the study. The positive attributes were converted into numbers using a nine-point Likert scale. This conversion enabled ranking of the cooking shows as a function of the total results obtained and considering the best show as the one with the maximum score attained.\nFindings: Evaluation of cooking practices by food safety professionals highlighted the most frequent safety errors and poor practices that are disseminated by the TV shows.\nPractical implications: While the repetition of good food handling and cooking practices risks antagonizing viewers, an increase in occasional emphasis of good hygiene would be of benefit to domestic viewers and potentially improve food safety practices among the public.\nOriginality/value: This is the first study that gives an European perspective on presentation of safety practices during food handling and preparation in a range of TV cooking shows as it examines 19 such shows broadcasted in six European countries over three months. Adherence to food safety standards and introduction of a star rating system for safety practices in TV cooking shows is proposed.\n",
"role": "assistant"
}
] |
CEOs' regulatory foci and firm-level product innovativeness in competitive environments
|
[
"Ghana",
"SMEs",
"CEOs",
"Emerging economies",
"Regulatory focus",
"Firm-level innovativeness"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: A growing management literature recognizes a firm's ability to innovate as one of the crucial factors for survival (Quinn, 2000; Story et al., 2015), regional dynamism (Audretsch et al., 2006), sustainable competitive advantage (Carbonell and Rodriguez-Escudero, 2010), local economic development (Malecki, 1991) and economic diversity (Desrochers and Sautet, 2008). A major challenge for firms is the need to continually explore innovative activities and thereby remain entrepreneurial (Hornsby et al., 2009).\n2. Theoretical background and hypotheses: The purpose of this study was to investigate how a regulatory focus orientation at the CEO level affects firm-level product innovativeness. One of the key components of firm success is the extent of their innovativeness (Appiah-Adu and Singh, 1998; Deshpande and Farley 2004; Hult et al., 2004). Innovativeness is a term that has been applied both at product (Szymanski et al., 2007) and firm levels (Wang and Ahmed, 2004; Akgun et al., 2012). Studies of innovativeness at the firm level refer to \"the propensity of a firm to innovate or develop new products\" (Story et al., 2015, p. 47). Incremental innovation is more appropriate within the context of emerging markets where the enabling factors for radical innovation such as effective intellectual property frameworks are uncommon (Robson et al., 2012). As such, this study focuses on the incremental changes in practices rather than radical new products.\n3. Methods: 3.1 Study setting\n4. Results: Table III presents the means, standard deviations and correlations for all variables. As indicated in Table III, CEOs' level of promotion focus is positively associated with innovativeness, while a CEO's prevention focus is negatively related to innovativeness. All the hypotheses were tested using hierarchical regression (Aiken et al., 1991). The variance inflation factors (VIFs) for all regressions in the study were calculated to account for multi-collinearity. The VIF value is 1.80, which is lower than the threshold value of 10, indicating that multi-collinearity is not a major issue in our analysis (Aiken et al., 1991). The mean-centered values are then used to plot the interactions following procedures used in previous studies (Dawson and Richter, 2006).\n5. Discussion: Guided by the regulatory and upper echelons theories, this study investigates the association between a CEO's regulatory focus and firm-level product innovativeness and the conditions under which regulatory focus is most or least beneficial to firm-level product innovativeness. This study aims to advance theory on entrepreneurship and innovation in SMEs.\n6. Conclusion, limitations and future research: This paper investigated the influence of CEOs' regulatory foci on firm-level product innovativeness. In addition, this paper examined the moderating role of a firm's industry environment (competitive intensity) on these relationships. This theoretically derived research model, which links a CEOs regulatory foci, competitive intensity and firm-level product innovativeness, was validated by means of an empirical study of 257 SMEs. The paper contributes to both regulatory focus and upper echelons theory.\n",
"role": "user"
},
{
"content": "Purpose: Using arguments from the regulatory focus and upper echelons theories, this paper aims to examine the impact of a chief executive officer's (CEO's) regulatory foci (i.e. promotion and prevention focus) on small- and medium-sized enterprises' (SMEs') level of innovativeness and how these relationships are jointly moderated by intense competition.\nDesign/methodology/approach: The empirical analysis draws on survey data gathered from 257 SMEs in Ghana.\nFindings: The study findings indicate that a CEO's level of promotion focus positively affects the firm's engagement in innovation, while a CEO's prevention focus is negatively associated with the firm's innovativeness. The positive association between a CEO's promotion focus and a firm's innovativeness is enhanced under conditions of intense competition. Additionally, the negative relationship between prevention focus and firm-level innovativeness is attenuated under intense competition.\nResearch limitations/implications: This study relied on a single informant and also used subjective measures for the dependent variable. As such, individual respondents might have biased perspectives on firm-level product innovativeness. Future studies may use multiple informants to examine the causal links of the variables.\nPractical implications: The study's findings provide managers with a deeper understanding of how to achieve superior firm-level product innovation. The understanding of this issue can promote the development and maintenance of further entrepreneurial ventures in emerging economies.\nOriginality/value: The paper has a strong theoretical value as it pioneers research on the effect of CEOs' regulatory foci on firm-level innovativeness in competitive environments.\n",
"role": "assistant"
}
] |
Effect of economic vulnerability on entrepreneurial competencies among Malaysian micro-entrepreneurs
|
[
"Entrepreneurial competencies",
"Micro-entrepreneurs",
"Economic vulnerability"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Poverty, vulnerability and inequality appear to be emerging issues found in developing countries. Aside from those mentioned, at the global scale, we are affected by crime, corruption, economic crises, climate change, environmental pollution, diseases and natural disasters (Tran and Korflesch, 2016). Without doubt, many developing countries have been experiencing a high level of material deprivation and wide dispersion of individual well-being. Therefore, alleviation of poverty and reduction in inequalities has been at the top of the agenda amongst developing countries, as well as the Millennium Development Goals (MDGs) of the United Nations. Furthermore, the very concept of vulnerability (risk of experiencing poverty in future) has been vastly discussed, followed by international economic shock, such as global economic crises, which have, unfortunately, hiked the rates of poverty and hardcore poverty (Heltberg et al., 2015). As for Malaysia, despite the noted reduction rates in poverty and hardcore poverty (Al-Mamun and Mazumder, 2015; Ahmed et al., 2016), inequality in income distribution and socio-economic vulnerability to poverty has remained a threat among its low-income households, similar to other developed nations (Nair and Sagaran, 2015).\nLiterature review: Theoretical foundation\nEconomic vulnerability and conceptual competency: Conceptual competency refers to competency linked to various conceptual abilities that are reflected in the behavior everted by entrepreneurs (Man and Lau, 2000), whereas economic vulnerability denotes the risk of exposure to shocks, as well as the potential to fall into poverty because of limited income and assets to absorb shocks from uncertainties (Fisher and Weber, 2004; McCulloch and Calandrino, 2003). Conceptual competency aids in forming the competitive scope of a firm. In addition, this competency depicts the intuitive thinking, the innovativeness and the capability to assess risks and address issues that may arise from the varied spheres of an entrepreneur (Man and Lau, 2000). Therefore, conceptual competency could lead to superior performance that increases income and assets accumulation. Meanwhile, as economic vulnerability has been closely related to poor households living along the poverty line, this element of economic vulnerability is assumed to have a negative relationship with conceptual competency. Based on the aforementioned, the following hypothesis is presented:\nEconomic vulnerability and opportunity recognition competency: Opportunity recognition competency refers to a competency associated to recognizing market opportunities through various means and forms from the competitive scope of a firm (Man and Lau, 2000), whereas economic vulnerability is the risk of exposure to shocks, along with the potential to end up being poor because of financial constraint to overcome uncertain shocks (Fisher and Weber, 2004; McCulloch and Calandrino, 2003). The competency of recognizing opportunities aids entrepreneurs to identify, assess and seek business prospects. This competency is crucial in identifying and assessing market gaps and shifts in environment, apart from seeking new business opportunities via marketing and promotion, so as to ensure sustainable performance (Man and Lau, 2000). Hence, opportunity recognition competency may lead to a superior performance that adds to wealth accumulation. Meanwhile, as economic vulnerability depicts one falling into poverty, it is predicted to have a negative relationship with opportunity recognition competency. With that, the following hypothesis is developed:\nEconomic vulnerability and organizing competency: Organizing competency denotes competencies associated to the organization of various internal and external human, physical, financial and technological resources that help to establish organizational capabilities (Man and Lau, 2000), whereas economic vulnerability refers to the potential of falling into poverty because of failure in absorbing uncertain shocks (Fisher and Weber, 2004; McCulloch and Calandrino, 2003). The aspect of organizing competency reflects the ability to plan, organize, lead, motivate, delegate and control both internal and external resources. This competency seems to be essential among owner-managers in executing their daily duties that incorporate planning daily operations, acquiring and allocating resources, leading and motivating employees and delegating and establishing rules and regulations. All these ensure smooth daily operations, apart from ascertaining long-term performance (Man and Lau, 2000). Hence, organizing competency may lead to an increase in enterprise performance, whereas economic vulnerability is expected to have an inverse effect on human capital development; therefore, economic vulnerability is predicted to have a negative relationship with organizing competency. As such, the following hypothesis is given:\nEconomic vulnerability and relationship competency: Relationship competency is the competencies associated to person-to-person or individual-to-group interactions that generate organizational capabilities and competitive scope of firms (Man and Lau, 2000), while economic vulnerability refers to the probability of falling into poverty because of inability of addressing diminishing shocks (Fisher and Weber, 2004; McCulloch and Calandrino, 2003). Relationship competency denotes the ability to build and keep networks and relationships with both existing and potential stakeholders. This is done by using networks and relationships to acquire and strengthen resources and business opportunities; building and securing trust among stakeholders; and communicating, negotiating and managing conflicts (Man and Lau, 2000). This competency is essential for any owner-manager toward maintaining a good working relationship with the existing and potential stakeholders so as to ascertain sustenance of performance (Man and Lau, 2000). Therefore, relationship competency is assumed to lead to superior performance and wealth accumulation, whereas economic vulnerability is associated to the potential of falling into poverty. With that, economic vulnerability may have a negative relationship with relationship competency, as portrayed in the following developed hypothesis:\nEconomic vulnerability and strategic competency: Strategic competency refers to competencies linked to setting, evaluating and implementing strategies devised by firms (Man and Lau, 2000), while economic vulnerability is the potential of becoming poor because of failure in addressing uncertain shocks (Fisher and Weber, 2004; McCulloch and Calandrino, 2003). In fact, strategic competency denotes the ability of setting goals and taking action toward reaching the goals, which can be done by forming and creating competitive scope and organizational capabilities (Man and Lau, 2000). Such competency is viewed as the most integral competency that brings together two aspects of entrepreneurial task of forming and generating competitive scope and organizational capabilities (Man and Lau, 2000). Thus, strategic competency may lead to superior performance that adds to one's wealth, while economic vulnerability is related to living along the poverty line. Therefore, economic vulnerability is presumed to have a negative relationship with strategic competency. Hence, the following hypothesis is developed (Figure 1):\nResearch methodology: This study had adopted the cross-sectional design by using quantitative data gathered via structured interviews so as to examine the effect of economic vulnerability upon entrepreneurial competencies. The respondents were selected from the low-income and poor households interested in socio-economic development by getting involved in working capital and enterprise development training programs. As such, all details and lists regarding the development organizations and the respondents were retrieved from the eKasih National Poverty Data Bank. As a result, a list of 400 randomly selected low-income and poor households from Kelantan, Terengganu, Kedah and Perlis had been obtained. Prior to data collection, a team of researchers contacted the selected households to briefly explain the objectives of the survey and to set interview appointments. The process of data collection took two months: from October until November 2017. As a result, a total of 300 respondents had agreed for the research team to visit their premises and to interview them.\nSummary of findings: Demographic characteristics\nDiscussion and conclusion: This study had investigated the effect of economic vulnerability upon entrepreneurial competencies (commitment, conceptual, organizing, opportunity recognition, relationship and strategic competencies) among selected respondents of various development organizations established in Peninsular Malaysia from the lens of the RBV. The outcomes of this study revealed that economic vulnerability does have a negatively significant effect on commitment competency, opportunity recognition competency, organizing competency and strategic competency among respondents of various development organizations in Peninsular Malaysia (H1, H3, H4 and H6). Nevertheless, economic vulnerability displayed a positively insignificant effect on conceptual competency, while positively significant on relationship competency (H2 and H5). The significantly negative results seem to support and further extend the existing literature (Barney, 1991; Man and Lau, 2000) through the evidence provided in this study that economic vulnerability negatively affects entrepreneurial competencies, specifically, commitment, opportunity recognition, organizing and strategic competency.\n",
"role": "user"
},
{
"content": "Purpose: This study aims to investigate the impact of economic vulnerability upon entrepreneurial competencies (i.e. commitment competency, conceptual competency, opportunity recognition competency, organizing competency, relationship competency and strategic competency) among respondents from varied development initiatives established by the eKasih program (National Poverty Data Bank) in Peninsular Malaysia.\nDesign/methodology/approach: Upon adopting the cross-sectional design, data were randomly gathered from selected 300 micro-entrepreneurs from the list of development organizations available in the eKasih (National Poverty Data Bank), located at four states in Peninsular Malaysia. The quantitative data were gathered by performing structured interview sessions from September until November 2017.\nFindings: The outcomes of the study displayed that economic vulnerability has a significantly negative effect upon commitment, opportunity recognition, organizing and strategic competency. On the other hand, the results showcased that economi c vulnerability has a significantly positive effect on competency, but insignificantly positive impact upon conceptual competency.\nOriginality/value: These study outcomes appear to extend the scope of the resource-based view, apart from enriching the existing entrepreneurial competency literature, particularly within the Malaysian context. Hence, it is recommended that the government of Malaysia and development organizations should focus on maximizing the level of competency among micro-entrepreneurs as a viable approach to decrease the effect of economic vulnerability.\n",
"role": "assistant"
}
] |
Dissemination strategy for Lean thinking in health care
|
[
"Sweden",
"Lean production",
"Dissemination strategy",
"Implementation approach"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: To be affordable, accessible, safe, thorough, efficient and cost-effective, healthcare staff adopt different improvement concepts. Recently, there has been increasing interest in implementing Lean thinking in healthcare organisations (de Souza, 2009; Esain et al., 2008; Poksinska, 2010); e.g. Womack and Jones (1996) proposed how Lean-thinking principles, derived from industrial manufacturing, could be applied to healthcare and suggested foregrounding and understanding the patient journey through care as a starting point. Many healthcare Lean studies have been published (Fillingham, 2007; Jones and Mitchell, 2006; Kim et al., 2006; Miller, 2005; Spear, 2005; Young et al., 2004), with results that show Lean's potential for reducing waiting times and costs (Dickson et al., 2008; Jimmerson et al., 2005; Kaplan and Patterson, 2008; Nelson-Peterson and Leppa, 2007). However, improving the entire system, and not just individual departments remains a major challenge (Esain et al., 2008; Massey and Williams, 2005; Papadopoulos and Merali, 2008; Poksinska, 2010) since multiple caregivers with different contexts, often in different locations, need to be coordinated (Sorensen and Iedema, 2008). This all-encompassing thinking is difficult to accomplish in healthcare, which is to a great extent organised as independent working silos (Kim et al., 2007). The literature offers few hospital-wide or system-wide Lean transformation examples (Balle and Regnier, 2007; Kaplan and Patterson, 2008) and little is known about effective disseminating and implementing strategies in large healthcare systems (Morrow et al., 2012; Smith and Barnett, 2008). This paper, therefore, contributes to Lean-thinking dissemination strategies in multiple healthcare organisations and compares an Ostergotland county council case study (COO), where staff developed and implemented a regional Lean-thinking dissemination strategy, with work at the National Health Service Institute for Innovation (NHSI), England and Odense University Hospital (OUH) in Denmark. Our research questions are:\nTheoretical background: Lean implementation in healthcare\nMethod: Case description\nResults: During the first Lean implementation phase (2008-2011), dissemination strategy development and testing was the primary focus. The second phase focused on the Lean improvement programme dissemination throughout all healthcare units within the county council.\nDiscussion: Table I compares three Lean dissemination strategies. The first column shows CCO results. The second describes the NHSI strategy (Morrow et al., 2010), launched as a national development programme called The Productive Ward: Releasing time to care, which draws on Lean-thinking principles. The third describes the OUH strategy developed (Hospitalsledelsen, 2007). The Swedish and Danish cases are similar in that they represent a regional initiative, including large university hospitals. However, the Swedish programme initiator is the county council management group, whereas in the Danish case it is the hospital management board. The Productive Ward is different since it is a national programme involving several hospitals. Despite the different analytical units, the three Lean dissemination strategies show similarities.\nConclusions: There are case studies describing Lean implementation in single healthcare organisations. There is, however, less knowledge how Lean can be disseminated in multiple healthcare organisations. By studying the CCO dissemination strategy and comparing it to the benchmark case dissemination strategies (NHSI, OUH), we identify common activities:\nLimitations: We did not evaluate whether the CCO Lean dissemination strategy was successful since it is a long-term process that could not be evaluated at this stage. We do not know if the strategy will result in full-scale implementation. However, we note that the dissemination strategy resulted in Lean's adoption within the CCO. Another limitation is that the dissemination strategies we studied addressed different healthcare systems in different countries. The CCO system was regional, NHSI was national and OUH targeted one large hospital. This could influence the dissemination strategies and limit the comparison possibilities. Despite these differences, dissemination strategies that we identified were similar.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to contribute to knowledge about dissemination strategies for Lean thinking throughout multiple healthcare organisations.\nDesign/methodology/approach: The Ostergotland county council, Sweden (CCO) was chosen as a case study for an healthcare Lean-thinking dissemination strategies. Document analysis and interviews were used and results were compared with similar strategies employed by staff at the National Health Service Institute for Innovation (NHSI) and improvement in Great Britain and the Odense University Hospital in Denmark.\nFindings: The Lean improvement programme was introduced to tackle challenges such as an ageing society, rising care expectations and budgetary and economic constraints. It was designed as a long-term programme to create added value for patients and employee involvement. The dissemination strategy was: forming clear visions and objectives; piloting; training potential adopters; and formal dissemination. The CCO strategy was focused primarily on managers and was not meant to involve all staff until the implementation stage. Staff at the NHSI attempted to address nurses' needs during dissemination, which questioned whether the CCO managers' dissemination strategy is sustainable.\nPractical implications: This paper inspires healthcare managers and decision makers who aim to disseminate Lean production in their organisations.\nOriginality/value: There are many case studies describing Lean implementation in single healthcare organisations, but little is known about effective dissemination and implementation strategies in large healthcare systems. The authors, therefore, suggest activities for developing and implementing dissemination strategies in multiple healthcare organisations.\n",
"role": "assistant"
}
] |
Detecting customers knowledge from social media big data: toward an integrated methodological framework based on netnography and business analytics
|
[
"Customer knowledge management",
"Big data",
"Social customer relationship management",
"Business analytics",
"Netnography"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The advent of social media affected customers' behaviors and the interaction with companies by generating users' empowerment, ubiquity of the information and continuous connections and sharing and causing a paradigm shift in the conception and execution of business strategies, mainly for what concerns customers' relationship management (Kietzmann et al., 2011; Erevelles et al., 2016; Agnihotri et al., 2016; Kannan and Li, 2017; Moe and Ratchford, 2018). Social media regard all that applications and social network platforms belonging to the Web 2.0 technologies that allow knowledge generation, dissemination and communication among communities of users (Zeng et al., 2010; Zhang et al., 2015). They also represent new customer-centric tools and a privileged locus for the interaction at level of customer-customer, customer-companies and customers-crowd (Kietzmann et al., 2011; Kannan and Li, 2017). Social media are considered as a crucial driver of innovation that lead companies on the bandwagon of social media (Guesalaga, 2016).\n2. Literature background: 2.1 Managing customers relationships and knowledge assets in the age of social media\n3. Research methodology: To address to the previous research question, this paper adopts the methodology of a single case study as suitable investigative tool for the analysis of a contemporary phenomenon in its natural setting (Yin, 1994, 2003), by using of multiple types of data collection (Eisenhardt, 1989). Consistent with the qualitative research approach described by Strauss and Corbin (1990) and Yin (1994, 2003), case study methodology has been preferred, as the boundaries between phenomenon and context are not clearly evident, and it is relying on several sources of evidence and for developing grounded theory (Glaser and Strauss, 1967; Miles and Huberman, 1984). In coherence with the linear model of a case research (Piekkari, et al., 2010), in developing our study, we referred to the typical phases of a case study process (Yin, 2003). Accordingly, the following key decisions have been assumed in relation to phases synthetized into Table 1 that follows:\n4. Research findings: The case study of Aurora company discloses interesting evidence related to the importance of social big data analytics for CRM with its implications in terms of CKM according to the three dimensions (from, about and for customer) defined in the literature. The integration of business analytics with desk analysis demonstrated how the analysis of social media big data can contribute in the value creation process and support the firms gaining its competitive advantage.\n5. Social media customer knowledge discovery framework: Framed on the literature background and on the evidence coming from the case study of Aurora, this paper aims to propose a methodological framework based on netnography and business analytics to answer the need of managing the interactions with customers on social media and to create value from the different knowledge assets generated by customers in their virtual journeys.\n6. Conclusions and implications: Social media became an important tool for companies for their marketing strategies and for capturing important customer knowledge assets. Embedded into virtual environments characterized by complex networks of relationships, customers play an active role in the process of value creation by contributing to nurturing the so-called Internet of People as synthesis of the huge amount of data generated by users through their interactions on the Web. This huge amount of knowledge represents a challengeable asset to acquire, monitor and manage and calls for investigative frameworks characterized by cross-disciplinary and based on the integration of views related to social CRM, CKM, ICT and big data management. It is in such a perspective that this paper aimed to offer an original contribution at the advancement of debate in the field of KM and social CRM through the embracement of interdisciplinary approaches by presenting the results of a case study on Aurora, a well-known Italian brand operating in the field of fountain pens and a successful example of a company able to innovate in the respect of its long tradition. Built through the integration of Web-based desk analysis and business analytics, the case study has highlighted how the implementation of a successful strategy of customer relationship management in the context of social media big data requires the adoption of new tools and approaches of analysis able to leverage on the huge amount of data available online for addressing the marketing strategy of the company toward a successful implementation. Specifically, the evidence of the case have allowed to comprehend how the adoption of a social CRM requires the adoption of investigative approaches able to assure the effective management of relationship with customers on social media and the higher exploitation of the customer knowledge assets under the perspective of knowledge for, from and about customers. The case has demonstrated how the integration of business analytics into the traditional phases of data collection can support companies in achieving such an ambitious goal by allowing the definition of a clear demographic picture related to its customers, the comprehension of what they want, how they interact with others and the company itself.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to demonstrate how the integration of netnography and business analytics can support companies in the process of value creation from social big data by leveraging on customer relationship management and customer knowledge management (CKM).\nDesign/methodology/approach: This paper adopts the methodology of a single case study by using desk analysis, netnography and business analytics. The context of analysis has been identified into the case of Aurora Company, a well-known producer of fountain pens.\nFindings: The case demonstrates how the integration of big data analytics and netnography is relevant for the development of a customer relationship management strategy. The results obtained have been categorized according to the three main categories of customer knowledge, such as knowledge for, from and about customer.\nResearch limitations/implications: This paper presents implications for the advancement of the theory on CKM by demonstrating, as the acquisition, storage and management of data generated by customers on social media require the adoption of a cross-disciplinary approach resulting from the integration of qualitative and quantitative approaches. The framework is structured as methodological tool to detect knowledge in virtual community.\nPractical implications: Practical implications arise for managers and entrepreneurs in terms of value creation from knowledge assets generated on social big data through the management of the customers' relationship and data-driven innovation patterns.\nOriginality/value: This paper offers an original contribution of integration of well-established research streams. The focus on the knowledge under the perspectives of information assets for, from and about customers in the debate on value creation and management of big data is an element of value offered by this study in addition to the comprehension of strategies of social customer relationship management as actual initiative embraced by a company in the leveraging of innovation and tradition.\n",
"role": "assistant"
}
] |
Conceptualizing e-government service quality under credence based settings : A case of e-counseling in India
|
[
"Service quality",
"E-Government",
"Credence based service setting",
"E-Government service quality"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nI. Introduction: Prominent adoption of information and communication technology (ICT) by governments has been brought about through various schemes such as online government, digital government, one stop government, etc., and these are collectively known as e-government (Andersen and Henriksen, 2006). The transformation toward e-government practices promise to pave the way for the efficient and effective delivery of services in the public arena. Increasing adoption of e-government practices across the world has been a center of attraction for researchers from multiple research disciplines. Dissemination of online information, digital interaction, and electronic transactions are the central activities that can support government operations performed by various departments. It is acknowledged that the nature of e-government services may vary based upon the types of targeted stakeholders, or the nature of interactions with the concerned stakeholder (Jaeger, 2003). Governments use ICT to design services and connect to their own institutions at different levels of governance, employees, business entities, and citizens for delivering various kinds of services such as providing information and consultation, facilitating interaction and transaction, etc. There is a widely held belief across countries that the adoption of e-government can result in many practical benefits such as reduction of costs, growing revenue, transparent operations, increased accountability, and improved services for citizens (Jaeger, 2003; Affisco and Soliman, 2006).\nII. Motivation: Involvements of a \"third sector\" in co-production of public services are explicitly described by the scholars (Pestoff et al., 2006). The presence of another entity in the service delivery process can create a substantial shift in the simplicity of the core service context, and requires more understanding about what kinds of entities are involved, what kind of role they are playing, and their expectations regarding the service outcome. Consideration of such contextual distinctions of core service settings can make the e-government environment even more complex, and demands theoretical rigor to conceptualize the service interactions. Emphasis upon core service settings can play a critical role in identifying the key dimensions of service conceptualization, and subsequent identification of device appropriate techniques for improving the service quality. This paper attempts to distinguish the e-government service environment by differentiating the core service contexts from generic service environments. Moreover, is the authors attempt to bring insights from the extant literature of e-service quality to indicate the potential bi-polarity in the conceptualization of e-government services. An Indian case of \"e-counseling\" is discussed to understand the credence-based core service context, and the authors argue that it has a different kind of conceptualization of credence-based e-government services. The provision of this kind of study will supposedly increase the emphasis on re-modeling the e-government services based upon the extent of complexity existing in multi-party interaction.\nIII. Conceptualizing the core service context: When understanding the core service context, it may be implicitly assumed that the role of e-government is viewed primarily as complementing or facilitating access to the core services provided by the government, through existing public departments and agencies that have established back-office operations, by bringing part of the informational, interactional, and transactional activities online. The veracity of this assumption can be tested across most kinds of e-service environments designed by any government or business organizations. The assumption reciprocates the definition of \"e-services\" provided by Rowley (2006) that says \"e-service is deeds, efforts or performances whose delivery is mediated by information technology (including Web, information kiosks and mobile devices).\" In the light of this definition, the significance of core service settings should neither be trivialized, nor simplified beyond a certain extent. Essentially, the nature, composition, and complexity of core services should be of prime consideration when designing an e-service interface. Further, the role performed by the actors involved in the service delivery is also a critical concern for enhancing the overall service experience. We revisit the dominant perspectives on services to gain insights about the nature of core service settings.\nIV. Distinguishing the credence-based e-government service context: Essentially, there is a need to understand the core nature of e-government services from both process and information perspectives. Process perspective is concerned with how services are delivered and what kind of impression is projected in the minds of the consumer. Information perspective takes into consideration the information search behavior of consumers while they make service purchase decisions by choosing the best available service provider among multiple alternatives. Above all, the concept of co-production is critical to understanding the service environment from both process and information perspective. Against the backdrop of initial discussions about core service setting in the previous section, here our attention is focussed on distinguishing the credence-based e-government service environments from traditional e-government service environments.\nV. Conceptualizing service quality in a credence-based e-government setting: e-Government services need regular assessment, evaluation, and continuous improvement over time, in order to ensure citizens' satisfaction. Therefore, it becomes essential to understand what is meant by the quality of e-government services, what are their important dimensions that require attention, and how they should be evaluated. The consideration of credence attributes in e-government accentuates the need to advance our understanding of service quality and demands for a more suitable quality assessment framework.\nVI. The case: \"e-counseling\" in India: Before proceeding immediately to describing the case, it seems essential to understand some fundamental notions, and usage of the terminology relating to \"counseling.\" Therefore, we first present a brief literature review and then come to the case description.\nVII Conclusion: Credence-based core service settings justified the need for a separate consideration of this service environment, and were properly illustrated using the case of \"e-counseling\" in India. Against the backdrop of previous discussions, we summarized the scope of e-counseling as a class of e-government services that falls under a credence-based setting; dedicated informational support to facilitate the decision-making process for millions of students; an interactional support where students have to participate to exercise their choices with adequate convenience and flexibility in order to reserve seats at their preferred institutions/programs; a transactional support that can facilitate securing appropriate seats by partial or complete fee payments; an opportunity to provide information about heterogeneous service offerings from various government and private institutions; and a common service interface that connects the three groups of stakeholders: students, academic institutions, and government counseling boards. Grounded by the fundamental nature of e-services of a credence-based core service setting, the bi-dimensional conceptualization of e-counseling requires more weighting to be given to the information quality elements without ignoring the interaction quality attributes. A two dimensional framework of service quality evaluation has been proposed in this work that is aligned with the nature of the credence-based core service environment. Additional research is required to find specific determinants of service quality for both the dimensions that could explicitly capture the information search behavior of consumers for high credence core services.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to acknowledge the potential complexities in conventional notions of e-government service environments, which have earlier been conceptualized as a two party interaction system where the government is the service provider and other entities such as businesses, citizens, etc. are the service recipients. In this regard, the authors extend their attention to a core service context and identify the existence of \"credence-based\" service setting that may involve more than two parties such as government, citizens, and third-party institutions (TPIs).\nDesign/methodology/approach: An in-depth literature review is performed to build a conceptual foundation for the proposed service environment. A case of an Indian context is exemplified to understand the credence-based service setting and advocate the suitability of broader dimensions of service quality assessment, and propose a preliminary model of service quality assessment in the given context. The case study approach is adopted for gathering and analyzing the data. The data are collected from government web sites, semi-structured interviews with government officials, and students.\nFindings: The paper finds that the existence of TPIs changes the nature of normal government to citizen (G2C) service interaction. Service quality conceptualization for this particular kind of e-government setting is bi-dimensional in nature which involves two separate but interdependent constructs quality of information, and quality of service Interaction that collectively contributes to service quality and empowerment of the recipients.\nOriginality/value: The present work attempts to differentiate the service context of credence-based e-government setting with the traditional conceptualizations of e-government service environments.\n",
"role": "assistant"
}
] |
Investigating current paradigms in supply chain risk management - a bibliometric study
|
[
"Research methods",
"Literature review",
"Supply chain risk management",
"Co-citation analysis",
"Bibliometric study",
"Supply chain risk’s"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Supply chain risks (SCRs) appear to be a major problem as unsettled economic environments. High volatility in supply and demand, coupled with rapid technological changes, and unforeseeable disturbances, are common issues these days (e.g. Barry, 2004; Christopher and Holweg, 2011; Tang and Tomlin, 2008; Waters, 2007). Judging by the increased number of articles published in scientific journals and germane trade periodicals, supply chain risk management (SCRM) is an area of growing importance (Chopra and Sodhi, 2004; Trkman and McCormack, 2009). The globalisation of markets, increased outsourcing and offshoring, reliance on external sources of know - how and new solutions, coupled with the emergence of information and communication technologies required to coordinate extended supply chain (SC) networks, are just a few reasons why this topic is relevant today (Lockamy and McCormack, 2010; Nasarimhan and Talluri, 2009; Pujawan and Geraldin, 2009; Ritchie and Brindley, 2007a, b). SCRM can be seen as a strategic activity undertaken by companies as SCRs can directly affect, business, financial, and operational performance of organisations (Nasarimhan and Talluri, 2009). These trends have expressed themselves through an increase of best cost country sourcing and collaboration with global SC partners (e.g. Lee, 2002; Martha and Subbakrishna, 2002; Norrman and Jansson, 2004). In an industrial environment mired in considerable uncertainty and complexity, firms are obliged to manage their SC in an efficient manner to advance reactivity and flexibility (Christopher and Lee, 2004; Manuj and Mentzer, 2008a, b; Peck, 2005; Wagner and Bode, 2008). As a consequence, risks, disruptions, disturbances, and uncertainties are evidence that the field of SCRM is ever broadening, and the theories, concepts, processes, and practices must develop with it (Cavinato, 2004; Neiger et al., 2009; Tang and Musa, 2011; Thun and Hoenig, 2011). Although the significance of SCRM research is not in question, diverse viewpoints concerning its intellectual contributions to the numerous research areas it relies on do exist (Pfohl et al., 2010). We aim to contribute to the SCRM literature by examining the intellectual structure of SCRM for improved understanding of its origins and its state of the science.\n2. Methodological background of the research: 2.1. Bibliometrics\n3. Research methodology and data collection: The general process for co-citation analysis adopted in this research follows a six-step approach (cf. Figure 1): selection of the database that contains bibliometric data and the journals to be included in the analysis; identification of the most relevant work within the area of study by using articles as unit of analysis; compilation of raw co-citation matrix to calculate similarity between pairs of referencing articles, using co-citation counts; multivariate statistical analysis of the co-citation matrix using multi-dimensional scaling (MDS), cluster analysis, and factor analysis; output preparation by using visualisations; and use of the resulting visualisation for discussion and evaluation (Boyack and Klavans, 2010; Chabowski et al., 2013; Zupic and Cater, 2014). Once the results are at hand, the analysed articles are allocated to the different paradigms identified.\n4. Results and discussion of the MVA: 4.1. Discussion of the results: multi-dimensional scaling and cluster analysis\n5. Conclusion and implications: Several conclusions and implications may be extracted as a result of the conducted analysis. Moreover, this research opens the door for further research endeavours. Accordingly, the subsequent conclusions and implications respond to the research objectives mentioned at the beginning of this paper. The main objectives were to: first, outline the intellectual and conceptual structure of scientific research on SCRM as embodied in the scientific literature analysing the intellectual structure of this discipline to date (i.e. shared or complementary methodological approaches, findings/results, etc.); second, define research streams that form both the intellectual structure and knowledge groups in the field as well as to reveal any relationships between these subfields; and to chart and illustrate the intellectual arrangement in a bi-dimensional sphere in order to evaluate identified clusters to support future academic research endeavours. In the following, we will present the conclusions that address these research objectives:\n6. Limitations and future research: With this paper, the latent structure fundamental to the SCRM literature was explored. The assumptions for the current investigation are that: research on SCRM is a foundation for the understanding of risks inherent network environments; and bibliometric methods provide an appropriate device for analysing the intellectual structure of a given literature. Co-citation analysis is especially enlightening in diagramming out the intellectual structure and research streams of connected colleges (i.e. the existence of an invisible school of thought) in the SCRM research. Findings in the study at hand can support the community of SCRM scholars to identify the core literature that may function as chief wellsprings of motivation for defining new research inquiries and developing new viewpoints. A shortcoming in co-citation study is that authors here and there refer to each other not so much because they have used each other's work for their research, but for some very different reasons. However, a recent research shows the growing significance of citations because references to citation counts may be an integral part of competitive contests among researchers (Aksnes and Rip, 2009). Different diverse intentions may induce citations (Gundolf and Filser, 2013), and, possibly, some will undermine the assumption that co-referencing is an appropriate measure of the resemblance and scholarly connection between articles (Hoffman and Holbrook, 1993).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyse the intellectual structure and research fronts of discipline of supply chain risk management (SCRM), in order to identify the knowledge groups in the research area to date, as well as to reveal any relationships between these subfields and the central influential trends.\nDesign/methodology/approach: By means of a bibliometric study, the 32 most co-cited articles on SCRM published in 16 top business-related academic journals are analysed using multivariate statistical techniques, i.e. multi-dimensional scaling, cluster analysis and correspondence factor analysis.\nFindings: The results demonstrate a clearly identifiable structure as a result of the performed co-citation analysis. The conducted cluster analysis and factor bring forward that the research field is arranged in five different areas of interest: explaining supply chain (SC) risk phenomena, concepts, frameworks and insights of SCRM; modelling risks for SCs; inventory risks affecting supply efficiency; SC and product design methods; and SC risk mitigating strategies.\nOriginality/value: Overall, the intellectual structure of SCRM is first examined through a bibliometric approach using quantitative techniques - for improved understanding of its origins, and to identify the state of the science - and to offer suggestions for future studies that could cover current gaps. This study represents the potential to advance the SCRM literature landscape.\n",
"role": "assistant"
}
] |
The winemaker as entrepreneurial marketer: an exploratory study
|
[
"New Mexico",
"Participant observation",
"Opportunity",
"Risk",
"Emerging wine region",
"Entrepreneurial marketing",
"In-depth interviews",
"Qualitative research",
"Winemakers"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The USA is the fourth largest wine producing nation in the world (Stevenson, 2005), with California accounting for 89 per cent of annual volume (Alcohol and Tobacco Tax and Trade Bureau [ATTTB] Report, 2011). Each of the 50 states and the District of Columbia is home to at least one winery (Gilinsky et al., 2010; Shaw and Bahney, 2003). The majority of these wine regions are fragmented with more individual and family-owned businesses than corporate wineries. Due to scarcity of extant knowledge of winemakers' life-worlds in developing and smaller wine regions in the USA (Dodd, 1995; Edwards, 1989; Spawton, 1989), wine business researchers have called for further studies on smaller wineries to understand the producers' side of the business (Beverland, 2000; Mora, 2006; Richardson, 2004). In wine research, despite multitude studies on consumers' perspectives, there is a marked lack of knowledge of the winemakers' viewpoint. Given that small wineries are prolific in the USA (ATTTB Report, 2011), this research contributes to extant knowledge by investigating ways in which small-scale winemakers sustain livelihoods. Scholars have outlined the need for small wineries to account for sustainable marketing strategies for survival in the long run (Beverland, 2000; Mora, 2006), yet little is known about how winemakers achieve (or fail) in this regard. To answer this call, this study investigates New Mexico (NM), an emerging wine region with challenging conditions where the winemaker is compelled to take bold action to simultaneously support production and create value for their customers.\nResearch method: Six winemakers located throughout NM with various backgrounds in terms of previous experience, heritage, winemaking knowledge, length of time in the business and overall business skills were recruited through purposive sampling (Lincoln and Guba, 1985). As the NM wine region is relatively small, to protect the identities of our participants, we offer a general description and use pseudonyms. Although modest, our sample offers a diverse group of people traversing the length and breadth of the state. Four wineries are located in central (one winery) and northern (three wineries) part of the state where weather and soil quality challenges are dissimilar. Similarly, the southern part of the state where we interviewed two winemakers face water and weather challenges unique to the region (e.g. windy season). Two of the winemakers had been in business between 15 and 20 years, and four wineries were within 5-9 years of start-up at the time of data collection. Between them, participants produce from a handful blends to over 70 blends per year. Some wineries operate under one label; others have over five labels.\nConclusion: This study contributes to existing knowledge by demonstrating underlying EM dimensions in winemaking business practices of an emerging wine region. Unlike other products where there may be an initial start-up period which goes onto steady operational processes with maturity and growth, winemaking requires a continuous learning commitment. Winemaking in NM is challenged with continuous uncertainties such as availability of grapes, steady flow of customers and acquiring and managing finances. Hence, irrespective of number of years in business, pursuing opportunities through proactiveness and innovation, minimizing risks through resource leveraging and creating value through customer interaction are often carried out with the basic idea of survival. NM winemakers exhibited gregarious tenacity to learn from mistakes to ensure business success. Such a dense inter-relationship amongst EM dimensions may be attributed to the need to survive given the uncontrollable risks associated with undertaking the wine business in a challenging region.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to examine, through the lens of entrepreneurial marketing theory, everyday business practices of small-scale winemakers in a challenging small wine region in the USA.\nDesign/methodology/approach: In-depth interviews and participant observations were utilized to elicit rich descriptions of entrepreneurial marketing efforts of six New Mexico winemakers.\nFindings: This article describes winemakers' entrepreneurial marketing efforts. We find that survival drives all other entrepreneurial marketing dimensions where accounting for risk is pervasive rather than a stand-alone dimension. Knowledge gained from intense customer focus is used for new product and service innovations. The leveraging of individual and shared resources is another dimension of entrepreneurial marketing that is demonstrated in our analysis.\nOriginality/value: As entrepreneurs stake their claim in developing small wine regions, understanding entrepreneurial marketing concepts will enable academics and practitioners to understand challenges of a business that is not only dependent on the economics but also on mother nature's whims.\n",
"role": "assistant"
}
] |
Customer relationship management (people, process and technology) and organisational culture in hotels: Which traits matter?
|
[
"Organizational culture",
"Hotel industry",
"Critical success factors",
"Customer relationship management",
"CRM components",
"People",
"Process and technology"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Customer relationship management (CRM) is a successful marketing strategy, which has been proven to aid in improving business performance, customer satisfaction and customer retention (Abu Kasim and Minai, 2009; Mohammed and Rashid, 2012). There is an ever-increasing use of CRM in tourism and hospitality businesses, and hotels as a part of the service industry largely applies CRM strategies to attain greater profits (Sarmaniotis et al., 2013). Wu and Lu (2012) in their study found that the successful implementation of CRM projects in hotels not only increases customer satisfaction but also has a significant and positive impact on business performance and customer lifetime value. Lo et al. (2010) mentioned that CRM plays a vital role in achieving the hotels' main objectives, which are increasing guest satisfaction, loyalty, retention, reducing guest acquisition costs and increasing profitability.\nLiterature review: Customer relationship management in hotels\nResearch method: Questionnaire and sampling\nResults: Demographic profile\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to examine the impact of four organisational cultural traits of adaptability, consistency, involvement and mission on the three components of customer relationship management (CRM), namely, people, process and technology, in the context of the hotel industry.\nDesign/methodology/approach: Required data are collected with a quantitative approach and a questionnaire adapted from the Denison organisational culture survey and the Mendoza CRM model. The questionnaire was distributed among 364 managers of a chain hotel in the UK and gathered data were examined using the structural equation modelling method.\nFindings: The results of this study reveal that the four traits of organisational culture (adaptability, consistency, involvement and mission) have positive and significant impacts on the three components of CRM (people, process and technology). A set of theoretical contributions and practical implications was also discussed.\nResearch limitations/implications: The study is conducted with a case study approach; hence, the findings cannot be generalised to a larger population, and the results might be different for other industries. Because of the limitation of access to all employees, only managers were selected as the sample, and future studies with all employees may show different results.\nPractical implications: Current study helps hotel managers to understand the role and importance of organisational cultural traits in successful implementation of their CRM strategy components.\nOriginality/value: The position taken in this study recognises the need to enhance the understanding of organisational culture's impact on implementing CRM components. Organisational cultural traits have different levels of impact on CRM implementation, and this is the first study to investigate the detailed impacts of the four traits of adaptability, consistency, involvement and mission on the three components of CRM, namely, people, process and technology.\n",
"role": "assistant"
}
] |
On the criteria of content web search enablement
|
[
"Content management",
"Information searches",
"Internet",
"Intranets",
"Worldwide web"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: 1.1 Overview\n2. The enablement model: 2.1 The model's block diagram\n3. Hybrid indexing process: Indexing always starts with building the searchable collections. Categories and knowledge trees are then created on the top of these collections. An indexing process depends directly on the nature of the data repositories required to be indexed into searchable collections and knowledge trees. The following cases were documented during work with the indexing process:* Case 1: When only file system documents are indexed - The indexing process would be simple as most of the available commercial indexing tools such as Verity's vspider and mkvdk indexing tools (Verity, 2001a, b) can directly handle this format. In this case, we need to know the directory names where these documents are stored or the list of pathnames for these documents.* Case 2: When only HTTP documents are indexed - The indexing process would be provided with the start points (http addresses); the indexing tool will crawl the web for subsequent linked documents as with the Verity vspider indexing tool (Verity, 2001b). Another way is to provide the indexing process with the list of all http addresses and then feed that list to the indexing tool, as with the Verity mkvdk indexing tool (Verity, 2001a).* Case 3: When only database metadata are indexed - The indexing process would involve using a custom gateway program that retrieves the metadata content from the database and stores that into a file on the file system. Then the retrieved content can be fed to the indexing tool to build the searchable collections. The content would be stored as XML/SGML dump, or as Verity bulk insert files (BIF) (Verity, 2001a).* Case 4: When both database metadata and documents are indexed - The indexing process would involve using a custom gateway program that retrieves the metadata content from the database. It is assumed that there is a clear association between the documents and the metadata. In this case, both metadata content and the content of these documents can be merged into hybrid content; the content is then dumped into XML, SGML or BIF, as in Case 3 above, before feeding the content to the indexing tool.In Case 4, the association between a metadata record and a document would be in the form of having the path or link to the document recorded in one of the record's fields. Also, the last case is considered a good way for monitoring the source content. In this case, the searchable indexes must always be updated by adding new records/documents to the index, i.e. re-indexing modified records. This may also entail removing records (from the searchable index) that were deleted (or marked to be deleted) from the database.Any indexing process would perform the following steps:1. Defining the content repository. Is it:* file-system documents?* web-based http documents?* database metadata?* database metadata and file-system documents?* database metadata and web-based http documents?* database metadata, file-system documents, and web-based documents?1. Defining (in the cases of file-system and web-based documents) all meta-tags that need to be captured from the documents during the indexing time. In this case, it should define all these meta-tags as additional fields to the indexing style files before running the indexing tool.2. Preparing and running an appropriate script (gateway program!) that can retrieve the source content. Such content can be from the local file system, the web, and/or databases. The retrieved content should be stored on the local file system in a pre-defined format (XML, SGML, BIF, etc.) that can be fed to the indexing tool.3. Feeding the stored retrieved content to the indexing tool to produce the searchable collections.An indexing process should result in one or more searchable collections. Categorization and knowledge trees can be produced on the top of these collections. Such knowledge trees can speed up the search process (see Figure 3).\n4. Configuring the search brokers/servers/collections: Search brokers do not perform the actual search, but manage search requests by passing them to selected search servers based on the availability and load of these servers. Therefore, the only configuration required for brokers is defining all attached search servers to that broker by defining the host TCP/IP (or the host DNS) and the port. When the broker passes the search to a search server, care must be taken to target the server that has access to the appropriate collection, while balancing the load. Also, each collection should be built based on a predefined set of custom style files. Such files define all custom fields and zones that a collection would have. A collection field can be searched as well as retrieved. A zone, on the other hand, can be search, but only fields associated with that zone can be retrieved. In general, searching zones is faster than searching fields.Figure 4 shows a sample style file that shows the list of fields to be captured by the indexing tool from all indexed documents during the indexing operation. Figure 5 shows a sample server configuration file that shows the list of collections to be opened by the server.\n5. Search process: As defined previously, search is the process of receiving user's queries (search requests), massaging their queries to enhance the search accuracy, and/or restrict the search requests to particular groups of the indexed documents, passing the massaged query to the search engine, receiving the search results from the search engine, and processing these results to display them back to the user. In general, a search process should remain transparent to the user. The user does not need to know where the search server is located. As described above in the Configuration section, the web server should be able to select the search server with the least load and maintain the search load distribution across the multiple search servers in the back end of the enterprise system. This means that the web server needs to be able to auto-determine which search server to select, and if the server of choice is down an alternative one is selected (fail-over scenario).After submitting the search query, the query can be managed such that the search scope is affected. For example, if the user is eligible to search only certain category of documents, then the search request would be enforced to search through that category of documents only (i.e. query becomes query docCategory categoryX).\n6. Developing the middleware component: The main objective of having a search middleware component is to integrate the front-end web pages with the back-end search servers. Such a component can be developed in any programming language that suits the web environment. Figure 6 shows the input/output parameters for a sample interface specifications for a middleware component. This sample was implemented in Java-based web pages (e.g. ATG Droplets; Art Technology Group, 2001a, b, c, d).Generally speaking, the development of a middleware component would require the following:* defining the components of the enterprise system (i.e. indexing tool, search engine, web server, application server, etc.);* finding out the programming languages supported by the web and application servers;* finding out the programming languages supported by the search engine;* developing a component that enables the web server to connect to the search engine, submit search requests, and receive search results; and* placing the developed component somewhere on the system such that it is accessible by the web-server.\n7. Criteria for selecting the enterprise components: The following factors would help decide which enterprise components to select for an enterprise system (i.e. web server, web-application servers, indexing tool, search engine, database servers, etc.):* Skills required to operate components - If the development environment is dominated by Java, then the web server would have to have Java Support. There are many web application servers that support Java to choose from such as ATG (Art Technology Group, 2001a, b, c, d), BroadVision (see www.broadvision.com), Tomcat (see http://jakarta.apache.org/tomcat/index.html), and many others.* Cost of components.* Ease of use of components.* Reliability of components.* Time since the first release of the components - the longer a component has been on the market, the better the chance that the component is defect-free.* Availability and quality of the vendor's technical support.* Programming languages support - Both the web application server and the search engine should support a similar programming language. Otherwise, integrating the two may prove difficult.\n8. Distribution of the enterprise components: With an enterprise system, the system's components do not need to be in one location. Indexing, search, database, web servers, and application servers may run on separate machines, and probably in different locations. Different locations can be within the same building, the same city, the same country, the same continent, and/or even in different continents. The following are the main issues that need to be considered when having the enterprise components running in different locations:* Indexing performance - Indexing local documents would be faster than indexing database metadata content as well as faster than indexing contents via the web server (http documents).* Search performance - Dealing with a search engine that is located within the same LAN would be faster than dealing with a search engine that is located in a different network.* Security - It is important to keep all search brokers/servers within the organization's firewall. In this case, when a customer searches for something, the search request is directed by the web-server to the search engine. The search engine performs the search operation and returns the results back to the web server.Generally speaking, the security layer of any enterprise system handles all security concerns (i.e. encrypting user IDs, passwords, and personal information, etc.). The search process can restrict search requests to particular groups of indexed documents/metadata.\n9. Conclusions: A general model for enabling content search on the web was investigated in this paper. The following indexing issues were discussed:* indexable content resources;* content retrieval and management; and* extracting hybrid content from documents and databases into the same collection (i.e. searchable index).The following search issues were also discussed:* cooking queries before submitting search requests;* building middleware components to communicate search requests and results with the search server;* configuring indexing tools and search engines;* managing the distribution of search servers across different locations; and* handling search security and performance.\n",
"role": "user"
},
{
"content": "Purpose: - To present an ongoing research effort based on previous industrial experience in the content management field.\nDesign/methodology/approach: - The project began with lessons learned while working in the IT industry. A model was then developed based on these lessons for managing content and enabling searching for content over the internet.\nFindings: - The author has used the proposed model both before and after joining academia a few years ago. The model works fine with some tune-up from one case to another.\nResearch limitations/implications: - As this is ongoing research, the model will be subject to enhancements based on further lessons learned in the future.\nPractical implications: - The proposed model has been used by the author on several projects and has led to good results in most cases.\nOriginality/value: - The model itself and the approach followed to integrate content into enterprise systems are both novel.\n",
"role": "assistant"
}
] |
The impact of corporate social responsibility in food industry in product-harm crises
|
[
"Corporate social responsibility",
"Food industry",
"Product‐harm crisis",
"Social responsibility"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Corporate social responsibility (CSR) has emerged, in the last two decades, as both an important academic construct and a pressing corporate agenda item. Therefore, a fairly large body of literature, both theoretical and empirical, has investigated the role of CSR in marketing and consumer behaviour; with almost all of these researches focusing on routine and everyday situations. However, the value of CSR is also challenged during non-routine situations, like product-harm crises, with a notable lack of research regarding the role of CSR in product-harm crises in the food industry. Accordingly, the examination of CSR as an antecedent assurance factor in times of crisis has critical importance, considering the high vulnerability of the food industry to product-harm crises.Food industry is highly vulnerable to crises and this explains why business history is full of crises caused by unsafe alimentary products. Crises in the food industry could be related to a specific brand and company or generalized to the whole product category or industry. Indeed, regarding the product-harm crisis for a specific brand, the case of Salmonella in Cadbury's chocolates comes easily to the minds (Carroll, 2009). On the other hand, the Bovine Spongiform Encephalopathy (BSE) ruffled the hole food industry at the beginning of the new millennium (Penning et al., 2002). Generally, the aforementioned cases prove that it is critical not only for the academics but also for the managers in the food industry to be able to detail the factors that shape consumer behaviour after food industry's crises. Given the globalization of the food supply, it seems utopian the restriction of crises in a local area or to a specific brand (Lindgreen et al., 2009; Lindgreen and Hingley, 2009).The general aim of this article is to shed light onto the role of CSR in food industry's product-harm crises. The current study utilized an experimental design to test the proposition that CSR has an impact on the perceived degree of danger, attribution of blame, brand evaluation and buying intention, after a product-harm crisis. Furthermore, the moderation effect of CSR importance ascribed by the consumers on the above relationships was examined.\n2. Literature review: 2.1 CSR in the food industry\n3. Conceptual framework: On the basis of the above and by using Klein and Dawar's (2004) idea of halo effect of CSR on consumer behaviour, the current study examines the effect of prior CSR information during a product-harm crisis on the perceived degree of danger, attribution of blame, brand evaluation and buying intentions in the food industry.There is only one research in literature concerning the relation of CSR with the consumers' perceived degree of danger of the crisis, that of Vassilikopoulou et al. (2009), concluding that CSR performance has a negative impact on the perceived degree of danger. This result is consistent with a previous research of Siomkos and Kurzbard (1994) indicating that the products of a company with a positive reputation, will be perceived as less dangerous. Based on previous findings regarding the positive effects of previous CSR performance on decreasing the perceived danger by the defective food product we predict that:H1. The crisis event will be perceived to have lower degree of danger when prior CSR is positive vs when prior CSR is negative.Concerning the attribution of blame, Klein and Dawar (2004) found that a company with a positive CSR will be blamed less than a company with negative CSR performance or a company with no CSR related information. Arguably then, this leads to our second hypothesis:H2. The attribution of blame will be lower when prior CSR is positive vs when prior CSR is negativeAs literature review reveals, CSR has a positive impact on brand evaluation after product-harm crises. As Dean (2004) pointed out, CSR has a critically important impact on the company's image, not only before the attribution of blame but also afterwards. Likewise, according to Klein and Dawar (2004) a company with positive CSR performance has higher brand evaluation because of the direct relationship of the two aforementioned variables and the indirect positive effect through the lower attribution of blame. Moreover, de Matos and Rossi (2007) suggest that the evaluation of a product after a recall decision is favourable in the case of a company with good CSR performance. Finally, Vassilikopoulou et al. (2009) suggest that previous CSR performance has a positive impact on the general impression of the company involved in the crisis. Consequently, we propose that:H3. The brand evaluation will be higher when prior CSR is positive vs when prior CSR is negativeRegarding buying intentions, Klein and Dawar (2004) claim that there is only an indirect relationship between CSR and buying intentions after a product-harm crisis because of the lower attribution of blame and more favourable brand evaluation. Furthermore, de Matos and Rossi (2007) have found that CSR has a positive impact on buying intentions after product recalls, a result that is replicated by Vassilikopoulou et al. (2009) in product-harm crisis. Thus, we predict that:H4. The buying intentions will be higher when prior CSR is positive vs. when prior CSR is negative.The results of previous researches indicate that the importance ascribed to CSR by consumers is a significant moderator of consumers' reactions for different CSR initiatives (Sen and Bhattacharya, 2001). The same result is found in Klein and Dawar's research (2004) regarding product-harm crisis situations, concluding that the perceived general CSR importance of companies moderates the relationship between attribution of blame and prior CSR. In accordance with these findings, our general hypothesis is that the importance of CSR that is ascribed by the consumers will moderate every aspect between the CSR of the company and perceived degree of danger, attribution of blame, brand evaluation and buying intentions.H5. The CSR importance level ascribed by the consumers will moderate the relationship between CSR performance and perceived degree of danger.H6. The CSR importance level ascribed by the consumers will moderate the relationship between CSR performance and attribution of blame.H7. The CSR importance level ascribed by the consumers will moderate the relationship between CSR performance and brand evaluation.H8. The CSR importance level ascribed by the consumers will moderate the relationship between CSR performance and buying intentions.\n4. Methodology: 4.1 Design and measures\n5. Results: 5.1 Manipulations checks\n6. Discussion: It has been suggested that CSR has a positive impact on consumers' attitudes and reactions for the companies, their products and brands (Hess et al., 2002; Fombrum et al., 2000; Bhattacharya and Sen, 2004), something, which is supported by previous experimentation (Brown and Dacin, 1997; Handelman and Arnold, 1999; Sen and Bhattacharya, 2001). The role of CSR in routine and non-routine situations is well documented in CSR and crisis management literature. However, the food industry has not drawn researchers' attention. This study fills the research gap that exists in the food industry literature, by indicating the importance of CSR for the food industry in product-harm crises with an integrative approach.The basic assumption of this study is that a socially responsible food company can cope with a product-harm crisis in a better way and will face milder consequences as well. This was tested with the four hypotheses made above. Therefore, this study clearly highlights that CSR has an impact on the attribution of blame, brand evaluation and buying intention but not on the perceived degree of danger. The second basic assumption is that the positive consequences of CSR after product-harm crises are observed only when it comes to consumers who have a high CSR importance evaluation. Moderation tests revealed that CSR importance has an effect on the relationship between CSR and blame attribution, brand evaluation and buying intention. However, the moderation hypothesis did not provide support for the relationship between CSR and the perceived degree of danger.Regarding the perceived degree of danger, our hypothesis was not confirmed because the results revealed that respondents react only in the condition of negative information, such as negative CSR performance. The explanation is that positive CSR information is not strong enough to create a \"halo effect\". On the other hand, consumers feel that the crisis is more dangerous because of the negative CSR performance. Our results contradict those of Vassilikopoulou et al. (2009) and this differentiation can be explained by the different products involved in the product-harm crisis situation. Our research focuses on food products and especially on margarine, while Vassilikopoulou et al. (2009) explore the impact of CSR in a product-harm crisis of mobile phones.The results of the experiment show that the prior CSR performance has an impact on the attribution of blame, something, which is consistent with the findings of Klein and Dawar (2004). It seems that the halo effect of CSR appears not only in product-harm crises like the one described in Klein and Dawar's research involving lubricants, but in the food industry as well. Likewise, consumers evaluate higher the brand, after a product-harm crisis, when they are presented with positive CSR information. This falls in line with the research stream suggesting that CSR has an important role to play in crisis situations (Dean, 2004; Klein and Dawar, 2004; de Matos and Rossi, 2007; Vassiliopoulou et al., 2009). Finally, buying intentions are higher in the positive CSR condition than the negative CSR condition, something, which is also validated in the food industry (Dean, 2004; Klein and Dawar, 2004; de Matos and Rossi, 2007; Vassiliopoulou et al., 2009).Previous researches support the existence of CSR impact on consumer behaviour, inasmuch the consumers evaluate the CSR as very important (Sen and Bhattacharya, 2001; Klein and Dawar, 2004). In the present research we obtain the same results concerning the attribution of blame, brand evaluation and buying intentions. Indeed, consumers in the high importance CSR group are affected by CSR performance, while consumers in the low importance CSR group don't seem to change their reactions because of CSR. An alternative explanation could be that consumers in the high importance CSR group identify themselves with the company (Sen and Bhattacharya, 2001). Moreover, a possible explanation could be found in consumers emotions' literature and Lazarus (1991) appraisal theory which suggests that goal congruence leads to positive emotions; goal incongruence to negative ones. Thus, the positive consumer emotions could have an impact on the attribution of blame, brand evaluation and buying intentions.Finally, in this study, we investigate the influence of CSR in a food industry's product-harm crisis. The importance of this investigation is emphasized in the ongoing discussions on the strategic role of CSR. However, it should not be disregarded that CSR is first of a societal obligation for the companies and the strategic role should not elude but enforce the ethical reasoning for CSR initiatives. Furthermore, the above mentioned argument seems more important nowadays since the positive impact of CSR on business performance is more difficult to establish because of the increasing CSR scepticism among all the stakeholders.\n7. Limitations and future researches: In the real world situation consumers tend to evaluate a company's CSR initiatives over time and recall this evaluation when and if a crisis takes place. Hence, one research limitation is the simultaneous act of reading about a company's CSR actions and judgement of the defect product and crisis situation instantly. Another limitation is the implementation of convenience sampling, allowing only narrow generalization of the results. Although margarine was purposefully chosen, results may not be generalized to other food products. Our results may be different in other crisis situations, other food product categories or other corporate associations, like corporate ability, or other cultures. For that reason, this initial effort should be repeated on the basis of these areas.\n8. Implications for practitioners: The results of this research allow us to make a series of managerial suggestions. Marketers should use CSR associations in order to build food brands, not only because CSR has an impact on brand attitude and buying intention in everyday routine situations but because it is a management tool in unexpected situations, such as product-harm crises, as well. Nevertheless, marketing managers should not exclusively focus on CSR association but on other corporate or brand associations as well, such as corporate ability and product characteristics, since they also have an impact on consumers' reactions after product-harm crises.Therefore, the contribution of our research to the discussion around social responsibility, states the CSR role in product-harm crises in the food industry. However, it goes without saying that managers should always consider the CSR limits. First of all, CSR should not be considered as an obligatory first-aid tool kit, being always there when managers need it. CSR can give benefits to companies, when used strategically and proactively, but in case of product-harm crises, CSR isn't a panacea. Managers should work hard to avoid the evaluation of their company as a socially irresponsible one, while at the same time should focus on CSR initiatives that are in accordance with the brand positioning and the real abilities of the food product and the brand.Likewise, companies and product managers in the food industry should remember that CSR has some drawbacks as well. First of all, we should always bear in mind that the good reputation of a company does not always protect the company from the crisis' negative impacts, in the presence of reliable and unambiguous proof of its responsibility. Furthermore, some authors suggest that familiarity, CSR, or good reputation could also be a disadvantage for the companies, because the Media and NGOs have their eyes always on them (Rhee and Haunschild, 2006; Park, 2008). Thus, big companies are more vulnerable to crises not only because it is more difficult to manage them (Cancel et al., 1999) but also because they attract the stakeholders' attention as well.\n",
"role": "user"
},
{
"content": "Purpose: - The first purpose of this study is to analyse the impact of prior CSR information on the perceived degree of danger, attribution of blame, brand evaluation and buying intentions after a product-harm crisis in the food industry. The second purpose is to examine the moderation effect of CSR importance ascribed by the consumers on the above mentioned relations.\nDesign/methodology/approach: - An experimental design consisting of three between-subjects conditions was applied and three CSR initiative conditions were selected (positive, negative CSR and no CSR information as a control condition). In this framework, three different scenarios were designed and tested under the condition of a product-harm crisis related to margarine.\nFindings: - This study highlights that CSR has an impact on attribution of blame, brand evaluation and buying intention but not on the perceived degree of danger. CSR importance has a moderation effect on the relationship between CSR and blame attribution, brand evaluation and buying intention.\nPractical implications: - Companies in the food industry should generate CSR strategies and should develop favourable CSR history not only because CSR has an impact on brand evaluation and buying intention in routine situations but because it is a part of crisis management and response strategy as well.\nOriginality/value: - There is lack of research directly emphasizing the role of CSR in product-harm crises, in the food industry. Besides, the assessment of CSR as an antecedent assurance factor in crisis situations has significant meaning due to the high vulnerability of food industry.\n",
"role": "assistant"
}
] |
Mindfulness and product failure management: an engineering epistemology
|
[
"Problem solving",
"Organizational learning",
"Product management",
"Team working",
"Ba",
"Mindfulness",
"Engineering epistemology",
"Engineering problem solving",
"Sensemaking"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Organizational learning has been heralded as essential for business success (Ellinger et al., 2002) - especially learning from failure (Cannon and Edmondson, 2001). Long recognized as a pre-condition for continuous improvement that can lead to bottom line enhancements (Levitt and March, 1988), organizational learning is often associated with higher levels of innovativeness (Hurley and Hult, 1998; Nonaka and Takeuchi, 1995) and superior firm performance (Huber, 1991; Levitt and March, 1988; Ruigrok and Wagner, 2003; Zahra et al., 2000). The literature is lean, however, with respect to organizational learning and knowledge creation in engineering contexts. Engineers solve technical problems every day using prescriptive approaches (e.g. Six Sigma), but how the solutions they generate become part of sustained organizational learning is underexplored in management and engineering literature. Motivated to address this gap, we designed a qualitative study based on semi-structured interviews with US product engineers charged with investigating product-related problems. Our goal was to generate a grounded theory about the way engineering problem-solving results in organizational learning and knowledge creation.Lloria (2007) argued, and we concur, that while organizational learning and knowledge creation have often been addressed as separate research streams, the latter may be a direct outcome of organizational learning. In this study, we define organizational learning and knowledge creation as having taken place when problem-solving outcomes changed organizational norms and assumptions. We designed an interview protocol that solicited rich descriptions of both successful and unsuccessful problem-resolution experiences to better understand the product problem-solving process and to identify conditions that promote or inhibit learning. Our findings, the result of iterative stages of rigorous coding, reveal that product engineers' real-world practices mirror Nonaka and Takeuchi's (1995) five-phase knowledge-creation model and the three stages of sensemaking in Weick's enactment theory (Weick et al., 2005). This discovery led us to dissect engineering problem-solving from an epistemological perspective, leading to the identification of \"engineering epistemology,\" an interpretative mindset grounded in intellect and practical know-how that individual engineers bring to situational contexts. Often neglected in practice, this highly situated cognition is a critical knowledge asset that gives rise to superior problem-solving performance.\n2. Theoretical foundation: Engineering knowledge creation may be viewed as a subset of general organizational learning that is facilitated in certain contextual settings. Huber (1991, p. 89) proposes that \"an organization learns if any of its units acquire knowledge that it recognizes as potentially useful to the organization,\" and Dodgson (1993, p. 377) defines a learning organization as one that \"purposefully constructs structures and strategies so as to enhance and maximize organizational learning.\" Accordingly, we conceptualized engineering knowledge creation stemming from product-related problems as a process that is instigated by product failure and supported by a failure management system, culminating in learning from failure in a mindful learning environment.2.1 Mindfulness, organizational learning, and knowledge creation\n3. Methods: We conducted qualitative research based on phenomenological interviews with product engineers in the automotive industry to generate a grounded theory about organizational learning and knowledge creation, relying on a research protocol recommended by Corbin and Strauss (1990). We digitally recorded face-to-face interviews and transcribed them. The rigorous analysis of these narratives detailing the \"lived lives\" of problem solvers yielded 31 successful and 29 less successful problem-solving stories to contrast and compare.3.1 Methodological approach\n4. Findings: Our data demonstrate stark differences between successful and less successful product-related problem-solving efforts by engineers. Analysis points to:* External conditions that influence the success level of problem-solving.* Product engineers as being not only technical investigators but also managers of relationships.* The extent to which problem-solving results in positive outcomes.We will first discuss these findings individually in this section then later show their relationships in the Discussion section.4.1 Five environmental conditions\n5. Discussion: Our study is, to our knowledge, the first research to model the dynamics of engineering knowledge creation in a product problem investigation context. Our model, building on the work of two great theorists - Nonaka and Weick - depicts an engineering team as an epistemic group embarked on a unique learning journey. The synthesis of Nonaka's knowledge-creation model and Weick's enactment theory and the identification of an engineering epistemology triad, often neglected in practice, are unique contributions.Our findings illuminated five external conditions - clear leadership vision, autonomy, controlled urgency, resources, and framework for sharing - along with five sequential phases through which rigorous and disciplined problem-solving efforts progress, culminating in system changes and successful sensemaking with cognitive convergence of stakeholders. These discoveries prompted us to delve further into two concepts explored earlier: Nonaka's (1994) knowledge creation process and Weick's notion of mindfulness (Weick and Sutcliffe, 2007). Additional insight about our findings was derived from Nonaka's model and Weick's enactment perspective of sensemaking (Weick et al., 2005), the genesis of Weick's idea of mindful organizing.5.1 Five external conditions\n6. Limitations: Several limitations to this study should be noted. Our sample was small and non-random. Most respondents were employed by, and worked at, the home office facilities of US Midwest firms, half of which were associated with the automotive industry. We acknowledge that regional and industry commonality may have predisposed the sample to location-specific and industry-prevalent attitudes. Although clear common themes emerged from our data, future research might seek to confirm our findings among a larger sample of engineers. Second, we asked engineer respondents to recall past experiences and narrate them in detail. D'Argembeau and van der Linden (2004) argue that both temporal distance and valence (positive versus negative) of events influence sensorial details and clarity of contextual information -, i.e. recent events and positive experiences tend to be more vividly recalled. We asked respondents to consider events occurring in the past 24 months, but some were motivated to narrate more distant experiences. Consequently, we acknowledge the potential effect of time on memory in the stories that resulted and recognize that \"successful\" stories may have been remembered with greater intensity and clarity than \"less successful\" stories.Finally, as is typical of a study like this, the categorization and interpretation of data depended on the perspectives and knowledge of the researchers. A research process is not entirely free of researcher bias, and so it is important for researchers \"to be continuously aware of the possibility that [they] are being influenced by pre-existing conceptualizations of [their own] subject area\" (Suddaby, 2006, p. 635). The principal researcher in this study is an experienced reliability engineer working closely with product design engineers in this capacity. While every effort was made to stay attentive to preconceptions in exercising interpretive judgment, we acknowledge the potential effect arising from our past and present experience.\n7. Implications for practice and further research: This study was conducted to inform practitioners and researchers of the dynamics of the engineering knowledge-creation process in product-related problem resolution. For practice, first, our findings demonstrate engineering problem-solving as being as much a social process as a technical exercise (Cannon and Edmondson, 2001; MacDuffie, 1997; Smith, 1998; Suchman et al., 1999). The ability of lead engineers to bring together varying view points of all stakeholders is a key to effective problem-solving and projects an image that is in stark contrast to the traditional notion of engineer as a pragmatic pursuer of people-free solutions (Schein, 1996). It is in the firm's best interest to provide an environment that allows the engineers to \"engineer\" relationships among various knowledge holders to cultivate mutual trust and commitment. At the same time, facilitation skills - in addition to technical competencies - may be an important attribute for employers to look for in engineering candidates. Second, social interactions may be far more important than historically viewed by management science. Sociality is an important part of engineering processes as it brings out tacit knowledge embedded in individuals and local contexts (Nonaka, 1994; Nonaka and Takeuchi, 1995; Orr, 1998). Amalgamation of tacit knowledge, which is unique to each firm, has far wider implications than just \"solving problems.\" The firm's core competitive advantage is built on the unique attributes of tacit knowledge assets (Edmondson et al., 2003; Nonaka et al., 2000, 2006), engineering epistemology being one of them. Socialization as a mechanism for tacit knowledge transfer and nurturing of group epistemology, going beyond superficial \"team-building\" exercises, merits management attention. Finally, this study also underscores the importance of leadership in motivating effective problem resolution and higher organizational learning. Our investigation shows that problem-solving can yield a great deal of useful knowledge for the firm, which is too valuable to let \"drift and evolve at the periphery of management's attention\" (von Krogh et al., 2001, p. 435).From a research perspective, this study contributes to the literature on organizational learning. We offer empirical evidence of double-loop learning following single-loop learning - rather than a dichotomy between the two (Argyris, 1976) - thus adding to \"more systematic empirical studies [that] will not find the two types of learning to be distinct in practice\" (Huber, 1991, p. 93). Further, our study contributes additional empirical evidence of the social nature of knowledge - that \"interpretations in organizations [are] socially constructed\" (Huber, 1991, p. 103), and so is organizational knowledge (Baba et al., 2004, p. 550). Our study also responds to a call for enriching the organizational learning research landscape by providing more \"cumulative and integrative work\" that \"cross-fertilizes and synthesizes the results from previous research\" (Huber, 1991, pp. 107-8).Additionally, more research on ba-an empirically under-explored construct (Nonaka et al., 2006) - should provide further understanding of the dynamics of engineering problem-solving and knowledge creation. Finally, even though the study was conducted in a particular engineering setting, our findings may carry across other engineering contexts such as new product development and innovation, as well as non-engineering problem-solving environments.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to provide empirical evidence of engineering knowledge creation in the context of product failure management, thereby extending knowledge about organizational learning and mindfulness to a largely unexplored context. The study addresses a gap in the literature by illustrating \"engineering epistemology\" as a critical knowledge asset that gives rise to superior problem solving - and potentially - superior business performance.\nDesign/methodology/approach: - The authors conducted qualitative research based on phenomenological interviews with product engineers to generate a grounded theory about organizational knowledge creation. Rigorous analysis of narratives detailing the \"lived lives\" of problem solvers relied on a research protocol recommended by Corbin and Strauss.\nFindings: - The findings show that engineers' real-world problem-solving practices mirror Nonaka and Takeuchi's five phases of knowledge creation and the three stages of sensemaking in enactment theory, the genesis of Weick's notion of mindfulness. A synthesized model illustrates how a five-step problem-solving process facilitated by environmental conditions resulting in organizational learning is influenced by an \"engineering epistemology\".\nResearch limitations/implications: - The sample was limited to engineers based primarily in the US Midwest. While the authors' methodology (grounded theory) was appropriate for theory generation, the results invite quantitative testing involving a larger and more diversified sample of engineers.\nPractical implications: - The paper highlights the social aspects of engineering problem solving that firms can optimize for effective problem investigation and higher organizational learning.\nOriginality/value: - The paper conceptualizes problem-solving teamwork as epistemic collaboration, with the often un-optimized potential of generating organizational learning. It is, to the authors' knowledge, the first research to concentrate on modeling the dynamics of knowledge creation in an engineering problem-solving context.\n",
"role": "assistant"
}
] |
Quality of life in breast cancer sufferers
|
[
"Clinical effectiveness",
"Health and safety",
"Quality of life",
"Breast cancer"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction and background: The World Health Organization defines health as complete mental, physical and social well-being, not only disease absence, so health measurement must indicate changes in disease severity and estimate well-being, which can be achieved by measuring quality of life (QOL) (Testa and Simonson, 1996, p. 1). QOL is defined as \"the individual's perception of his/her position in life, within the cultural context and the values in which he/she lives, as well as relation to his/her objectives, expectations, standards and concerns,\" which includes physical, mental and social well-being (World Health Organization Quality of Life Group, 1998, p. 1). There are many factors affecting QOL, one is health status, which refers to the QOL component determined primarily by health status and focusses on the physical, psychological and social domains (Juniper, 2001). Most clinicians agree that traditional medical outcomes such as mortality and morbidity must be complemented by health-related QOL (HRQOL) assessment to understand the disease process and care. In some cases, well-being may increase even if physical fitness or health deteriorates (Eronen et al., 1997).\nMethodology: Our cross-sectional study measured QOL among early stage female BC patients. Our purpose was to determine the main factors affecting their QOL. Sample size, using PASS11, was calculated at 64 patients (Hintze, 2011), which produces a two-sided 95 percent confidence interval with a distance from the mean equal to five when estimated standard deviation is 20 (Al-Naggar et al., 2011). The oncology outpatient clinic was open three days per week during the study (April-June). Two days were selected randomly and all early stage BC female patients (average three cases per day), taken from all attending patients, agreed to be involved. Three interviewer administered questionnaires were used to collect data. Follow-up telephone calls were used to complete missing data in the questionnaires.\nResults: The study sample included 64 subjects. Respondent's age ranged from 28 to 68 years (mean =51.05+-9.25years). In total, percent were educated; 3 percent were single, 69 percent married, 27 percent were widowed and 2 percent divorced; 95 percent had children ranging from 1 to 42 years (mean, 20.43+-9.51 years); 9 percent lived in a rural area; most (82.8 percent) were housewives. In total, 59 percent in the full sample reported insufficient family income and had to borrow money for treatment. All patients had public health insurance. In total, 80 percent had a BC family history. The physical domain score was most (2.22+-1.10) and the functional domain was least affected (3.18+-0.93). There was a positive statistically significant correlation between patient age and total QOL score, emotional, functional and BC-specific domains. There was a statistical difference between illiterate and educated patients regarding QOL scores except in the BC-specific domain. Married patients had lower QOL scores but this was statistically insignificant. The QOL scores were higher among all those who had children and this association was statistically significant in different QOL domains except in the BC-specific domain. There was positive correlation between total children and QOL scores, but this correlation was statistically insignificant. There was a statistically significant positive correlation between youngest child age and QOL scores except in the physical (p=0.252) and functional domains (p=0.066). The QOL score was higher among those living in urban areas except in the functional domain, but all associations were statistically insignificant. The QOL scores were higher among working patients in all QOL domains except in the social, but this difference was statistically insignificant. Patients reporting sufficient family income had higher QOL scores in all domains and associations were statistically significant. In total, 67 percent did not know about early BC detection methods and none who had this information did breast self-examination or attended for mammography. In total, 52 percent had another chronic disease and 55 percent had a family carer (Table I).\nDiscussion: BC is the most frequent cancer among women (25 percent of all female cancers) and the second most common cancer among both sexes (11.9 percent) after lung cancer. It ranks as the fifth cause of death from cancers overall (522,000 deaths), but it is still the most frequent cancer death in women worldwide (14.7 percent of all cancer deaths). In Egypt, according to the NCI, BC is the most common malignancy among Egyptian females (33.5 percent) and it is also the most common cause of cancer deaths among women (21.6 percent) (Globocan, 2012).\nConclusion: The physical domain was most affected among BC patients in our study and the functional domain least. Socio-demographic factors that significantly affected QOL scores for BC patients were age, education, children and income. Carer presence was the specific patient characteristic that affected QOL. Disease-related factors affect QOL scores for early stage BC include age at diagnosis, whether the predominant hand was affected, post-operative chemotherapy and difficulty obtaining medication, which had statistically significant effects on QOL scores.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to measure health-related quality of life (QOL) among patients with early stage cancer breast under curative treatment at department of oncology and nuclear medicine at Ain Shams University Hospitals. Identify factors affecting QOL among these patients.\nDesign/methodology/approach: A cross-sectional study measured QOL among early stage female breast cancer (BC) patients and determined the main factors affecting their QOL. Three interviewer administered questionnaires were used.\nFindings: The physical domain mostly affected in BC patients and the functional domain least. Socio-demographic factors that significantly affected BC patients QOL scores were patient age, education, having children and family income. Specific patient characteristics include caregiver presence - a factor that affected different QOL scores. Age at diagnosis, affection in the side of the predominant hand, post-operative chemotherapy and difficulty in obtaining the medication were the disease-related factors that affected QOL scores.\nOriginality/value: The final model predicting QOL for early stage female BC patients included age, education and difficulty in obtaining the medication as determinants for total QOL score. Carer presence was the specific patient characteristic that affected different QOL scores.\n",
"role": "assistant"
}
] |
Using the kaleidoscope career model to analyze female CEOs' experiences in sport organizations
|
[
"Women",
"Careers",
"Sport"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Women's participation in the workforce has increased over the past 20 years; however, most organizations are still male dominated (Burke and Major, 2014). One field in which this disparity is particularly clear is that of sport. Despite increases in women's participation in sport at all levels over the past 30 years, the lack of female representation at the higher levels of sport administration has been well documented (Burton, 2015; Knoppers and Anthonissen, 2008). Sport organizations' structures and practices routinely undermine women's progression (Hovden, 2000), and simply often do not take women seriously at the highest leadership level (Shaw and Hoeber, 2003). Despite long-term calls for changes to sport organizations (Hums et al., 2007; Women's Sports Foundation, 1994), little has changed for women's representation in high levels of sport administration over the past 20 years. Similarly, much of the research focus on women's experiences in sport organizations has focused on barriers and constraints to their progression (Adriaanse and Schofield, 2014).\nTheoretical framework: There is a need to focus on women's positive achievements which may provide a counter-balance to the constraints focus of prior research (Burton et al., 2011; Claringbould and Knoppers, 2012; Shaw and Hoeber, 2003). To achieve this, we need to focus on successful experiences but, as Knoppers and Anthonissen (2008, p. 94) have argued, \"little is known about how senior managers or executive directors of (sport) organizations construct [...] their work\". Much of the career literature focuses on \"structures, routes and processes through which employees gain promotion\" (Harrow and Mole, 2005, p. 80). This structured approach to understanding careers does not provide a complex enough framework to allow for the flexible, transient nature of careers (Baruch et al., 2014) and, in particular, many women's experiences (Eagly and Carli, 2007). Consequently, we turned to Harrow and Mole (2005) who have explored CEO's career accounts. Rather than focusing on identifying clearly defined pathways through careers, this approach recognizes recent (1980s and onwards) changes to career structures in which constantly evolving economic and social circumstances have led to the development of non-traditional careers, the reduction in contract certainty and length and the greater integration of women in the workforce (Halford et al., 1997).\nMethodology and methods: To address our research questions, we interviewed female CEOs of New Zealand National Sport Organizations (NSOs). At the time of the research, there were ten such CEOs identified through an Internet search and the researchers' personal networks. All ten were invited to take part in a face-to-face interview and asked questions regarding their educational experience and career, particularly their experiences in sport management. Seven agreed to participate and Table I provides an overview of their backgrounds. This information is limited by our ethical approval and promise to the participants to respect their confidentiality.\nFindings and discussion: The elements of the KCM are presented here as discrete elements. However, where possible, we show how the sub-themes allowed us to indicate overlap between them.\nConclusions: We analyzed the career accounts of female CEOs in NZ sport organizations. The KCM was developed to examine the career experiences of women. Following Sullivan and Baruch (2009), we challenged the KCM by using qualitative inquiry. In so doing, we explored the usefulness of the KCM in understanding women CEOs' experiences and explored women's experiences by using KCM. First, we demonstrated that authenticity has many nuances and was experienced in many different ways. This reflects the range of individual backgrounds and work contexts. As women understand what drives them, so organizational decision-makers need to recognize the multiple ways in which women identify authenticity, rather than focus on conventional understandings of a good employee.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this research is to examine the experiences of female CEOs in New Zealand sport using a career account approach overlaid with the Kaleidoscope Career Model (KCM). This research focuses on their successful careers, rather than constraints and barriers, which is a feature of much previous research in the area.\nDesign/methodology/approach: Semi-structured interviews and analysis were informed by the principles of the KCM of career authenticity, balance and challenge. Seven chief executive officers in New Zealand sport organizations were interviewed. Data analysis was undertaken through a three-step coding process.\nFindings: Respondents highlighted the importance of authenticity, balance and challenge in different ways. The KCM model is extended by identifying sub-themes within each principle. These were, under authenticity, passion and relationship-building; under balance, self-awareness and influencing the organization; and under challenge, taking opportunities and working in sport's gendered environment. Findings indicate that a more nuanced development of KCM is required and that decision-makers in sport organizations need to be aware of the varied drivers of women's positive experiences.\nResearch limitations/implications: This research contributes to an understanding of successful women's careers, rather than focusing on barriers to success. It also extends the KCM. Future research is necessary to examine the viability of our suggested extensions to the KCM.\nPractical implications: While not generalizable, this study's findings suggest that sport organization decision-makers would be well advised to understand the unique ways in which authenticity, balance and challenge are understood by their female employees. This may encourage organizations to look more closely at their cultures to ensure that they are more welcoming and supportive to women in a male-dominated industry.\nOriginality/value: This research contributes to an understanding of successful women's careers, rather than focusing on barriers to success. It also extends the KCM. Future research is necessary to examine the viability of our suggested extensions to the KCM.\n",
"role": "assistant"
}
] |
Hold me responsible: The role of corporate social responsibility and corporate reputation for client-perceived value
|
[
"Corporate social responsibility",
"Corporate reputation",
"Strategic orientation",
"Client-perceived value"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Today, companies are expected to do business in a responsible and good way. Due to these expectations, corporate social responsibility (CSR) is becoming an imperative in strategizing, planning, and tactics. CSR encompasses customer perceptions of environmentally and socially responsible companies. However, there is no definite formula as to how to implement CSR and communicate its manifestations to stakeholders (Nielsen and Thomsen, 2007). Researchers are debating how to communicate CSR (Perez, 2015), and how to distinguish between CSR activities and pure public relations (Frankental, 2001); however, parallel to the corporate side there is still a lack of knowledge regarding how stakeholders perceive such actions, whether and why CSR has an effect, and what kind of feedback/outcomes companies may expect from stakeholders when pursuing CSR actions.\n2. Literature review and conceptual framework: With the conceptualization of CSR, De Bakker et al. (2005) report a tendency for the continual proposal of new constructs and new linkages. We take the standpoint of corporate communications and use a specific definition pertaining directly to customers. Walsh and Beatty (2007, p. 133) posit that CSR represents \"customers' perceptions that the company sees and acts on environmental and social responsibilities.\" Sen and Bhattacharya (2001), in their seminal work, address company-specific (product quality) and individual-specific factors (customer general beliefs) of when, how, and for whom CSR initiatives work. Hence, we assess the perceived CSR as customers' impressions based on information they gather from communications and relationships with providers. When providers are acting responsibly, their acts join the general pool of attributes assessed by service clients when evaluating the service in all phases of the process (Schreck, 2011). Stakeholders' attitudes toward companies' CSR activities are critical for maximizing business benefits from CSR activities, pointing to a need for companies to effectively communicate CSR to stakeholders (Du et al., 2010; Golob and Bartlett, 2007).\n3. Methodology: In order to empirically test our conceptual framework, we selected business clients and professional service providers' relationship. Due to the high level of expertise required to implement a professional service, business clients usually lack knowledge and need to rely on other cues in order to form their perceptions. Corporate reputation and CSR are thus likely to play an important role in the perceived value and, more precisely, the functional value.\n4. Results and discussion: Descriptive statistics, diagnostic tests and statistical inference for items for all constructs were first conducted using SPSS V.20. Covariance-based structural equation modeling using LISREL 8.71 was then used for testing the direct hypothesized effects, following the two-step approach (Anderson and Gerbing, 1988). Finally, Hayes' (2013) approach was used to estimate the full conceptual model.\n5. Discussion and conclusions: Three important findings are revealed in this paper. First, communication practices that concern CSR and corporate reputation positively and significantly influence client-perceived (functional) value. That is, corporate communication framework elements are important signals of quality in business services. This again stresses the vital importance of communicating issues related to CSR and actions for building corporate reputation vis-a-vis clients. Well-communicated CSR activities generate favorable customer attitudes and support behavior, as well as strengthen relationships with customers.\n",
"role": "user"
},
{
"content": "Purpose: Corporate communication practices are becoming ever more important for business service clients, as they signal quality and hence are related to client-perceived value. The purpose of this paper is to examine the interplay between corporate social responsibility (CSR), corporate reputation and client-perceived value, and to assess the moderating role of strategic orientation in business service relationships.\nDesign/methodology/approach: The conceptual framework based on the corporate communication framework, signaling theory and relationship marketing theory has been tested on a survey sample of 228 client firms, using covariance-based SEM and additional procedures for assessment of mediation and moderated mediation.\nFindings: This paper reveals that communication practices concerning CSR positively and significantly influence client-perceived value. The authors show that reputation fully mediates the effect of CSR on client-perceived value. Finally, the effect of CSR on value is stronger if the client firm has a short-term strategic orientation, while long-term strategic orientation boosts the effect of corporate reputation on customer-perceived value.\nResearch limitations/implications: Further research on the topic may involve developing links between other elements of the corporate communication framework and client-perceived value.\nOriginality/value: The originality of the paper lies in better understanding the effects of CSR and corporate reputation on client-perceived value. The authors provide empirical evidence of the mediating role of reputation between the CSR (seen as \"actions\") and client-perceived value.\n",
"role": "assistant"
}
] |
Who needs delight? The greater impact of value, trust and satisfaction in utilitarian, frequent-use retail
|
[
"Trust",
"Value",
"Customer satisfaction",
"Customer loyalty",
"Delight",
"Hygiene factors"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In the first decade of the twenty-first century, Portugal and other countries witnessed a proliferation of supermarkets, most of them incorporated in shopping malls (APED, 2011). As a result, consumers have been granted access to an increasingly wide range of products and brands, and competition among major retailers has intensified. In such a competitive market consumers' store selection is of significant importance (Furey et al., 2002).\nConceptual development: Because the majority of the paths in this model have already been empirically proven and the focus is on the strength of the paths when all constructs are included, the following sections will provide only a brief discussion of constructs. Greater emphasis is placed on justifying the expectations of their relative strengths.\nMethodology: Sample and data collection\nResults: The PLS model is analyzed and interpreted in two stages. First, suitability of the measurements is assessed by evaluating the reliability of the individual measures and the discriminant validity of the constructs (Hulland, 1999). Then the structural model is appraised. Item reliability is assessed by examining the loading of the measures on their corresponding construct. Items with loadings of 0.707 or more should be accepted, which indicates that more than 50 per cent of the variance in the observed variable is explained by the construct (Carmines and Zeller, 1979). In this study, all items (Table II) display loadings equal to or above 0.748 and were therefore accepted. Composite reliability was used to analyse the reliability of the constructs as this has been considered a more accurate measurement than Cronbach's a (Fornell and Larcker, 1981; Sanchez-Franco and Roldan, 2005). Table II indicates that all constructs demonstrate acceptable composite reliability with values over 0.7 (Nunnally, 1978).\nDiscussion: Exploration of the relative importance of delight in the supermarket setting is extremely relevant because of the high costs typically associated with delighting retail customers. Among the reasons customers often provide for their delight are the availability of easy parking, wider store aisles, highly trained salespeople who are empowered to make decisions and equipped with problem-solving capabilities, the absence of stock outs, and extreme markdowns (Arnold et al., 2005). Each of these sources of delight requires heavy financial investments by the retailer as well as time and spatial resources. As the results of this study demonstrate, those may be unnecessary expenses in retail settings that customers visit frequently.\nConclusions: The main goal of this study was to determine whether the cumulative effect of satisfaction, perceived value and trust provide more predictive power of consumers' loyalty intentions than does the construct of customer delight in a routinely-patronized, utilitarian service setting. The findings reveal that not only does the cumulative effect of these variables have a greater impact on loyalty intentions than customer delight, but also that delight has no significant impact on loyalty intentions in a utilitarian service setting that encourages frequent patronage, such as that of a supermarket.\n",
"role": "user"
},
{
"content": "Purpose: The aim of this study is to determine whether the cumulative effects of satisfaction, trust, and perceived value may, under certain conditions, provide more explanatory power for customer loyalty intentions than the often studied and more elusive customer delight. Herzberg's two-factor theory is used to explain why the frequent nature of grocery shopping, a primarily utilitarian experience, might introduce considerations that have not yet been addressed in the study of delight.\nDesign/methodology/approach: A survey is administered to a quota sample of Portuguese supermarket shoppers via phone, using a CATI system.\nFindings: Research findings suggest that perceived value, trust, and satisfaction have a greater impact on behavioural outcomes than customer delight in the grocery shopping setting. In such a setting, cognitive drivers may be even more important for customers who are primarily concerned with hygiene factors (rather than motivators).\nResearch limitations/implications: Retailers are encouraged to focus on the more mundane factors that influence consumers' perceptions of value and trust rather than trying to invest in the substantial resources required to continually delight consumers. Future research may explore other determinants of loyalty intentions and test the extended model in different service sectors, cultural contexts and countries.\nOriginality/value: This study applies Oliver et al.'s consumer delight model in a utilitarian, frequent-use setting, finding previously undiscovered limitations to its validity.\n",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.