title
stringlengths 6
301
| keywords
listlengths 0
23
| messages
listlengths 2
2
|
|---|---|---|
From artefact to effect: the organising effects of artefacts on teams
|
[
"Team working",
"Culture (sociology)",
"Artefacts",
"Injuries",
"Sweden"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The purpose of this article is to describe the organising effects of artefacts on teams. It has long been argued that common social understanding can influence organisations (Weick, 1979). Culture is said to be the glue that makes up a common identity between different individuals (Smircich, 1983). This means that there are good reasons for determining which cultural phenomena influence teams, and how it is done. It can occur, for example, through the use of symbols, metaphors, anecdotes, and artefacts. These can take on certain values and make people act in a different way than expected (Smircich and Morgan, 1982; Smircich, 1983; Wilkins and Ouchi, 1983). A trauma team with strong symbol values has been studied. Its artefacts have been identified and categorised, and the organisational effects have been analysed. The research question is: do artefacts contribute to team organisation in a culture of fragmented and differentiated healthcare?The healthcare team - an idealised phenomenon\nMethodology: Data for this study have been collected through semi-structured interviews and observations. A total of 30 key personnel from a trauma team were interviewed, including team coordinator, team leader, doctor, and registered and assistant nurses. They represent the entire personnel hierarchy of the team. The purpose was to understand the differences in the meaning of artefacts for different groups of the team. In addition, surgeons and anaesthetists were also interviewed. The purpose was to learn whether symbol value perceptions varied between colleagues from different areas. Only data from participants with more than three years of team work experience was used. This demarcation was chosen so that the information collected only came from participants well acquainted with the organisation (Kvale, 1996)The observations were done at two accident exercises and four 12-hour real emergency events. The purpose of following the trauma team during real emergency responses was to complement the statements of the participants with our own observations. This provided information that had not been revealed during the interviews. The interviews helped us to investigate whether the events we had registered were coincidental or if they were part of the normal behavioural pattern of the team (Spradley, 1979).The collected data were analysed in three stages. First, we selected and coded the relevant portions from the interviews (Constas, 1992; Miles and Huberman, 1994). Second, we documented and coded the accompanying observations (Kvale, 1996). Third, the interviews and the observations were analysed as a whole. By combining the two data collection methods, we have been able to provide a broad and varied database (Miles and Huberman, 1994).\nResults: The trauma team\nAnalysis: The artefacts that were registered in the trauma room are here divided up into three levels. According to Wartofsky's (1979) methodology, primary artefacts are made up of physical phenomena needed to complete a mission. Secondary artefacts are made up of action models. Tertiary artefacts are different assumptions about the team's position. The organising effects of these three levels of artefacts are interpreted in this study.At the primary level, artefacts were symbol-creating utensils, trauma equipment and lapel pins. The trauma room and its floor marking would be considered primary artefacts. Those crowding outside the lines were not part of the team. The coloured clothing, hats, visors, face masks, and gloves made the team members unrecognisable, but there was no doubt about their identity. They were the trauma team, independent of any divisions among the various professional groups or clinics. These artefacts demarcated members from non-members and contributed to their team affiliation. The primary artefacts of the team had a demarcating effect.The secondary level of artefacts consisted of action schedules, role distributions, and the way the patient was divided up among the team members. Even though the purpose of these artefacts was to favour integration and create order, competition could arise in tense situations. The physical space in the trauma room was claimed by individual members as well as by groups within the team. Certain areas were \"hot\", or desirable. Team members would block access to such popular areas in order to protect them from invasion. The patient was also object of competition. Different parts of the patient were assigned to certain team members and were then guarded and defended. This competition created steep vertical hierarchies with the trauma leader at the top. The secondary artefacts had a systemising effect on the team by distributing tasks, areas, and positions.The tertiary level of artefacts consisted of assumptions about the team's position. The team members' images of themselves were manifested by interpretations and hypothetical assumptions. They would compare themselves with trauma teams at nearby hospitals. They would mark their superior position by being abrupt or demeaning to others. When a trauma team at another hospital was perceived as more successful, the resident team was motivated to participate in intensive training in order to regain its role as the most successful team. They developed new techniques and polished existing ones. Observers in the trauma room and in the university hospital's general areas contributed to the team's unique position. The attitude of the assistant nurses, for example, signalled exclusivity. Outside the yellow floor lines were potential future team members that were still undergoing training. However, instead of feeling welcome, they were made aware of the team's elevated position. There was an unwritten requirement for perfection, and inability and uncertainty were covered up. The ideal self image did not allow for mistakes. Instead, the team members could act conspicuously and show exaggerated self confidence. This was expressed by special terminology, dominating actions, and \"wide walking\" while meeting with other hospital personnel. The tertiary artefacts had a positioning effect.The empirical analysis is summarised in Figure 4.The organising effects of artefacts\n",
"role": "user"
},
{
"content": "Purpose: - Earlier studies have identified artefacts, but have only to a lesser degree looked at their effects. The purpose of this paper is to investigate how artefacts contribute to organisation.\nDesign/methodology/approach: - A trauma team at a university hospital has been observed and its members interviewed.\nFindings: - The trauma team showed itself to be rich on artefacts since it had strong internal driving forces, high legitimacy, and tried to live up to high expectations from the outside. Its members were motivated to be in the forefront of trauma care. Through renewal, the team succeeded in maintaining demarcation. It also succeeded in systemising internal work tasks and made for itself a position in relation to the outside. The team's capacity, however, came to be limited by internal conflicts and battles for prestige.\nPractical implications: - The study shows that informal logic has a strong influence on teams. Teamwork contributed to the development of organisational structure and motivation for the personnel.\nOriginality/value: - Earlier studies advocate the important role of artefacts in order to communicate, collaborate, negotiate or coordinate activities. The conclusion is that artefacts also have an organising and developing effect on teams in a fragmented and differentiated healthcare.\n",
"role": "assistant"
}
] |
A supply chain study of technology trust and antecedents to technology internalization consequences
|
[
"Supply chain management",
"Automotive industry",
"United States of America",
"Technology led strategy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Trusting a technology, within an organizational setting, may place users at risk if the technology fails to meet operational expectations (Lippert, 2001). Trusting an inanimate object, on which a supply chain member is dependent for sharing information, produces uncertainty that orders may not be properly executed. The change to a new technological solution includes risk and the possibility of experiencing an information loss associated with the adoption or usage behavior. In some cases, these losses are not directly quantifiable which suggests that uncertainty, associated with the loss, is a significant part of risk (Yates and Stone, 1992). The introduction of technologies throughout supply chains poses great uncertainty and risk as multiple players, with diverse technological backgrounds and systems are requisite to fully integrate supply chain activities. Increased risk and uncertainty in supply chains can lead to significant inefficiencies and the inability to deliver products to final consumers (Zsidisin, 2003).Organizational strategies to overcome risk and uncertainty include inter-organizational information sharing via information technology (IT) (Chopra and Van Mieghem, 2000) as well as supply chain members' commitment to supply chain planning (Van Landeghem and Vanmaele, 2002). Organizations are relying more on ITs to share information and reduce risk (Chen and Soliman, 2002) as long as the information is both accurate and accessible (Hull, 2001). If employees understand the importance of supply chain management concepts and the (positive) long-term consequences (LTC) of using a new technology, they are more likely to adopt and internalize the new system (Forman and Lippert, 2004).Internalization is the effective and consistent use of a technology over time (Lippert and Forman, 2005b; Forman and Lippert, 2004). As such, understanding the factors impacting user internalization along supply chains can aid in the reduction of uncertainty and risk through managerial investment in activities designed to cultivate positive user attitudes and beliefs regarding the technology. As supply chains are becoming more dependent upon technology particularly to provide part-level visibility (Schoenthaler, 2003), the level of interest in the technological innovation continues to expand (Harreld, 2001). Many companies are reluctant to implement internet-based real-time data visibility products throughout their supply chains out of concern for trust barriers (Songini, 2000).The study's population consists of first-members of a major US automotive manufacturer's supply chain using a technology called the collaborative visibility network (CVN). This web-based technology was designed to streamline inbound parts operations by providing real-time part-level visibility to the supply chain members. CVN maintains a database of cubic dimensions for manufacturers' parts, components, assemblies, and subassemblies supplied to the automotive company's consignees. CVN also incorporates functions such as a customer management capability that enables suppliers, customers, and receiving locations to update parts and packaging information. Order management features allow supply chain members to view and track shipment details. A business intelligence mechanism includes updated interactive reporting and analysis capabilities to support operational and payment processes. These functions allowed the fourth-party logistics (4PLs) firm to better manage the inbound flow of product.Sharing information requires a significant amount of trust on the part of the supply chain participants since the human experience is predicated on various forms of trust (Lippert, 2001). Trust may develop between individuals (Mayer et al., 1995; Johnson-George and Swap, 1982), organizations (Gulati, 1995), individuals and organizations (Zaheer et al., 1998), individuals and social institutions (Barber, 1983), and between a user and an IT (Lippert, 2001). In each case, the object of trust differs in that trust is bestowed upon a person, place, event, object (Giffin, 1967) or technology. Building trust in technology is becoming more important (Jones et al., 2000) as individuals develop greater dependence upon technologies (Lippert, 2004). This is particularly true for supply chains that are becoming more global in nature requiring inter-organizational trust and trust in the technologies used to link international partners (Olson and Olson, 2000).Inter-organizational trust (Dyer and Chu, 2003), information sharing (Chopra and Van Mieghem, 2000), and perceived consequences of adopting a technology (Cheung et al., 2000) have been examined in the literature. Users' technology trust (TT), which should lead to a propensity to internalize an IT and the organizational antecedents to users' perceptions of LTC of using a technology have not been thoroughly explored. As such, the purpose of this study is to examine organizational antecedents of perceived LTC and TT in the technology internalization process which can ultimately affect supply chain member exposure to uncertainty and risk. Specifically, the current research aims to achieve two main objectives:1. to empirically investigate the influence of TT on perceived LTC which lead to technology internalization in the supply chain context; and2. to empirically examine the attitudinal antecedents leading up to TT and perceived LTC of adopting a new technology.The remainder of the paper is organized as follows. The next section provides the study's theoretical foundation. A discussion of risk, uncertainty, trust - particularly TT - and technology internalization in the supply chain context is offered. The subsequent section provides a series of hypothesized relationships between specific attitudinal antecedents of TT and the perceived LTC of technology internalization. The following section is the research methodology and the results. Managerial implications based on the results are addressed. Limitations and directions for future research are discussed.\nLiterature review: Social systems approach\nAttitudinal factors in technology internalization: Satisfaction with the existing system\nMethodology: A structural equation modeling (SEM) approach was used to evaluate the research questions. SEM is a statistical approach designed to test hypothesized relationships among observed and latent variables (Hoyle, 1995). The SEM represents a series of hypotheses identifying how the variables are generated and related (Chou and Bentler, 1995) and provides an assessment of predictive validity (Byrne, 1998). LISREL is appropriate for this type of data analysis since the model fit to the data is evaluated through the significance of the individual causal paths. Hypotheses are simultaneously assessed in the context of the entire model rather than through independent regression analyses of each distinctive hypothesis. The results generated from a SEM offer rigorous parameter estimations (Kaplan, 2000; Joreskog and Sorborm, 1996; Bollen, 1989), consider questions of both measurement and prediction (Kelloway, 1998), and are appropriate when latent constructs interact with one another (Bollen, 1989).Data collection\nData analysis: Data analysis was conducted in two stages. In stage one, reliability and construct validity of independent and dependent constructs were evaluated using Cronbach's a coefficients and a principle components factor analysis (PCFA). In stage two, a confirmatory factor analysis (CFA) was conducted prior to evaluating the hypotheses using SEM with the use of LISREL 8.5 (Joreskog and Sorborm, 1996).Construct reliability and validity\nDiscussion: Previous research found a strong relationship between perceived consequences of a technology and implementation of that system (Bergeron et al., 1995). This study identified and empirically validated relationships between antecedents to TT and the perceived LTC of using a technology, thus extending the findings of Bergeron et al. (1995). First-tier members of a major US automotive supply chain were examined and a SEM was derived from the literature. The data supported the proposed SEM. Three constructs, SES, TTF, and PSEs, positively and significantly impact TT, which in turn, positively influences LTC of using a technology. Both PSEs and TTF exhibit a direct positive effect on LTC. As expected, there was a negative relationship between SES and LTC suggesting that it would be more difficult to get individuals satisfied with existing technologies to favorably view the LTC of a new IT.\nManagerial implications: These findings have a number of managerial implications for supply chain managers within the specific supply chain investigated or other supply chains within the US automotive industry using this technology. These findings are important because supply chains are continually relying on technology to support their geocentric nature and overall increased structural complexity (Minner, 2003). The first of these findings identify and empirically validate specific root antecedents of perceptions of LTC of a new technology. These are SES, TTF, TT, and PSE. The study offers supply chain managers using the CVN technology a richer understanding of important antecedents to LTC of the new system.This knowledge combined with the empirically validated directional relationships between the above antecedents of LTC give supply chain managers responsible for CVN, a framework to guide and shape organizational perceptions of LTC, thereby making technology internalization more efficient. For example, to increase an adopting entities' positive perception of the LTC of using a technology, management have several options including implementing corporate initiatives to ensure that organizational users understand that a new technology will be a good fit with existing and future organizational tasks and responsibilities.Additionally, understanding the positive relationship between PSEs and LTC is valuable to know when management (e.g. purchasing personnel) contemplates upgrading technology to improve supply chain efficiencies and efficacies. For example, seeking out new technology similar to existing systems will assuage the uncertainty for adopting entities attempting to integrate a new radically different technology with their work activities.It is also important for management to understand the implications of these same antecedents on TT, especially since TT is a significant antecedent to an individual's perceptions of the LTC associated with using the IT. Trust is the glue that holds everything in society together (Luhmann, 1979) and is an important element of human relations (Larzelere and Huston, 1980). Trust is central to transactions (Dasgupta, 1988), because without trust, individuals and organizations are frequently immobilized through an inability to make a prediction or fulfill expectations. In organizations with high levels of trust, productivity consistently exceeds other businesses where trust is low or latent (Sitkin and Stickel, 1995; Davis et al., 1995).PSE and TTF can also facilitate TT, making the focus on these two antecedents critically important when internalizing a new technology. The directional differences among SES, LTC and TT are significant in this case. One way for management to address this is to try to build on users' SES and channel those positive attributes toward trusting technology as a genre as opposed to only a specific existing technology. By addressing these root antecedents, management can directly enhance the adopting entities' perceptions of LTC associated with using a technology, the LTC and indirectly through TT.\nLimitations: Several limitations have resulted despite efforts to guard against such occurrences. Most notably, two threats to validity are highlighted. The first is the threat to internal validity, commonly labeled as, response error due to social desirability. The second is a threat to external validity and involves the generalizability of the findings beyond the boundaries of the study.Social desirability, in the context of this study, consists of the desirable response by supply chain members who are implementing a new technology under influence from the original equipment manufacturer or the value-added reseller. In the situation where a response might have a negative effect on a business practice, respondents are likely to take minimal risk regardless of their individual beliefs. Therefore, the authors recognize the potential contamination of unidentified influences on the study respondents.Another significant threat (external validity) constitutes the generalizability of these findings to other supply chain technologies, alternative supply chain organizations, or applications to supply chains in other industries. A limitation in this study is directed at better understanding the phenomena of TT independent of the specific technology, organization, or industry. The implication for this threat is the replication of this study to other information system technologies or production organizations.\nFuture research: This study represents an important first step in examining the relationship between uncertainty reduction relating to incorporating technology in the supply chain context. The study, although significant in terms of its power and the industry from which it is drawn, focuses on one-tier of a US automotive supply chain. It would be useful for future research in this area to include multi-tier supply chain members. Additionally, it is useful to understand what causes individuals within supply chains to continue using a technology after it is introduced (Lippert and Forman, 2005a). This study is useful in the automotive industry but future research into other industries is needed to validate these findings thereby making them more generalizable. Finally, since many supply chains are global in nature, it would be constructive to conduct a similar study with less geocentric supply chains.\n",
"role": "user"
},
{
"content": "Purpose: - User uncertainty and risk are inherent in any technology adoption due to the perceived long-term consequences (LTC) associated with the internalization process. This study seeks to investigate antecedents of technology trust (TT) and perceived LTC associated with using a technology in order to understand the underlying attitudes and beliefs influencing supply chain members' trust in a specific information technology (IT).\nDesign/methodology/approach: - A theoretical model was tested through a survey of 449 supply chain members using an IT innovation for part-level visibility along the entire first-tier of a major US automotive supply chain.\nFindings: - Empirical results confirm the basic structure of the model, including the role of TT impacting users' perceptions of LTC relative to internalizing the technology. Results also identify several uncertainty reducing antecedents of TT and perceived LTC including satisfaction with the existing system, task-technology fit, and prior similar experiences.\nResearch limitations/implications: - The study, although significant in terms of its power and the industry from which it is drawn, focuses on one-tier of a US automotive supply chain. It would be useful for future research in this area to include multi-tier supply chain members.\nOriginality/value: - This study represents an important first step in examining the relationship between uncertainty reduction relating to incorporating technology in the supply chain context.\n",
"role": "assistant"
}
] |
The role of drug and therapeutics committees: Perceptions of chairs and information officers
|
[
"Committees",
"Drug controls",
"Health services",
"Patients",
"Sweden"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Costs, quality and safety of drug use\nStudy design: A qualitative method\nResults: There were 31 chairs and 43 information officers responding the research questions. The two groups were analysed separately. However, similar categories were found and the results are presented together. Conceptions in both groups were categorised according to how respondents viewed the role of DTCs in relation to the prescribers. We found four different conceptions among chairs and three among information officers (see Figure 1).The goals of the DTC expressed by chairs were \"optimal drug use\", \"good quality in drug use\", or both. These notions were grouped in the different categories (see below): \"optimal drug use\" in the two first categories (D and KD); both notions in the next (C); and primarily \"good quality in drug use\" in the last (SoL). This distribution could not be seen among information officers. On the contrary, 15 goal notions were evenly distributed over the three conceptions, and furthermore, many of the respondents had their focus on themselves instead and did not express any committee goal.Distant (D)\nWhether patients were included in the answers: Some chairs (20 per cent) and information officers (9 per cent) were patient-aware in their answers. However, this was not a major aspect. Among chairs there was one knowledge distributor, two catalysts and three supporters who mentioned patients. Of the information officers who mentioned patients, two were knowledge distributors, one a catalyst and one a SoL.Chair 6 wrote:The most important task is rational drug use that includes best possible drug use in all aspects, pharmacological, medical, economic, and not least, the best benefit for the patient, which is not always taken into consideration. ... the enormous impact the industry has had on prescribers for a long time and still has due to their tremendous engagement on all levels. Now also directed to the patients.Information Officer 6 wrote:Important that patients and the public have confidence in the drugs that are recommended in \"The list of recommended drugs\" and that these drugs are asked for by the patients.\nDiscussion: This seems to be the first study exploring the professional role of health care employers working with producer-independent information and education about medications, as perceived by the professionals themselves. We found a variation in how the DTC professionals relate to the prescribers: one conception was classified as \"distant\", whereas the other extreme was described as \"the link and supporter\". Furthermore, both chairs and information officers of the DTCs primarily have their focus on the prescribers and a few of them included the patients in their answers.Owing to the magnitude of drug use within health care and its relation to patient safety, public health and health economics, it is of considerable importance to develop strategies for public drug promotion. Studies have shown that traditional teaching and learning are unlikely to influence professional practice (Thomson O'Brien et al., 2003) and that personal outreach visits, especially combined with other methods, are a promising approach (Thomson O'Brien et al., 2001).Conceptions reveal how people perceive different phenomena. How a person relates to a phenomenon is crucial for the direction of this person's consciousness; moreover, the direction of the consciousness determines what a person will focus on, i.e. which aspects will be in the foreground and which will be in the background (Marton and Booth, 2000). Studies have shown connections between peoples' conceptions of their work and their competence in this work. For instance, in a study of physicians and nurses meeting patients with diabetes, the professionals' conception of diabetes care was related to the quality of how they performed patient counselling (Holmstrom et al., 2000; Holmstrom and Rosenqvist, 2001). In a study on medical students, the researcher first interviewed the students, focusing on their understanding of medical practice and then observed them during one or two patient counselling situations. Those students with a more complex understanding were less frustrated and were more able to listen to patients' concerns during the encounter (Dall'Alba, 2002). A study of anaesthetists did not explore conceptions but \"habits of action\": it showed that an interpretative action was superior to a reactive action when utilising tools, e.g. monitoring information during anaesthesia (Klemola and Norros, 2001). The researchers assume that ways of thinking about things and ways of doing things are intimately connected. This closely relates to our idea that conceptions are connected to action and competence. Another study on conceptions and competence within a field outside health care (developing engines for new models of cars) showed that engineers who had a more complex conception of their work were considered to have higher competence according to the engineers themselves (Sandberg, 2000).Conceptions are expressed in action and speech. Accordingly, a knowledge distributor probably does the planning, implementing and follow-up the work differently to a catalyst or a SoL. How this influences the results of the DTCs' work is beyond this study. However, this study reveals the existence of various conceptions, and the implications of these findings need to be discussed further.Conceptions are imbedded in our consciousness but are often so natural to ourselves that we do not notice them until something or somebody shows us that differences exist. When we can see that different conceptions exist, we can start to reflect on these differences and be aware of our own consciousness. In this way, a possibility of learning is created.Although legislation in Sweden (Swedish Commission, 1997, 1999) has strengthened patients' involvement and patients' rights within health care, only a minority of respondents seemed to be thinking about the patients when answering our questions - primarily, the prescribers were in focus. However, the prescribing doctor is not the end-user of drugs; rather, the patients are.We did not ask explicitly for the professionals' opinions of patient involvement in their work. Thus, we cannot conclude that the respondents lack \"patient awareness\". Still, by asking open-ended questions without focusing specifically on patients, we think we captured something that correlates to the degree of their patient involvement. If we had asked explicit of opinions of patient involvement we probably had got more patient-aware respondents. However, studies show that even when professionals say they intend to include patients, it is difficult to do this in practice (Rosenqvist, 2002; van Dam et al., 2003).Do chairs and information officers in DTCs need to include patients in their work? Are their activities influenced if they are \"patient aware\" or not? And if so, how? We do not know, but most consider that work within health care should be done for patients: thus, it may be advantageous to have \"patient awareness\" at all levels within health care.Methodological considerations\n",
"role": "user"
},
{
"content": "Purpose: - Improved quality and safety in drug use is a public health goal of major importance. In Sweden, local drug and therapeutics committees (DTCs) have adopted the task of working for safe and rational drug use. This study aimed to explore how chairs conceived the role of the DTCs, to explore how information officers conceived their own role, and to determine whether the respondents included patients in their answers.\nDesign/methodology/approach: - Data were collected using questionnaires and the answers were analysed according to phenomenographic method to identify conceptions. \"Patient awareness\" was studied by content analysis.\nFindings: - In both groups the prescribers were the focus of attention, and only a few respondents mentioned patients. A variation of four conceptions was found among chairs and three among information officers. It would be beneficial if DTCs used this knowledge in their development.\nOriginality/value: - The importance of \"patient awareness\" within DTCs must be further explored.\n",
"role": "assistant"
}
] |
Study on driver's braking intention identification based on functional near-infrared spectroscopy
|
[
"Machine learning",
"FNIRS",
"Cooperative driving",
"Driving intention identification"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Compared with the traditional driver assistance system, the cooperative driving system is required to be more intelligent and to have the capacity to identify drivers' driving intentions in real time. It can also adjust the condition of the vehicle to adapt to the driver so that it can achieve coordinated control and relieve the stress of the driver. Driver's intention identification is an extremely challenging problem in the field of cooperative driving and it now leads to a proliferation of studies. There are two methods to identify a driver's driving intention: one is based on driving operation data (Klingelschmitt et al., 2014; Jin et al., 2012; Dang et al., 2013) and the other on brain activity data (Ikenishi and Kamada, 2014; Ikenishi and Kamada, 2015).\n2. Methodology: The flow diagram of experimental data acquisition is shown in Figure 1. Firstly, we designed an experiment for the study of driver's braking intention, which was implemented on the DS. During the experiment, the DS could collect driving data in real time, such as speed, engine speed, acceleration, lateral acceleration, steering wheel angle, front wheel angle, brake pedal pressure and acceleration pedal pressure. The fNIRS device could record the oxy-hemoglobin concentration changes (DHbO) and deoxy-hemoglobin concentration changes (DHb) of the driver's cerebral cortex in real time. After that, we established the driver's braking intention recognition model through these experimental data.\n3. Results: Figure 10 shows the relationship between the number of sample features and the test accuracy of BPNN: the x-axis represents the number of sample features after PCA dimensionality reduction and the y-axis represents the test accuracy of BPNN. It can be seen that the samples have different feature quantities corresponding to the different test accuracy of BPNN. When the sample feature number was 15, the network achieved the best test accuracy of 80.39 per cent.\n4. Discussion: ADAS has raised expectations for the reduction of human error while driving and released or freed people from the task of driving. We called the pattern that ADAS sharing controlling with human driver and completing driving task together as cooperative driving. To achieve better cooperative driving, the driver's driving intention must be perceived by ADAS in advance. Driver's driving intention identification is of great significance to automatic driving, and identifying driver's driving intention with ADAS in advance is helpful for vehicle safety, stability and comfort. This study aimed to establish a driver's driving intention identification model. To this end, we designed an experiment based on DS and NIRx devices to study the driver's braking intention. The simulation experiment was carried out on the deceleration road (Figure 5). During the experiment, the DS collected driving-related data and the NIRx device collected cortical activity data. Then, the machine-learning algorithm was used to establish the identification model of driver's braking intention.\n5. Conclusion: In this study, the driver's braking intention identification model was established through PCA and BPNN. In addition, the data of three drivers driving under deceleration condition were gathered randomly to verify the model. The research results showed that the accuracy of the model established in this paper was 80.39 per cent. This study can be used as a reference for future research on driving intention through fNIRS, and it has a positive effect on the research of brain-controlled driving. At the same time, it has developed new frontiers for intention recognition of cooperative driving.\n",
"role": "user"
},
{
"content": "Purpose: Cooperative driving refers to a notion that intelligent system sharing controlling with human driver and completing driving task together. One of the key technologies is that the intelligent system can identify the driver's driving intention in real time to implement consistent driving decisions. The purpose of this study is to establish a driver intention prediction model.\nDesign/methodology/approach: The authors used the NIRx device to measure the cerebral cortex activities for identifying the driver's braking intention. The experiment was carried out in a virtual reality environment. During the experiment, the driving simulator recorded the driving data and the functional near-infrared spectroscopy (fNIRS) device recorded the changes in hemoglobin concentration in the cerebral cortex. After the experiment, the driver's braking intention identification model was established through the principal component analysis and back propagation neural network.\nFindings: The research results showed that the accuracy of the model established in this paper was 80.39 per cent. And, the model could identify the driver's braking intent prior to his braking operation.\nResearch limitations/implications: The limitation of this study was that the experimental environment was ideal and did not consider the surrounding traffic. At the same time, other actions of the driver were not taken into account when establishing the braking intention recognition model. Besides, the verification results obtained in this paper could only reflect the results of a few drivers' identification of braking intention.\nPractical implications: This study can be used as a reference for future research on driving intention through fNIRS, and it also has a positive effect on the research of brain-controlled driving. At the same time, it has developed new frontiers for intention recognition of cooperative driving.\nSocial implications: This study explores new directions for future brain-controlled driving and wheelchairs.\nOriginality/value: The driver's driving intention was predicted through the fNIRS device for the first time.\n",
"role": "assistant"
}
] |
Knowledge workers deserve differentiated offices and workplace facilities
|
[
"Workplace",
"Survey",
"Facilities management",
"ANOVA",
"Knowledge workers",
"Offices"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Today, most industrialized countries are highly dependent of their knowledge workers' productivity. This paper is a study of how workplace design, facilities and services facilitate knowledge workers' ability to do their job. The European standard EN15221-1 defines Facilities Management (FM) as \"Integration of processes within an organization to maintain and develop the agreed services which support and improve the effectiveness of its primary activities\" (CEN, 2006). Application of FM ideas and principles is, thus, of high relevance for those who aim for workplaces that support and improve the effectiveness of organizations' primary activities; hereunder improving organizations' ability to achieve their goals through application of their knowledge workers' knowledge and skills. This explorative study of factors that are important for the knowledge workers' (the end-users') perceived productivity at their offices and workstations took place in three Norwegian knowledge organizations, one institution for research and higher education and two consultancy companies.\n2. Literature: The past decades' development of ICT has made virtual workplaces far more common. However, physical offices are still the most common workplace for knowledge workers. The present research emphasizes physical workplaces and offices.\n3. Methods: Knowledge work is often intangible and thereby \"difficult to map and assess\" (Greene and Myerson, 2011). It is thus rather difficult to measure knowledge workers' productivity and thereby the effect of different kinds of workplaces and interventions on workplaces (Ramirez and Nembhard, 2004; van der Voordt et al., 2016; de Been et al., 2016). One approach to overcome these challenges is to measure the \"perceived (impact of facilities on) productivity\" (de Been et al., 2016, p. 149). This approach was chosen in the present research. One possible alternative to measurement of the respondents' perceived productivity through a questionnaire survey is an experiment where randomized experiment and control groups of knowledge workers with similar background working with similar tasks in different office environments where the knowledge workers' actual productivity in each environment is measured. However, such an experiment is difficult and most likely also rather costly to arrange. There are also most likely still several X-factors in the productivity equation, which are difficult to control even in experimental settings. Thus, given the available time and resources, the chosen approach measuring the respondents' perceived productivity seems justified.\n4. Results: This section includes three subsections, the first about the respondents, the second about the EFA and the last about testing two-way ANOVA models of the effect of the two independent variables Type of office and Hours spent at the workstation during the week on the dependent variables.\n5. Conclusion: This paper has presented the results from an explorative large N survey (N = 1670) of knowledge workers' perception of different categories of offices and factors facilitating productive workplaces. The response rate of 38.2 per cent is good for this kind of survey in Norway. The respondents are fairly well representative for their organizations. In all, 1,053 respondents (67 per cent) had cell offices. In all, 459 (29 per cent) had flexible offices with reserved positions, and 50 respondents (3 per cent) had flexible offices with free seating.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate factors that influence effective workplace designs for knowledge workers.\nDesign/methodology/approach: During spring 2016, the employees in a large institution for research and higher education, a large consultancy company and a medium-sized consultancy company (in total 4367 employees) in Norway received invitations to participate in an anonymous online survey about workplaces and facilities. In all, 1,670 employees answered the survey (38.2 per cent response rate). The data have been analyzed with IBM SPSS version 23, among others through use of exploratory factor analysis and two-way ANOVA.\nFindings: Most respondents at the institution for research and higher education have cell offices. Most respondents in the two consultancy companies have open and flexible offices. This paper indicate the respondents' preferences or perception of their workstation and the workplace's fit for their tasks is affected both by the respondents' type of office and how much time they spend at their workstation during the week. There are also possible age or generation effects.\nResearch limitations/implications: One methodical weakness in the present paper is that two-way ANOVA has been applied on survey data. Experiments are usually arranged to provide almost equal numbers of observations in each category. This is usually not possible with survey data. However, despite this weakness, the present paper provides several findings that challenge some of the workplace research's taken for givens.\nPractical implications: The present paper indicates that facility managers and others responsible for office and workplace design are advised to take the employees' tasks and work patterns into consideration when designing workplaces and providing offices and workstations to their end-users. The present paper also indicates that employees require different kinds of support facilities and services depending on what kind of offices and workplaces they have.\nOriginality/value: This is a large N empirical study among knowledge workers in three organizations, one public administration and two private enterprises. The present paper indicate that provision of offices and workstations with supporting facilities should be differentiated according to the end-users' work tasks and work patterns.\n",
"role": "assistant"
}
] |
Golf tourism - World Cup Cricket 2007 - Guyana prospects
|
[
"Golf",
"Guyana",
"Tourism",
"Cricket",
"Sports"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: With the phenomenal growth and expansion of sport tourism (Williams, 2003; Ritchie and Adair, 2004; Hudson, 2003; Weed and Bull, 2004) there has been a corresponding proliferation in the range of activities engaged in, participated in, attended and marketed from a tourism standpoint. Gammon (2004) in a tradition that recalls Graburn (1989) interprets this phenomenon as part secular pilgrimage where homage is paid to gods/heroes at shrines/stadia and other sacred places/sports venues. The volume of media attention (Hudson, 2003) that has been directed towards sports events has had the effect of making both sport and tourism images instantly accessible to a large international audience.Golf is, pre-eminently, a growth sport (Williams, 2003) that has been expanding in terms of the devotees (both professional and amateur) who play the game, the courses on which the sport is played worldwide and the revenues earned by those occupying the higher rungs of the professional ladder (Palmer, 2004). Despite its often controversial evolution (Swarbrooke, 1999; Backes et al., 2002; Palmer, 2004) it remains a sport almost hallowed in its place in the sports portfolios of countries rich and poor, developing or developed. It should therefore come as little surprise that it now enjoys a symbiotic association with a completely different sport - the game of international cricket - at the world championship level. Venues selected for the hosting of 2007 World Cup Cricket (CWC 2007) matches are now paying attention to their golfing facilities as ancillary entertainment for those who will be travelling to the Caribbean to participate in the CWC 2007 event. Golfing opportunity of course translates into additional revenues for those host countries.This paper discusses the phenomenon of golf tourism and explores the main theoretical issues and controversies that emerge in the literature. The developments relating to the organisation and administration of the sport in Guyana, its facilities and infrastructure are then examined against the background of the upcoming CWC 2007 event. Finally an assessment is made of the prospects for the development of a golf tourism sector in Guyana in the context of an overarching sports tourism framework.\nGolf and tourism: Emergence of golf\nTheoretical issues: Despite the prominence that golf enjoys in the panoply of world sporting events and the popularity of golf tourism as a tool for diversifying tourism product development (Readman, 2003; Palmer, 2004) extending the traditional tourist season and generating increased tourist revenues, its development is nevertheless the subject of some controversy in the literature. Part of the controversy that is generated by the growth of golf and golf tourism relates to issues that are not necessarily specific to golf or to golf tourism per se, but in fact are part of the general problematique of tourism development especially as that occurs in developing or less developed countries, and as it involves the relationship between (usually wealthy) tourist-generating regions and the (recently emergent) tourist-receiving regions.The broad areas of controversy can be itemised in four areas:1. the North/South political economy;2. the environmental domain;3. the socio-political arena; and4. the sociological consequences.The North/South political economy\nBenchmarking the golf tourism sector in Guyana: Classification\nThe CWC 2007 stimulus: Audience and scope\nThe prospects for Guyana: There are four factors critical to any assessment of the prospects for the development of a golf tourism sector. At least one of these factors relates to and directly recalls an issue explored in the literature. These factors are:1. the present size of the \"golf establishment\" in Guyana and the state of facilities;2. the level of recognition accorded golf as a growth area in a national sports tourism portfolio;3. the scope of current golf development plans;4. perceptions regarding the social milieu of golf and golfers.The golf establishment\nConclusion: Well established in the literature (Stoddart, 1994; Pattullo, 1996; Readman, 2003); is that the exclusivity of the golf milieu in no way inhibits its growth and development potential, nor retards its success as a tourism generator. So its elitist feature should not be understood, or misunderstood, as a factor restricting its development in Guyana. It is, however, the force and weight of the other three factors that prompt the author to express a doubt that CWC 2007 would to any significant degree accelerate the march of Guyana as a golf tourism destination.The limited golf infrastructure currently in place, the abbreviated scale of development plans for the facility, the absence of any golf tourism policy noises, the modest profile of the golf establishment all bespeak limited increments by 2007. The most likely, and hopeful, scenario is that CWC 2007 will constitute a pressure of missed opportunity and unfulfilled demand that could be the catalyst for a reassessment of the whole concept of golf tourism and its product and market potential in Guyana.The other more uplifting scenario is that publication of this conceptual paper will serve as a \"wake-up call\" to the Guyana Tourism Product Development establishment to focus scarce resources on this opportunity that carries remarkable return on investment potential. Perhaps motivated international corporate partners could be identified to accelerate golf tourism infrastructure/superstructure development in a timely manner to capitalize on the probably once-in-a-lifetime, international tourism market exposure of Guyana as a blessing of CWC 2007.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to advance the sport of golf as a compelling enhancement of the Caribbean region tourism product and assess prospects for the development of golf tourism in Guyana, catalysed by World Cup Cricket, 2007.\nDesign/methodology/approach: - Documentation derived from leading researchers in the field certify contemporary golfing's international steep growth curve and global tourism implications. Additional theoretical issues include environmental consequences of Caribbean golf course development are explored. Attention is then directed toward analysis of golfing infrastructure/superstructure in Guyana.\nFindings: - Participation in the sport of golf is indeed experiencing a major global upswing, especially among non-traditional devotees, largely due to the high international profiles of non-traditional golf professionals exemplified by Tiger Woods and Vijay Singh. While some Caribbean destinations will profit considerably, the golf tourism carrying capacity of Guyana will require major upgrading to reap significant benefits.\nResearch limitations/implications: - The implications should constitute a clear wake-up call to Guyana and other Caribbean regional tourism establishments if golf tourism is to be taken seriously as an enrichment component of the tourism product available to participants of CWC 2007.\nPractical implications: - Tourism planners and developers will be able to utilize these findings as a road map to establishing or enhancing golf tourism in their respective destinations, particularly in the Caribbean region.\nOriginality/value: - The authors believe that their particular conceptual approach to the challenge of exploiting the potential golf tourism windfall from CWC 2007 represents a valuable contribution to the literature of tourism development and a seminal research piece that will find its way into the Professional Golf Management degree curriculum resources of regional and international universities that have a legacy of post-colonial national development.\n",
"role": "assistant"
}
] |
Samuel Selvon under scrutiny: an annotated bibliography of selected criticism of Selvon's novels
|
[
"Bibliographies",
"Literature",
"Literary criticism",
"Trinidad and Tobago",
"Caribbean"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: More than five decades after the publication of his first novel, A Brighter Sun, the writings of Samuel Selvon continue to attract the attention and interest of literary researchers on an international level. A number of West Indian, Canadian, English, French, Australian and German authors have offered critical analyses and explorations into the fictional works of this Trinidad-born writer who helped to define the literature of the English speaking Caribbean. Between 1952 and 2007, over two hundred scholarly items have examined the style and themes in Selvon's novels, a few of which were written in foreign languages. The richness and the quality of these analyses, have created the need for this select annotated bibliography of the critical appraisals of Selvon's novels.The literature review reveals that three comprehensive bibliographies of the critical reviews of Selvon's novels are incorporated in Critical Perspectives on Samuel Selvon, Foreday Morning and Something Rich and Strange. An examination of these books helped to guide the research process and highlighted the honorary distinction and consideration afforded to the writings of Selvon. Susheila Nasta's Critical Perspectives on Samuel Selvon is comprised of articles, pamphlets and notes, 13 of which were issued in the 1970s and eight in the 1980s. Most of these were reproduced in Foreday Morning edited by Kenneth Ramchand and Susheila Nasta. Something Rich and Strange by Martin Zehnder offers 75 critical studies of Selvon's works which spans the period 1955 to 2001. The citations of Zehnder reveal that the majority of critical works (41) were published in the 1990s.This bibliography places emphasis on the critical works of Selvon during the period 1989 to 2007. The timeframe serves to fill the literary gap since Nasta's publication in 1988 and includes the contribution of selected critics in the 1990s who are highlighted in Zehnder's compilation. Moreover, it should be noted that the bibliographies of Nasta, Ramchand and Zehnder are not annotated. The researchers have now taken the opportunity to produce the first annotated compilation of the critical perspectives of Selvon's novels. The purpose of the bibliography is to provide a resource guide for researchers and students studying Caribbean literature and also to assist institutions embarking on a collection development drive to support Caribbean diasporic studies.Since Nasta's work (1988), many writers have attempted to analyze, evaluate and study the writings of Selvon in the form of books, journal articles, book chapters, dissertations, newspaper articles, conference reports and audio-visual materials. Whilst Zehnder outlines three dissertations in his list of citations, research reveals a complete absence of academic treatises in the bibliographies of Nasta and Ramchand. As a result, building upon the valuable contribution of doctoral and master's students in the evaluation of Selvon's works; dissertations issued prior to 1988 are reviewed and included in the study. The forte of this bibliography lies in the annotative aspect, the treatment of dissertations and the inclusion of non-English speaking documents. However, an obvious limitation is the omission of newspaper articles, conference reports and audio-visual materials. This annotated bibliography is by no means exhaustive, but it seeks to identify the English and non-English books, journal articles, unpublished works and electronic resources that have analyzed various aspects of Selvon's novels primarily during the period 1989-2007.\nBackground: In 1923 Samuel Dickson Selvon was born in Trinidad and Tobago to a father of East Indian descent and a mother of East Indian and Scottish descent. After a brief journalism career in Trinidad, he migrated to the United Kingdom in 1950 and two years later he gained renowned attention with the publication of A Brighter Sun. His writing career gained momentum with other publications, including An Island is a World (1955), The Lonely Londoners (1956), Turn Again Tiger (1958), I Hear Thunder (1963), The Housing Lark (1965), The Plains of Caroni (1970), Those Who Eat the Cascadura (1972), Moses Ascending (1975) and Moses Migrating (1983). In addition to his novels, Selvon co-wrote a screenplay with Horace Ove entitled Pressure (1978) and a number of other plays, which include You Right in the Smoke (1968) which was written for radio, Anansi, the Spider Man (1974) a television play for BBC and Switch (1977) a stage play performed at the Royal Theater, London. In recognition of his literary accomplishment he received a number of awards, including two Guggenheim Fellowships (1955, 1968), Trinidad and Tobago Hummingbird Gold Medal for Literature (1969), Best Novel Award from the Writers Guild of Alberta for his novel Moses Ascending (1983), an honorary doctorate from the University of Warwick (1989) and Trinidad and Tobago Chaconia Medal Gold for Literature (posthumously 1995). In 1978, he moved to Alberta, Canada where he was a writer in residence at the University of Winnipeg and died during one of his regular visits to Trinidad and Tobago on April 16, 1994.\nRationale for selecting Selvon: The Samuel Selvon's Special Collection resides at the Main Library, the University of the West Indies, St. Augustine, Trinidad and Tobago. In 1985 and 1988, Selvon donated his personal collection, which included essays, manuscripts and stories dramatized for radio and television, to the University of the West Indies. At the time of Selvon's generous presentation, Max Richards, the principal of the University of the West Indies, St Augustine commented, \"what we have been given is a life's work and the inner record of a soul's progress.\" In 1985, Selvon received an honorary doctorate from the University of the West Indies and maintained a close relationship with the institution.Selvon is counted among the literary giants in the English speaking Caribbean, and his works form part of the Caribbean diasporic literature. His Tiger and Moses novels are studied by researchers and students at both secondary and tertiary levels. Salick (2001) correctly advocates that \"the bulk of Selvon criticism focuses primarily and generally on his style, too often emphasizing style and neglecting all other aspects of his fiction.\" The exploration of themes such as racial identity, Caribbean language, community life, migration, colonialism and independence has resonated with those who have read his works. Looker (1996) states that in \"Selvon's fiction various histories collide: local and imperial, rural, urban, popular and elite.\"Ironically, despite the absence of a higher education, the eloquence and charm of Selvon's work provoke an ongoing curiosity among scholarly critics and reviewers. Selvon's novels capture various aspects of the historical and social experiences of the people of his native land. His writings transcended cultural borders at a time when West Indians had little or no channel to express their unique speech patterns; he captured the migrant's sense of displacement in metropolitan centers with an idiosyncratic sense of humor. James (1999) confirms and supports the pioneering role that Selvon played in the Anglo-Caribbean diasporic movement in his admission that \"the first major novel of Trinidad development was Sam Selvon's A Brighter Sun.\" The writers who have taken the time to critically examine the fiction of Selvon have sought to highlight his understated creativity in his use of language to create unforgettable characters, plots, and style. This select annotated bibliography seeks to call attention to the critics and researchers who have made valuable contributions, and serves to generate further discourse on Selvon's work.\nMethodology: The quest for resources that offered critical analyses and reviews began by searching the Main Library's Online Public Access Catalogue (OPAC) at the University of the West Indies, St. Augustine. Apart from the university's St. Augustine campus, the OPACs of the other campus libraries at Mona, Jamaica and Cave Hill, Barbados were also accessed. The card catalogue of the library's Special Collections was searched and proved to be a valuable source for articles in non-West Indian journals. CARINDEX, an index hosted by the University of the West Indies, was a resource for West Indian journals and theses. Another useful tool in the data collection process was the accessibility of several online electronic resources. For example, a keyword search of Samuel Selvon in the Modern Language Association database retrieved 86 hits which included full-text articles from scholarly journals, abstracts to journals, theses, and books relating to Selvon. The documents that were not held locally were accessed through inter-library loans, which included foreign language materials that were requested in English. The challenge of translating non-English documents was resolved through the use of online translators, English abstracts and in-house translators. Another search tool used to locate articles and documents was Google Scholar. Using the keywords Sam Selvon, 296 hits were retrieved and three quality articles were found.Unpublished works, such as doctoral and masters' theses, were selected for the annotated bibliography because they have undergone the process of peer evaluation. It is, for this reason, over thirty undergraduate projects were identified at the Main Library, West Indiana Division were not included in the bibliography. However, postgraduate theses originating from West Indian and non-West Indian universities are included, and the selection criteria of the latter relied on the availability via UMI ProQuest Digital Dissertations.To a certain extent, the retrieval and selection process was also controlled by the globally dispersed locations of Selvon's critics. The entries in the bibliography capture various authors and publishing bodies scattered across regions and countries. In addition, many writers have submerged and allied the critical works of Selvon with that of other literary writers in and out of the Caribbean, such as V. S. Naipaul, George Lamming, Earl Lovelace, Roman Chalbaud and Patrick Chamoiseau. Whilst emphasis was placed on the selection of critical perspectives solely on Selvon, the study also considered the weight and attention given to Selvon in documents covering multiple authors.\nBibliography: Articles\nArrangement of bibliography: Entries adhere to the Modern Language Association citation style and are organized according to the physical format of the item, such as articles, books, book chapters, dissertations and electronic resources. Within each format, the annotations are arranged alphabetically by authors. The bibliography offers translation of the foreign language documents and these are placed according to respective formats.\nConclusion: This bibliography provides a common destination for researchers who are interested in the critical perspectives of Selvon, and it also brings to the fore the multiple formats available for individuals and institutions wishing to augment their Caribbean literature collections. Originality is evident in its annotated element and the treatment of non-English published and unpublished works. In the final analysis, this paper serves to pave the way for future bibliographers to uncover and highlight critical studies of Selvon's other works, such as plays and short stories.\n",
"role": "user"
},
{
"content": "Purpose: - This paper intends to provide an annotated bibliographic guide for scholars and researchers studying the critical works of Samuel Selvon, one of the founding fathers of the Anglo-Caribbean literary movement.\nDesign/methodology/approach: - The authors identify and evaluate items from print bibliographies, online and card catalogues, electronic databases and the world wide web.\nFindings: - Numerous scholars have critically assessed the novels of Selvon and this is evident in the existence of published and unpublished works in English and other foreign languages.\nOriginality/value: - The forte of the bibliography lies in the annotative aspect and the inclusion of unpublished works and non-English speaking documents.\n",
"role": "assistant"
}
] |
Making sense of dementia: A phenomenographic study of the information behaviours of people diagnosed with dementia
|
[
"Information",
"Dementia",
"Phenomenography",
"Diagnosis",
"Behaviour",
"Need"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Dementia\nStudy aims and objectives: The overall aim of this study was to develop a better understanding of how people recently diagnosed with dementia need, seek and utilise information.\nMethods: Research design\nResults: Characteristics of the sample\nDiscussion: People with dementia have different needs for information. Participants used and understood information given to them in different ways, dependent on their reaction to the diagnosis, their existing perceptions of dementia and the extent to which they felt that they could influence their situation. The people with dementia in this study actively attributed different meanings to dementia and understood dementia in different ways. This affected their experience of making sense of dementia which, in turn, impacted on information behaviour. This confirms Dervin's (2005) proposition that the value of the same information to different people is likely to vary, according to how they conceive of their situation and the nature of the gap that they experience (Ford, 2015).\nConclusions: Although information from external sources can be used to help make sense of a diagnosis, people often rely on their internal beliefs and existing perceptions of dementia; consequently information given by health professionals will often be put to one side and not considered. For those that did need and use information to try to come to terms with dementia, some found the information they were given caused anxiety about the future and did not meet current needs. Understanding individual preferences for information of people with dementia requires bespoke information packages, distinct from the information given to carers. Developing strategies to encourage information acquisition (by overcoming feelings of futility) may help the person with dementia to feel more in control, to take a more active role in the management of their condition and to plan for the future.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the information behaviours of patients newly diagnosed with dementia.\nDesign/methodology/approach: This is a cross-sectional qualitative study, using in-depth interviews with 13 people recently diagnosed with dementia.\nFindings: Reactions to a diagnosis of dementia varied and these influenced the perception of the value of information when making sense of the diagnosis. Information was avoided if participants did not feel that they could influence their situation; instead, participants relied on internal explanations to normalise their memory loss. Barriers to information seeking and use included not knowing who to speak to, perceived stigma associated with dementia and difficulty of applying generic information to own situation. Some participants valued information that confirmed their suspicions and provided explanations.\nResearch limitations/implications: This study was based on a small sample size (n=13), the findings may not be generalisable to all people with dementia; however, the findings may be transferable to people who have recently been diagnosed with dementia.\nPractical implications: There is not a one-size-fits-all approach to information provision for people with dementia at diagnosis, information should be tailored to individuals.\nSocial implications: There is a need to address the feeling of powerlessness and futility that some people with dementia experience at diagnosis, as this precludes independent information seeking and use. People receiving a diagnosis may need additional support and information pertinent to their specific circumstances, separate from the information needs of their carer(s).\nOriginality/value: The study provides a new understanding of the information behaviours of people recently diagnosed with dementia and how these differ from those of informal carers.\n",
"role": "assistant"
}
] |
Convergence as organization: Blockupy against the ECB
|
[
"Crisis",
"Organization",
"Social movements",
"Activism",
"Convergence",
"Protest"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: On 18 March 2015, thousands of protesters from the transnational Blockupy coalition, angry with the effects of crisis and austerity, descended on Frankfurt. Their intention was to block the inauguration ceremony of the European Central Bank (ECB) thereby \"interrupt[ing] this celebration of power and capital [...] in the heart of the beast of the European crisis regime\" (Blockupy Frankfurt, 2014). Confronted by a considerable police presence, a range of organizations participated in the event, including activist groups and networks, movements, NGOs, trade unions, and political parties. The protest unfolded as a dynamic between resistance and police repression, as Frankfurt became a space of barricades and tear gas, blockades, and water cannons. For some, the event was chaotic and lacking accountability; for others it was an efficient mode of organization and paralyzation of the city; for the majority it was probably a combination of both. Based on participant observation of 18M, and in-depth interviews with Danish activists, this paper argues that the organization of a mass protest event like Blockupy's \"18M\" is not adequately captured by mainstream conceptualizations, and should be analysed as a complex convergence of various modes of spatiotemporal organization at different scales.\nOrganization of the mass protest event: Mainstream conceptualizations of organization cannot fully capture some of the principal modes of organization evident in mass protest events. As two organization scholars note: \"when we move away from the terra firma of corporations, non-profits, and social organizations, we find ourselves on shaky ground in determining what is and what is not organizational\" (Wilhoit and Kisselburgh, 2015, p. 573). Organization theory must be pushed towards new critical directions capable of developing an analytical framework in which the \"shaky ground\" of the mass protest event can be interrogated. This is particularly important at a time where networks are disproportionately emphasised, to the extent that \"[i]t may appear as if organization [...] is a phenomenon of the past\" (Ahrne and Brunsson, 2010, p. 84). Defining organisation as \"a decided order, including one or more of the elements of membership, hierarchy, rules, monitoring and sanctions\", Ahrne and Brunsson (2010, p. 84) suggest that \"complete\", in contrast to \"partial\" organizations involve decisions on all these elements (see also Haug, 2013). While this broadens the concept of organization, it is not enough to capture many of the organizational modes involved in contemporary social movements (Reedy, 2014). Competing theories of social movement and organizations elsewhere pose similar limits. For example, the popular concept of \"social movements organization\" (SMO) describes an organization that parallels a firm as it rationally competes with other SMOs for survival in a wider social movement industry - its primary goal (McCarthy and Zald, 2015). Beyond the problem of such \"marketization\" and depoliticization of organization and struggle, this theory also rests on too formal a notion of organization, based on economistic and positivist premises of rational behaviour and self-interest. Similarly, overtly \"cultural\" approaches to social movements suffer from an unhelpfully narrow conceptualization of organization. In this context, Jasper (2014, p. 82) refers to the situation in which \"organizational\" issues of funds, staff, rules, offices, etc. come into conflict with the movement's \"original mission\", that is, its political goals as \"organization dilemma\". Such a formal division of organization and politics does not take us far in analysing the organization of the mass protest event that is both complex and political.\nMethods: The methods employed for this study were participant observation during 18M, complemented by in-depth interviews. There are two interrelated reasons for this choice of methods. First, in attempting to grasp the organizational dynamics of the mass protest, it is necessary to get as close to the empirical phenomenon as possible. We had previously participated in other mass protest events, and in the build-up to 18M we participated in coordination meetings and then in 18M itself, on equal terms with other activists. It was only during that day in Frankfurt that we began talking about writing an article about the event. One of us travelled with the Danish participants on one of the buses to and from Frankfurt. This participation gave us the opportunity to give a rich, while necessarily partial, narrative of how the event unfolded. To add to our own observations, in Denmark, we conducted in-depth, semi-structured interviews with a group of 11 Danish Blockupy activists (two of them had not participated in 18M but in previous Blockupy protest events). At the time of the interviews, we had not decided on a specific research question, or even the theme of organization. Rather, we formulated our approach on the basis of our participant observation and the experiences and concerns of the activists interviewed. To identify these experiences and concerns, we did semi-structured interviews, beginning each interview with the very broad question of how the interviewee had experienced 18M. While this approach did not obliterate all the differences between \"academic interests\" and \"activist interests\", this academic paper can hopefully further conversation about organization among sociologists and activists alike.\nBlockupy's organizational roots: We begin our analysis by looking at the organizational macro-scale and the organizational roots of Blockupy. Blockupy is a transnational coalition of various European left groups and networks that recently organized a series of protests events and other actions. The name refers both to the coalition and to its protest events. The coalition contests the role of the ECB in the crisis management of the Eurozone debt crisis. It seeks to create solidarity with people affected by this crisis management and a transnational convergence space capable of facilitating a pan-European movement.\nActivist organization in the streets of Frankfurt: Having now established the development of the Blockupy coalition's organizational macro-structure preceding the 18M, the following section discusses activist organization as it played out in three different spatiotemporal settings at the mass protest event. While the first two subsections operate primarily on the meso-scale of minor informal organizations, the third includes the micro-organization of the subject. First, the bus journey from Denmark to Frankfurt proved to be a crucial organizational setting for many participants as they used it to organize affinity groups; then, the initial organizational \"chaos\" the Danish activists experienced as they unwillingly became part of a black bloc; finally, a bridge set the scene for a confrontation between protesters and police which became an important micro-scale organizational encounter for the activist subject. Together, these three discussions construct a spatiotemporal narrative of the modes of organization unfolding through and producing a convergence space in the streets of Frankfurt.\nOrganizational routes: This last section turns to the organizational routes that emerged from 18M as they were produced at the scale of the individual activists, the meso-scale groups, and the scale of the Blockupy coalition as a whole. As our interviewees were rooted in different organizations, identities and histories before 18M, they had multiple routes coming out of it. As hinted at above, the black bloc experience played an important role in this respect. Some interviewees felt that their activism of choice had been silenced and this decreased the prospect of them participating in or organizing mass protest events again. One activist was of the impression that Blockupy had been \"a demobilizing experience for many\" (C3). Another said about the black bloc actions: \"You could have done this any other day but instead of doing this, you are a parasite on an existing event, end up de facto pulling off a coup and make it about something altogether different\" (C2). He described the event as \"totally alienating\" (C2) and thought it unlikely that he would participate in a mass protest event again. 18M had made him more of a Leninist, he said, implying that he was now convinced of the organizational importance of a sort of \"democratic centralism\". These activists felt a breach of organizational trust and accountability at the scale of the coalition, as Blockupy had not been able to manage its \"troops\", but also on the scale of the other activist groups and individuals who had not acted according to the action codex. If the action codex is an organizational tool by which different identities and tactics are linked, not abiding by this codex has organizational consequences beyond the organizational horizon of the particular protest event.\nConclusion: With its analysis of Blockupy's 18M, this paper has demonstrated how a mass protest event is a convergence space of multiple modes of organizations, identities, and tactics. This goes against other approaches asserting that the modes of organization of contemporary activism and protest are necessarily rigid and formal, or that they can be reduced to fluid or flat spaces with little or no conflict.\n",
"role": "user"
},
{
"content": "Purpose: On 18 March 2015, the transnational anti-austerity Blockupy coalition protested the inauguration of the new European Central Bank premises in Frankfurt. The purpose of this paper is to analyse this mass protest event by highlighting the organizational differences, possibilities, and conflicts that was involved.\nDesign/methodology/approach: The paper is based on participant observation of the Blockupy event and interviews with a group of Danish activists who also participated.\nFindings: The paper constructs sociospatial narrative that unfolds through three different scales of organization: the Blockupy coalition, the participating formal and informal organizations, and the activist subject. This narrative explicates the mode of organization as a \"convergence space\" (cf. Routledge, 2003), with different \"roots\" and \"routes\" of organization (cf. Davies, 2012).\nOriginality/value: Thus, through an analysis of the modes of organization constituting this mass protest event, this paper restates the relevance of the concept of organization, which have recently been ignored or understated in favour of master-narratives of networks or the dichotomy of horizontalism and verticality. It concludes by posing a set of questions for further discussion among both activists and sociologists.\n",
"role": "assistant"
}
] |
Factors affecting quality of management education in India: An interpretive structural modelling approach
|
[
"Management education",
"MBA",
"ISM",
"Interpretive structural modelling",
"Management institutes",
"Quality of education",
"Quality of management education"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Growth of management education in the last few decades has been phenomenal in India. India has outdone all other countries of the world in the increase in the number of management institutes and seats (Philip, 2008). Where in the 1950s, 1960s and 1970s, merely four institutes were set up on an average every year, this number became as large as 283 in the new century (see Table I) meaning thereby that the average annual addition increased more than 40 times in the last 35 years. At the time of birth of the All India Management Association, the largest network of management institutes in the country in 1988, there were just 100 of them. Thereafter, they \"doubled every five years, following a kind of a geometrical progression\" (Bowonder and Rao, 2004). According to the regulator of management education in India, the All India Council for Technical Education, the increase in numbers has been the maximum in the case of management programmes as compared to any other post graduate stream. The number of management institutes increased from 1,888 in 2005-2006 to 3,858 in 2010-2011 indicating an almost 100 per cent growth rate in five years (see Table II). The seats grew almost fourfold from 95,000 in 2005-2006 to 360,000 in 2010-2011 resulting in a five year compounded annual growth rate of 30 per cent (ASSOCHAM, 2013).\n2. Theoretical background: Quality\n3. Factors affecting quality of management education: In all, 42 research papers written and published between 1974 and 2011 on the theme of management education in India and abroad listed on the EBSCO database were considered to identify the factors (see Table III). An account of these factors is given ahead.\n4. Methodology: The objectives of the study were met through primary and secondary research. It was completed in three phases:\n5. Results: ISM-based model\n6. Discussion: Based on the results, some points of relevance for quality improvement of management institutes operating in the contemporary environment are discussed here.\n7. Conclusion: This study was based on the identification and modelling of factors affecting the quality of management education in India. All the identified factors were found to be important and leadership emerged as the most important factor. Significant interrelations were revealed which are sometimes not exposed by mere observation. Many institutes lack focus and clarity in devising and employing measures for quality improvement. Consequently, they fail to make an impact. The framework proposed in this paper can be utilised by them to develop strategies which are focused, practical and effective.\n",
"role": "user"
},
{
"content": "Purpose: The purpose and value of management education was always under the critics' scanner but the proliferation of institutes impelled a serious debate on its quality. The purpose of this paper is to identify the factors affecting quality of management education in India and explains their nature, significance and mutual influences using interpretive structural modelling (ISM).\nDesign/methodology/approach: The factors were listed through literature review. They were then validated by empirical research conducted through questionnaires administered electronically and personally to 220 master of business administration students and alumni. On 13 such factors finalised, a qualitative and interpretive tool, ISM was applied.\nFindings: Leadership emerged as the most important factor followed by organisational structure and practices. Interrelations otherwise not easily observable established their prominence. An important fact that evolved is that almost all the factors have strong interdependence and have to be seen in coherence when analysing their impact on students.\nOriginality/value: The literature until now has been highlighting the factors and their association with management education largely in isolation. This paper contributes to the existing literature by proposing a framework of the interrelationships of the factors which have a role in improving the quality of management education.\n",
"role": "assistant"
}
] |
Public private partnership (PPP) outcomes in e-government - a social capital explanation
|
[
"Social capital",
"Critical success factors",
"E-governance",
"Public-private partnerships"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: This paper addresses the question of how and why public-private partnerships (PPP) projects succeed or fail, especially in the context of complex e-government initiatives. The challenges faced by a PPP project during its implementation and the issues of relationship and co-operation between the public and private partners need to be studied in detail.\n2. Literature review: There are several definitions of e-government, but the general consensus among authors is that electronic government entails much more than the evolutionary process of systems and technologies; it includes organizational, institutional, and administrative practices too (Gil-Garcia, 2012; Gronlund and Horan, 2004). Therefore, electronic government is a much more complex phenomenon, one that extends well beyond internet technologies (Luna-Reyes et al., 2008). Gronlund and Horan (2004, p. 714) in their review of definitions further add that \"all definitions of e-government go beyond services to the citizen to include organizational change and the role of government.\" Hence, we build on these views (i.e. socio-technical systems, and an intent of organizational change) and define e-government as a capacity to transform public administration through the adoption of ICTs to enhance the ease of access of government information and services to citizens, non-government organizations, and other government entities. This definition articulates the complexity of e-government initiatives, with emphasis on \"transformation\" and \"capacity\" creation expectations from e-government projects, and the challenges encountered in the adoption of ICT as an innovation. Transformation in e-government projects consists of various stages - beginning with billboard, partial service delivery, portal with integrated service delivery, seamless integration of e-services across administrative boundaries, and finally interactive democracy (Torres et al., 2005). The internal management capacity, consisting of knowledge and skills, is significantly associated with performance, productivity, and organizational effectiveness of e-government projects (Brudney and Brown, 1998; West and Berman, 2001; Chen and Perry, 2003).\n3. Methodology: We, first, constructed a theoretical framework based on the review of the relevant literature (e-government, IS outsourcing, and PPP) and formulated a set of exploratory propositions, which wait to be tested and validated. We, then, used a case study of an e-government project with a view to only illustrate or instantiate our propositions to improve the comprehension of our theoretical framework and the attendant propositions. Such a use of case study for instantiation is in line with earlier studies (Foster and Heeks, 2013; Cordella and Willcocks, 2012). We chose the e-government project called Nemmadi for this purpose because - it was a pioneering project, it was a complex large-scale project with a strong citizen-interface and consisted of automation of workflows too, it was a failed PPP, with several challenges during the implementation[1], and we were fortunate to get permission from the e-government department, government of the State of Karnataka, in India, to study the project.\n4. Explanatory framework for PPP outcomes: 4.1 PPP and organizational determinants of social capital\n5. Project Nemmadi case study - PPP in e-government: Nemmadi was an e-government project initiated in 2004 by the government of the State of Karnataka, located in the southern part of India (Das Aundhe and Narasimhan, 2012). Aimed at improving the transparency, accountability, and efficiency of the government administration at the village level, Nemmadi was the first and the largest G2C (referring to information, interactions, transactions, etc., between the government and citizens) e-government project.\n6. Case analysis and discussion: 6.1 Nemmadi case: project complexity, outsourcing, and outcomes\n7. Conclusion and implications: This study is an attempt to explain how and why certain factors, which have been reported in the PPP literature as the soft CSFs, influence the success of PPP models in e-government, a public ICT innovation adoption. This study has developed two propositions based on the extant literature of e-government and PPP, and used the social capital theory to integrate the propositions into a tentative explanatory framework. An exploratory case study of Nemmadi, a complex Indian e-governance project was used to instantiate the elements of the proposed explanatory framework, grounded in the extant literature.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explain how and why the intangible critical success factors (CSFs) determine the outcomes of public-private partnerships (PPP) projects in e-governance.\nDesign/methodology/approach: Drawing from the literature, PPP was conceptualized as an organization which facilitates the creation of social capital. It is argued that the intangible CSFs correspond to the key dimensions of the social capital that drives the development of intellectual capital in the course of addressing the challenges faced during the execution of an e-governance project. These efforts determine the accomplishment of the desired project objectives. The emergent framework was applied to an e-governance PPP project to anecdotally exemplify how the deficiency of organizational characteristics impedes the formation of social capital, resulting in project failure.\nFindings: A theoretical framework was developed to illuminate the mechanisms and the attendant propositions that explain how and why the intangible factors influence the PPP outcomes.\nResearch limitations/implications: This study fills a critical gap in the literature on PPP projects in general, and on e-governance projects in particular. It also extends the application of the social capital framework from an intra-organizational to an inter-organizational context.\nPractical implications: The results of this study foster a better understanding of the drivers of success in managing a PPP model in e-governance initiatives.\nSocial implications: This research work would help in improving the formulation and management of PPPs in the emerging economies, which could potentially enhance the societal outcomes.\nOriginality/value: The explanatory framework of this research serves as a useful perspective to address policy and program implementation issues of PPP initiatives in e-governance.\n",
"role": "assistant"
}
] |
Mitie made mighty by a woman at the top: Diversity pays rich dividends for outsourcing specialist
|
[
"Diversity",
"Gender",
"Women",
"Top executives",
"Health and safety",
"Awards"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: A company whose chief executive is the first Asian woman to head a FTSE 250 company has been named one of The Times Top 50 Employers for Women for a second consecutive year.\nMaking the most of the company's diverse talent: Ruby McGregor-Smith, who set up and remains involved in a mentoring circle for black, Asian and minority-ethnic women, said: \"We are delighted to have been recognized for a second consecutive year by The Times for our commitment to making Mitie the very best it can be for female staff.\"Mitie is a fast-growing and ambitious business that will reach its potential only if it makes the most of all the diverse talent within the company, including women. By making sure that we have the right environment and support for our female staff we will succeed together as a company.\"MITIE also has a dedicated diversity steering group that is externally chaired and made up of 14 diversity \"champions\" who have undertaken management courses in equality and diversity and are responsible for supporting the gender agenda and implementing initiatives across the business.Gender diversity in recruitment and mentoring is a strong focus for Mitie, with the company recently embarking on a targeted recruitment campaign to encourage women into engineering apprenticeships.Ruby McGregor-Smith, recently named one of the Financial Times' top 50 women in world business, herself combined motherhood with full-time work. She took an 18-month career break to take care of her second child, who had been ill.She believes that what got her through was having a lot of support and flexibility at work, and not having to feel guilty about needing time off when she did. Now she is determined to offer this to her employees, too.Since Ruby McGregor-Smith was promoted from chief financial officer to chief executive in 2007, the company has more than doubled in size to 63,000 employees across the UK. It reported a 26 percent rise in its order book, to PS8.6 billion, and an 8.9 percent increase in profits, in the year to March, as the public and private sectors sought efficiency savings.Helen Wells, director of Opportunity Now, which manages The Times Top 50 Employers for Women list and judging process, said: \"Any organization included in the list has to be commended for its commitment to actively recruiting, retaining and progressing female talent.\"Businesses are dependent on having a highly motivated and diverse workforce and being included in The Times Top 50 Employers for Women list demonstrates a focus on creating an inclusive and diverse workplace that uses the skills and talent of a diverse workforce. It is a clear statement that senior leaders see this as a commercial imperative, not a women's issue.\"\nOutstanding risk control: In a third major accolade for 2012, Mitie's facilities-management business has won the gold medal in the latest Royal Society for the Prevention of Accidents (RoSPA) Occupational Health and Safety Awards.This is the fifth consecutive year that the business has won gold. The award is for its approach to the prevention of accidents and ill health, recognizing its achievements in demonstrating outstanding risk control.The awards look not only at accident records, but also the entrants' overarching health and safety-management systems, including practices such as strong leadership and workforce involvement.Mitie has developed a Work Safe, Home Safe communication campaign, designed to raise awareness of health and safety for all employees and managers. Its innovative approach to staff communication has ensured more employees are aware of health and safety procedures, thereby lessening the risk of accidents.The business has a target of reducing its accident rate by 10 percent a year. It targets such problematic areas as slips and trips - one of the largest health and safety issues across most industries.Jeff Hilton, head of quality, health, safety and the environment at Mitie, said: \"By involving the workforce and ensuring health and safety are part and parcel of their day-to-day work, the business is able to develop pragmatic and effective solutions to health and safety issues.\"We are constantly thinking of new ways of working to help to reduce the number of accidents - for example, by access to the right equipment and advanced cleaning systems.\"David Rawlins, awards manager at RoSPA, said: \"The RoSPA awards program provides well-deserved recognition for the winners and spurs on other organizations to raise their standards of accident and ill-health prevention.\"We congratulate Mitie on its success and encourage it, and all our other winners, to remain committed to safety and health, an approach that is well recognized to be good for workers and the bottom line.\"\nNote: David Pollitt, Human Resource Management International Digest Editor, wrote this article.\n",
"role": "user"
},
{
"content": "Purpose: - This article's aim is to describe how strategic outsourcing and energy-services company Mitie Group plc is helping more women to reach senior executive posts.\nDesign/methodology/approach: - The measures the company has taken are described as well as some of the results they have achieved.\nFindings: - The article highlights measures including: a program to encourage women to return to work and develop their careers after taking maternity leave; a maternity-coaching program to help to retain senior female talent; flexible working hours and a flexible-leave system; home working and a compressed working-fortnight option; child-care vouchers; additional annual-leave purchase schemes; a diversity week; a mentoring program for female rising stars; and a twice-yearly employee forum at which gender achievements, progress and initiatives are discussed.\nPractical implications: - It is explained that the Mitie chief executive, the first Asian woman to head a FTSE 250 company, has set up and remains involved in a mentoring circle for black, Asian and minority-ethnic women.\nSocial implications: - The article points up the importance to society of ensuring that women enjoy equal opportunities in the workplace.\nOriginality/value: - The article discusses a company that places huge importance on attracting and retaining female talent while delivering a supportive and inclusive environment for all its people.\n",
"role": "assistant"
}
] |
The impact of customer equity drivers on loyalty intentions among Chinese banking customers: The moderating role of emotions
|
[
"China",
"Loyalty intentions",
"Value equity",
"Emotions",
"Brand equity",
"Relationship equity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In the contemporary world, banking industry around the globe is under constant rivalry. Gradually increasing customer desires along with intensifying product and service portfolio of the banks are some of the prime reasons underlying in the competitive banking industry. In order to survive in this ever-evolving economic workplace, banks have to sustain a healthy relationship with their customers (Ernst and Young, 2011). This crafting of solid relationships with the clients becomes more important when you are in the field of giving financial services as compared to any other service providing industry in the market (Tung and Carlson, 2013). Therefore, a primary responsibility resides on the shoulders of the bank managers to comprehensively cater the changing needs of their customers in order to restrict them from moving to other banks (Chiu et al., 2005). Now the marketing activities of many banks have been transformed into accomplishing customer loyalty (Izogo and Ogba, 2015), the reason being losing customers in the banking sector is considered as a potential loss and a point of severe concern (Sweeney and Swait, 2008; Chi and Gursoy, 2009; Fathollahzadeh et al., 2011; Akhter et al., 2011).\n2. Literature review and hypotheses development: 2.1 Customer loyalty\n3. Emotions: Emotions are viewed as the most integral part for the survival of human being (Niedenthal and Ric, 2017). In order to comprehend emotions in a better way, first, there is a need to understand the underlying theories of emotions to exactly know their background and how they are developed. Schachter and Singer (1962) came up with a theory called two-factor theory of emotions and proposed that emotions are formed as a combination of two things: automatic arousal - which can be triggered by anything; and a label - that defines the experience with respect to that specific situation. Two-factor theory envisages that a human being upon experiencing a stimulus would label himself or herself \"happy\" or \"fearful\" by the nature of situation or environment he or she experiences, and the result would be termed as emotion. On the other hand, development theories of emotions acknowledged that our emotions are also formed by the social experiences, comprising cultural traditions and customs (Saarni, 2008). Development theories encapsulate that emotion is the amalgamation of both nature (gene) as well as nurture (experiences), i.e. the place where the individual is being brought up (Burnette et al., 2013). Furthermore, Darwin suggested that human emotions evolved because of their adaptive value. According to him, fear, for instance, evolved because it aids individuals to act in ways that ameliorate the probabilities of survival, as it alarms about the possible dangers in the surroundings (Darwin, 1859).\n4. Methodology: 4.1 Survey design and sampling\n5. Results: 5.1 Measuring instruments reliability and validity\n6. Discussion: Considering the prior discussion on the importance of emotions and equity drivers in service encounters, this study presents empirical research that supports the link of emotional aspect of loyalty in creating loyalty intentions. Moreover, both positive and negative emotions have been detected as playing a vital role with respect to Chinese bank customers.\n",
"role": "user"
},
{
"content": "Purpose: The implementation of customer equity drivers (CED) as a crucial marketing tactic to surge customer loyalty intentions has received a considerable importance in the literature. However, most of the research done in the past has mainly centralized around western societies. To make it even more interesting is the fact that the significance of customer emotions has been ignored by the previous studies. Therefore, the purpose of this paper to explore the impacts of CED on loyalty intentions along with exploring the moderating role of customer emotions (positive emotions and negative emotions).\nDesign/methodology/approach: A sample of 661 Chinese banking customers was collected by making the use of store-intercept survey design. The gathered data were then utilized to empirically validate the proposed model by making the use of hierarchical moderated regression.\nFindings: Loyalty intentions were found to be driven by emotions of Chinese banking customers. Consequently, in order to better forecast the loyalty intentions of the customers, the emotional aspect is vital and therefore should be incorporated along with other cognitive aspects (value equity, brand equity and relationship equity).\nPractical implications: The managers of the banks should make every effort to make the visit of their customers as pleasant as possible as the emotional responses of customers have a significant impact on the formation of loyalty intentions.\nOriginality/value: The current study holds its unique contribution by including emotions in the service-oriented settings.\n",
"role": "assistant"
}
] |
The nature of peer coaching: definitions, goals, processes and outcomes
|
[
"Coaching process",
"Peer coaching",
"Coaching outcomes",
"Theory",
"Definition",
"Peer interaction"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Coaching, from a managerial perspective, has been represented in the literature since the 1950s (Evered and Selman, 1989), thus literature on the topic spans several decades and the nature of coaching behaviors and skills has been widely discussed in the literature (Hagen and Peterson, 2014). Development of this construct began as a transitioning of the coaching concept from an athletic context to that of organizations. This has been illustrated by a move from an apprenticeship-type relationship in the 1950s to a more traditional approach to organizational coaching as a tool for improving \"problem work performance\" then finally to one in which learning and development are core activities (Ellinger et al., 2006, p. 3). Along with this transition, organizational coaching, in general, appears to have taken on a life of its own with an explosion of new articles and research on the topic (Beattie et al., 2014). The topic of coaching more generally has been researched thoroughly; however, both scholarship and practice overwhelmingly have focused on two specific types of coaching: executive coaching, which generally utilizes a professional coach, and managerial coaching, which generally uses coaching between manager and subordinate (D'Abate et al., 2003). Meanwhile, fewer studies on peer coaching are available in publication. Research on peer coaching has been substantively limited, with some researchers referring to it as \"embryonic\" and calling for specific peer coaching tools and models to improve scholarship (Kowalski and Casper, 2007).\nThe importance of peer coaching: According to Parker et al. (2015) and Bennett and Bush (2014), one area that is particularly under-explored is that of peer coaching in the for-profit organizational context. Their research noted an increase in the practical implementation of peer coaching within organizational context, as indicated by an increase in professional development in the area, noting that many organizations are implementing peer coaching as a low-cost tool for professional development (Parker et al., 2015). This increase is illustrated by the International Coaching Federation's (ICF) Reciprocal Peer Coaching program. This program, designed as a peer development opportunity for professional coaches, connotes the growing importance of peer coaching within the broader field of coaching (ICF, 2016), and helps to illustrate one of the many ways in which peer coaching is filtering into the organizational context. One particularly well-used model for peer coaching in the for-profit context is Whitmore's GROW model. This model from popular literature has been applied to both managerial and peer coaching contexts. The model emphasizes the coaching process through goals development, review of multiple options and sustained effort. What is notably missing from the model, however, is a distinction in that process between managerial coaching, when the coach has hierarchical power, and peer coaching situation, when no power differential is evident.\nResearch purpose and question: Hamlin et al.'s (2008) important work reviewed a range of coaching definitions and helped to define coaching as a single, unified concept that spanned several variants, providing some clarity to what coaching is, and what it is not. However, it is noteworthy that much of the literature reviewed in Hamlin et al.'s work fell within the areas of general coaching, executive coaching, business coaching and life coaching; peer interactions that fall within the coaching context are not addressed. Of particular note is that their review and definition dealt primarily with coaching from the perspective of coaching as a profession - one in which the coach was a trained professional, generally external to the situation, who took on the role of a professional coach; this definition is not as useful in this situation, as it does not apply well to the peer coaching context.\nMethodology: This research uses a combination of inductive and deductive qualitative approaches through the implementation of qualitative content analysis (QCA). The use of content analysis has become increasingly popular among management researchers, and has become particularly popular in organizational studies, as it is useful in learning about the complexities of organizational life and systems, with themes derived via a qualitative application process (Cho and Lee, 2014; Erdener and Dunn, 1990; Duriau et al., 2007). Most recently, QCA has been described as a \"method for systematically describing the meaning of qualitative material\" (Schreier, 2012, p. 1). It is worthy of note that this research is not intended as a semiotic analysis, which generally is associated with the deeper meaning of themes (Neuendorf, 2002), nor is it intended to be interpretive research, which is generally constructivist in nature (Walsham, 2006). Rather, the methodology proposed for this research is intended to provide an \"objective, systematic, and quantitative description of the manifest content\" (Berelson, 1952, p. 18) contained within the peer coaching literature in order to make \"valid inferences\" from the ensuing text (Weber, 1990, p. 9).\nResults: To answer the research question, the researchers chose to analyze only those data that applied specifically to the issue at hand, definition (Cho and Lee, 2014). Thus, only those articles that included a direct reference to peer coaching definition were included in the analysis. Inclusion was determined in two ways:\nDiscussion: The purpose of this research was to aid in the development of a definition for peer coaching that was derived from current literature for use in multiple contexts. The resulting definition, and the methodology from which it derived, helps to create a deeper understanding of the practices that contribute to its effectiveness. This definition is intended to help organization development (OD)/training practitioners, managers and scholars apply the concept of peer coaching consistently throughout its many varied applications, contexts and environs in a way that makes sense. Regardless of the focus of interest or intent, it is important that peer coaching participants and facilitators have a thorough understanding of what peer coaching is, and what it is not. As suggested by D'Abate (2003) over a decade ago, there is confusion on the concept and definition of coaching; this issue continues in the area of peer coaching even today. In particular, there are three areas that continue to cause confusion or contradiction within the literature:\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this research is to create a definition of peer coaching using literature from various and disparate organizational and educational contexts. This research is intended to clarify what constitutes peer coaching overall, and guide the ever-growing practice of peer coaching.\nDesign/methodology/approach: This research uses a combination of inductive and deductive qualitative approaches through the implementation of qualitative content analysis (QDA). The research used a data reduction process with 87 existing scholarly articles in the area of peer coaching, in which the researchers focused on selected aspects of the data and followed the practice of staying close to manifest artifacts within the data.\nFindings: The results of the qualitative data analysis indicated that five themes emerged within the literature. Themes were as follows: program structure, purpose and goals of peer coaching, peer coaching processes and mechanisms in which peer coaching is conducted, outcomes of peer coaching and relational contexts and functions of peer coaching relationships.\nOriginality/value: While there is an increase in peer coaching, research on this type of coaching is scarce with few empirical studies on the topic: what has been conducted has not been particularity reliable (Hagen and Peterson, 2014); and most of the published research is limited to education, nursing, other medical contexts and non-profit organizations. This research helps to clarify the nature of peer coaching and create a cogent definition that defines formal peer coaching within all peer coaching contexts.\n",
"role": "assistant"
}
] |
Assessing the relationship between market orientation and business performance in the hotel industry - the mediating role of service quality
|
[
"Service quality",
"Market orientation",
"Hotel industry"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Market orientation research suffered a boom in the 1990s. Former studies on market orientation by Kohli and Jaworski (1990) and Narver and Slater (1990) carried out the foundations for the following research on this subject. The 1990s were fruitful for market orientation literature. During this period, several market orientation scales were developed (Deshpande et al., 1993; Deshpande and Farley, 1998; Jaworski and Kohli, 1993; Kohli et al., 1993; Narver and Slater, 1990; Ruekert, 1992), although MKTOR scale (Narver and Slater, 1990) and Jaworski and Kohli (1993) scale, or its reduced version, MARKOR scale (Kohli et al., 1993), established the cornerstone on market orientation measurement.\n2. Literature review and hypotheses development: 2.1 Market orientation and business performance\n3. Methodology: This work aims to analyze the relationship between market orientation and business performance and to evaluate potential mediating effects of service quality on this relationship. To perform this analysis, a survey was sent to the managers of a sample of Portuguese hotels.\n4. Results: This research used a component-based structural equation modeling (SEM) and SmartPLS 3.0 (Ringle et al., 2015) software to analyze data (PLS: partial least squares).\n5. Discussion, conclusions and limitations: 5.1 Discussion and conclusions\n",
"role": "user"
},
{
"content": "Purpose: This study aims to explore the nature of the relationship between market orientation and business performance in the Portuguese hotel industry and to evaluate the mediating role of service quality in this relationship.\nDesign/methodology/approach: A literature analysis on market orientation, service quality and hotel business performance was conducted. The structural model was developed, and the measurement scales' items were generated. Working data were obtained by sending a survey to Portuguese hotel managers. A purification process was conducted, and the convergent and discriminant validity, as well the reliability, of the measurement models were assessed. The structural model was tested by conducting a path analysis, as well by evaluating the coefficient of determination (R2) and the predictive relevance indicator (Q2).\nFindings: Research results show that market orientation has a positive direct effect on business performance. Moreover, results indicate that market orientation effects on business performance are mediated throughout the hotels' ability to provide service quality.\nOriginality/value: Despite the large amount of research on market orientation, research on the hospitality sector has been conducted mainly using tools developed for the manufacturing and consumer goods sectors. On the other hand, little has been done to explore the nature of the mediating effect of service quality on the relationship between market orientation and business performance in the hotel industry. Moreover, despite the contribution of the tourism industry to the Portuguese gross domestic product, research on the hotel industry is scarce.\n",
"role": "assistant"
}
] |
Complexity science approaches to the application foresight
|
[
"Complexity",
"Disruptive events",
"Emergence",
"Reframing",
"Variation",
"Policy making",
"Complexity theory",
"Uncertainty management",
"Strategic planning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Foresight techniques: To date, foresight techniques (see http://hsctoolkit.tribalhosting.net/The-tools.html for a summary of techniques, applications and case studies) have developed out of need (see www.andyhinesight.com/category/foresight-2/ (accessed August 2011) re practitioner usage); they are highly pragmatic, were originally developed for use in commercial organisations, and are used because they deliver useful insights However, other than some post-graduate foresight programs (provided on the basis that foresight is an interdisciplinary field grounded in a variety of social science), there has never been much underlying theory, or rationale, for either foresight or its techniques, Gheorghiou et al. (2008) and Miller (2011).In many ways, foresight techniques also originally arose out of the need to deal with unexpected events in the technology and political arenas, for example the classic Shell scenarios. More recently the development of techniques for coping with disruptive events, often referred to as wild cards, has focussed on adapting scenario techniques, Ramirez and Van der Heijden (2008) or on wild card management, for example that based on weak signal methodologies, Mendonca et al. (2004). As complexity theory itself has developed, Mitchell (2009), so has some limited work in the foresight field, Ringland (2010) and Miller and Poli (2010), but there has been little specific consideration of how foresight techniques need to be developed more generally to accommodate complexity theory and its specific implications for disruptive events.In addition, there has been little published work, or understanding, on how to really apply (implement) foresight and how to extract real value from the results of foresight studies and projects. Much foresight work, while very interesting and great fun to discuss over dinner, goes unused, unappreciated, and makes little observable difference. Foresight may be either an art or a science, Bell (2003), and may or may not be, a discipline, (or perhaps be developing into one), but either way, a critical thinking approach can be helpful. Bringing together complexity theory, reframing and some critical thinking provides a way to understand how techniques can work better, be applied better, and most importantly lead to better decision-making and policy making (and perhaps contribute to use of theories of social change in foresight).\n2. Complexity: So what is complexity theory and what is its relevance to foresight? There are of course as many answers as people, including Prigogine and Stengers (1984), Byrne (1998) Mitchell (2009), Goodwin (1994), Strogatz(2003), and Waldrop (1992), but like any developing area there are many common themes. This paper will explore some selected elements of complexity theory and their implications for foresight techniques, and later derive implications for the application of foresight in policy making.There are five relevant elements of complex systems:1. A system cannot be explained by breaking it down into its component parts because the key element is the interaction between the parts. The system needs to be considered as a whole. As a result of these interactions complex systems exhibit emergence - (self-organised) behaviour that results from these interactions. For example, consider taxies in a city. The location and availability of taxies in a city cannot be explained by breaking the system down into its individual parts - drivers, cars, customers, fares, other taxies etc. Rather it is an emergent property of the whole system resulting from the interaction between all the parts - the taxies and customers, the road system, the traffic (itself an emergent property of a city's transport system), city taxi rules, and even the weather (the system environment):\n3. Improving foresight application - reframing: One way to think of all of these ideas together is as different ways of seeing, of being conscious and aware about the nature of the mental model, i.e. being a description of perception rather than a description of reality (Schwartz, 1991), or more simply as reframing the future landscape and potential strategies, options, and potential policies. In general we are trained to think in terms of linear causality; doing anything else is difficult, disturbing, and different. Complexity theory can provide a newer way of looking at change and shows why reframing becomes important. Whatever we say about the future has an implicit idea of change underlying any inference and most theories of social change have an underlying implied pattern, although we are not always aware of it. And while no theory explains everything, complexity potentially offers us the closest we have ever come to having an overall theory of social change.In simple terms Reframing is:* changing the way you think;* changing the way you see things;* changing the way you understand things;* not trying to solve problems using the same frame they emerged from;* generating new solutions to issues; and* enabling alternative actions.An example of Reframing from Battram (2000): After much searching in many different areas of the city, you have finally found a possible flat to buy and you go for a walk to explore the area a little more. You pass a newsagent and decide to drop in for a newspaper and some chocolate. You pay and take your purchases but the newsagent gives you too much change. You realise, keep quiet and start to walk out of the shop; no-one will know... But then you think ... I may buy the flat, I may be coming here again, the owner will know who I am. You turn round and return the excess change.By simply thinking about a possible future (the purchase of the flat), you have changed your frame. And you have changed your opinion and even changed your actions. Axelrod and Cohen (2001) call it the \"shadow of the future\".From the perspective of applying foresight the benefits of reframing are:* realising that there are more choices than thought, and that those possibilities can be created;* bringing something new into existence;* creating something outside the cone of possibilities;* deconstructing an existing frame to explain or demonstrate to others that they are stuck in one frame; and* an ability to deliberately switch frames.In addition to using Reframing to enable a change in the way we see, think about, and understand things, we are often utterly unaware of how our ways of seeing are driven and constrained by our values. When unrecognized and unexamined, our preferences often pass for objective assessments of plausibility and probability; \"Impossible is just and opinion\". Some foresight techniques do use Reframing, for example in causal layered analysis, Inayatullah (1990), employs reframing at the deep myth/metaphor level.\n4. Results - policy making implications: We all have our favourite story of a failed or ludicrous policy where a lack of foresight is obvious - in retrospect. The lesson is often that the environment/time/space/area in which one works and lives in determines how one thinks (the mental model or frame), and not only is it difficult to get out of that \"thought channelling\" process, but one is usually not even aware that thoughts are being channelled.The key insight of complexity-based foresight for policy making is that command and control approaches do not work in complex systems. The resulting implication is that a system cannot be controlled from \"above\" and so policy operating in a complex system cannot achieve a specific outcome directly. Instead, policy making needs to embrace emergence - to exploit the developing behaviour that results from interactions between the parts of a complex system; specifically policy makers must focus on the idea of interactions rather than a system's constituent parts and develop a vision of the self-organised behaviour that could result from interactions between the parts.What does \"embracing emergence\" actually mean? There are different ways of understanding the idea and what it may mean for policy making. For example, emergence is a process of change and embracing it can mean choosing to respond, continuously, to that change, sympathetically and synergistically rather than a controlling, combative style: embracing not resisting. Another way is to accept that change is the force of control rather than being the result of controlling intentions. In this case, embracing emergence can mean supporting and encouraging the current situation, and its emergent change, in a way that ensures that generally desirable outcomes will thrive, irrespective of what those specific outcomes might be. It could also be about accepting that it is the route being travelled, rather than the unknowable destination, that matters, particularly when that route is uncertain.What then could embracing emergence mean for policy making? Simplistically, it is about understanding the system, in terms of a system's interactions rather than its component parts, so that policy is designed to stimulate evolution rather than to force control. Such policy making requires iterative monitoring of the emerging changes, to ensure that the desirable ones are supported and the undesirable ones diverted. Axelrod and Cohen (2001) describe the idea of harnessing complexity - to deliberately change the structure of the system, and to do so by changing the way in which a system is perceived - use of different mental models, or reframing. Again, policy makers need to watch for the \"emergent\" properties that arise as a system organises itself following a policy intervention, and use that policy to preserve the conditions in which the best solutions arise. With these perspectives, the link with, and importance of, the concepts of optimisation and exploration (resulting from the non-equilibrium state of a complex system - see section 2.5 above) also becomes clearer. The activities of embracing emergence, stimulating evolution, and harnessing the system, can all enable policy makers to see the system from a different perspective and to generate both optimum and (currently) non-optimum alternative potential strategies and options.In terms of policy making, \"embracing emergence\" can be seen as very different to many FTA techniques. Horizon scanning for example focuses is on the possible results and outcomes (futures) resulting from a change process, rather than working with(in) interactions of the systems components itself. Similarly with techniques such as trend or driver analysis, and even scenario development, the onus is on a vision of the final outcome, and policy options for that final outcome, rather than on a vision of the self-organised behaviour that could result from interactions between the components, and how that can be harnessed.Some writers, Battram (2000) and Swanson and Bhadwal (2009), have considered how complexity-based foresight can be applied to policy making. For example looking for shorter term, finer grained measure of success that can usefully stand in for longer-run, broad goals, \"trying it on and see what you get\", and using social activity to support the growth and spread of value criteria. Other methods include observing patterns, relationships, and rhythms rather than events (a specific example of focusing on interactions rather than constituent parts of a system), promoting effective neighbourhoods, building networks of reciprocal interaction, and promoting variation.Much as it might be interesting and intellectually satisfying to re-invent the whole policy making process in the light of complexity-based foresight, the current (and historic) general policy making environment is likely to be the one in which implications for the application of complexity-based foresight in policy making will take place, at least for a while. We will therefore use a simple, generic, policy-making model, Table II adapted from Bhimji (2009) - direction, design, and delivery - to explore one example of a complexity-based technique to deal with disruptive events in policy-making - Promoting variation.Promoting variation can provide a response to several of the requirements of foresight techniques identified above in Section 2. For example it can contribute to generating a range of future options and alternative potential strategies through both optimisation and exploration. Promoting variation can also contribute to embracing emergence and to enabling visions of phase change situations. In this case promoting variation also constitutes a risk management approach, whereby a policy is more able to work as its environment changes - while many of the policy interventions will fail (and failures are a normal feature of complex systems) having several options increases the likelihood that at least one option will succeed.Promoting variation in practical policy making terms can mean using several options to achieve an intended outcome as implementing a variety of policies to address the same issue increases the likelihood of achieving desired outcomes. In this case, reframing can bring the benefits of realising that there are more variations than originally thought of and of creating something outside the original cone of possibilities. But promoting variation can also be viewed as a set of \"parallel experiments\" being undertaken simultaneously with the aim of achieving a common objective. Here, reframing can bring several benefits. For example, providing an awareness of existing frames and ways of seeing, demonstration that those ways of seeing are driven and constrained by particular values or expectations, and being able to deconstruct an existing frame to demonstrate that a particular way of seeing is being used. In addition, reframing enables deliberate switching between frames to generate new experiment variations and options.There are at least three key ideas, Battram (2000) and Swanson and Bhadwal (2009), which can help policy makers to promote variation:1. Specifically promote variation by designing and implementing a range of alternative policy options to meet the various needs of different stakeholders. Using a mix of policy instruments, exploring synergies with other policies, providing opportunities for risk-spreading, and undertaking cost-benefit analysis.2. Create an enabling environment for variation by facilitating conditions that enable societies to create alternative approaches to achieve a common objective or to respond to a common issue. Identifying influencing factors and removing barriers to facilitating variation.3. Study from past and current experiences and adapt as needed by reviewing what policy interventions would usefully create or destroy variety and considering if the variety that results can offer potential value. Considering alternative sources of variety which could have greater promise and arranging organisational routines to generate a good balance between exploration and exploitation options.Table III shows how promoting variation can be applied in practice at each stage of the simplified policy making process:\n5. Conclusions: Complexity theory can provide a different perspective on how and why future disruptive changes may happen in a system. It also provides insight into how foresight techniques need to be developed to perform better in complex systems to enable better decision-making and policy making.Foresight methodology\n",
"role": "user"
},
{
"content": "Purpose: - This paper aims to present an exploration of recent work in complexity theory to explain why and how disruptive events happen in systems and how responses could be better, particularly in the policy-making arena.\nDesign/methodology/approach: - The main method applied is critical thinking combined with a review of selected aspects of complexity theory and a general experience of applying foresight. Several new and practical implications for foresight techniques and their application are derived. Promoting variation is examined as one way to make policies more resilient in a complex system.\nFindings: - Complexity science demonstrates that disruptive events do not need an associated trigger, as they are a normal part of a complex system. This insight implies that if we are always looking for weak signals we will certainly be caught unawares.\nPractical implications: - The assumption that disruptive events can be managed by planning and forecasting is not a workable option. Instead, policy making needs to assume that unexpected disruptive events will happen even with the best horizon scanning system in place. Foresight techniques need to be developed to embrace emergence and to provide capabilities such as reframing to visualise systems from very different perspectives, including those considered impossible now.\nOriginality/value: - Although neither complexity theory nor the concept of reframing is new in the area of foresight, the derivation of practical implications for foresight techniques is original.\n",
"role": "assistant"
}
] |
Analysis of the innovation value chain in strategic projects of the Brazilian Army
|
[
"Innovation process",
"Brazilian army",
"Defense sector",
"Innovation model"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Even though innovation has been more and more present in entrepreneurial competitiveness, it also affects other sectors, e.g. civil defense. The military activity, known for its work intensity and troop ships, started to receive intensive capital and innovation investments from the half of the nineteenth century onwards (Markusen, 1986); there was, therefore, a shift from an weaponry competitiveness toward a scientific competition (Paarlberg, 2004; Schmidt, 2013). If during the Second World War the source of the military supremacy corresponded to the industrial production capacity of weaponry of countries like the USA (Paarlberg, 2004), by the end of the war, the scientific capacities started to focus on military powers (Schmidt, 2013), producing successive technology generations and quick changes in the strategic military environment.\n2. Theoretical framework: 2.1 Organizational innovation process\n3. Methodology: Our research presents a qualitative approach and a multiple case study, whose focus is on the strategic projects of the Brazilian army. In the current study, we used a descriptive and exploratory approach to analyze the innovation process by means of primary data collection from interviews and secondary data from official documents.\n4. Analysis: According to the interviewees, there is an integration among projects as it is possible to identify common objectives and synergy, whose purpose is to develop an innovation system and a more modern army.\n5. Conclusion: implications and future research: Our study described the innovation process in the Brazilian army with four stages outlined in the conceptual model: idea creation, selection of the best ideas, development and adoption and diffusion. In order to meet the research goal, a qualitative study was carried out, focusing on seven strategic projects not only because they intend to change the land forces by modernizing and equipping it, but because they are focused on relevant issues, i.e. the protection of society, maintenance of law and order and defense of the country.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to describe and compare seven case studies of strategic innovation projects of the Brazilian army; these projects present high transformational potential and high investments and are supported by technology and science policies.\nDesign/methodology/approach: The authors present herein multiple case studies in which the authors conduct a documentary analysis of the innovation processes in the Brazilian army, as well as semi-structured interviews conducted with eight servicemen with more than 15 years of working experience.\nFindings: The results obtained suggest that the innovation process occurs in four stages: creation, selection, development and diffusion of ideas.\nPractical implications: The research is relevant because it presents how the interaction between the Brazilian army, companies and academia strengthens the innovation ecosystem, stimulating the development of best practices for the management of strategic projects.\nOriginality/value: The main contribution of this study is to present the strategic project management of innovation based on public policies and investment in projects of the Brazilian army, which are drivers for the development of ecosystems that promote the creation and expansion of companies, the diffusion of technological knowledge in universities, and suitable solutions for the military sector.\n",
"role": "assistant"
}
] |
Journals used for the publication of English psychiatry, surgery and paediatrics research
|
[
"Research work",
"Psychiatry",
"Surgery",
"Paediatrics",
"Journals",
"Financing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The National Health Service (NHS) spent about PS500 million on research in the year 2002/2003 (Department of Health, 2004). In an era of growing accountability it is relevant for questions to be asked about what knowledge is being produced and where it is being published (Buxton et al., 2000; Croxson et al., 2001; Harrison and New, 2002). At the same time, bodies that fund health research are increasingly being expected to show the benefits from their expenditure in terms that go beyond the criteria traditionally used in academic assessments such as peer review of the knowledge produced (Buxton and Hanney, 1996; Hanney et al., 2004; Lewison, 2003, 2004). This is especially true for institutions such as the NHS in the UK that could, instead of spending money on research, use it directly on providing health care.One aspect of this issue leads us, in turn, to questions about how far publication in peer-reviewed journals, the primary output from research, is an appropriate mechanism to use to disseminate research findings and encourage the use of research and achieve broader benefits. Although Coomarasamy et al. (2001) found that journals had not been an effective route for the dissemination of research findings, they concluded that medical journals were in a unique position to improve dissemination. Schein et al. (2000) reported that clinicians considered peer-reviewed journals to be their most important source of information.Given the large number of journals that are published, a key question becomes how important are the individual journals in which the research is published. An increasingly important way of assessing this is through the impact factors of journals, which show the average number of times an article in a particular journal is cited. This is a fundamental citation-based measure of the \"significance and performance of scientific journals\" (Glanzel and Moed, 2002). There is some evidence of correlations between the impact factors of journals and measures of quality such as rejection rates (Yamazaki, 1995), but various technical criticisms are made about the way in which journal impact factors (JIFs) are constructed (Glanzel and Moed, 2002). Despite warnings from the creator of impact factors (Garfield, 1996) they have been used as a convenient shorthand in the evaluation of researchers; an approach that has been strongly criticised in leading medical journals (Seglen, 1997). For example, problems arise because journals in different fields can have very different impact factors (Wellcome Trust and NHS Executive, 2001), which means that unsophisticated application of impact factors could damage the researchers in certain fields (Schwartz and Lopez Hellin, 1996). Furthermore, in the recent debate about the pressures currently faced by academic medicine (Bell, 2003) one of the key issues has become the sometimes inappropriate demands being made for researchers to aim their papers at journals with a high JIF (Abbasi, 2004).If the issues raised above are to be better addressed, several strands of analysis would be informative. It would be useful to know the extent and nature of the publication of health research and, in particular, for the link between publications and specific research funding to be better understood (Croxson et al., 2001; Butler, 2001). The research outputs database (ROD) provides bibliographic information on these issues in the UK. It is constructed by the addition of various details, especially on funding acknowledgements, to the data about all biomedical publications with at least one UK address that are held on the science citation index (SCI) and the social sciences citation index (SSCI) (Dawson et al., 1998). The NHS ROD, a subset of the ROD, contains details of papers from England that involve some element of NHS financial support (Wellcome Trust and NHS Executive, 2001). The determination of which papers involve NHS funding is not straightforward unless there is a specific NHS funding acknowledgement. Therefore, any paper describing research carried out on NHS premises is considered to have some element of NHS funding and is included in the NHS ROD. The nature of its construction means that the NHS ROD includes a wide range of research, including trials, much of which is primarily paid for by other funders but conducted on NHS premises.Many practitioners in the medical field still continue to rely on print journals (Tenopir et al., 2004). Therefore, notwithstanding the increased focus on individual papers that comes as a result of electronic access and the use of bibliographic databases such as Medline, it is still important to conduct some analysis at the journal level. Attention is now being given to the attitudes taken by clinicians towards individual journals. Saha et al. (2003) concluded from such a study that impact factors may be a reasonable indicator of perception of quality for general medical journals. Other studies, however, suggested that the relationship between impact factors and the journals viewed as significant by practitioners varied depending upon the type of science being conducted (Lewison et al., 2001).To gain a fuller understanding of these issues and the nature of the research funded by the NHS, it is useful to examine the pattern of journal usage for publication in various clinical specialities. This should allow common themes to emerge as well as allowing analysis of any speciality-specific issues. The first report on the NHS ROD (Wellcome Trust and NHS Executive, 2001) provided considerable information about overall patterns of publication in each of a number of specialities. This current article focuses on the three specialities of psychiatry, surgery and paediatrics, chosen for their strongly clinical nature, and examines the pattern of journal usage. For each speciality we report on three main strands of analysis.First, we consider how many, and what proportion of, English publications, in each of the three specialities, has some element of NHS funding and analyse the number of journals used to publish NHS and non-NHS research. Second, we identify the journals most used as publication outlets for each speciality and examine their JIFs. Third, we draw selectively on survey findings to show the variation in importance attached by clinicians to the journals most widely used to publish English NHS research.\nMethodology: The number of papers and journals\nFindings: The number of papers and journals\nDiscussion: Limitations\n",
"role": "user"
},
{
"content": "Purpose: - To identify the papers, and publishing journals, describing psychiatry, surgery and paediatrics research funded by the National Health Service (NHS) in the UK. To make comparisons with non-NHS research and examine the journal impact factors, and importance to clinicians, of journals publishing the most NHS research. To consider the implications, including those for research assessment.\nDesign/methodology/approach: - Existing databases were examined: the research outputs database (ROD), which contains information on UK biomedical papers; NHS ROD, which contains details of papers on ROD funded by the NHS; lists of journal impact factors. These were combined with selective findings from surveys conducted to identify journals read and viewed as important for clinical practice by psychiatrists, surgeons and paediatricians.\nFindings: - In each specialty many papers publish NHS-funded research and they out-number the non-NHS papers in the ROD. They appear in a wide range of journals but in each specialty one journal is clearly the most used. The impact factors of journals publishing the most NHS research vary considerably. In each specialty the journal containing most NHS publications is widely perceived to be important by clinicians.\nResearch limitations/implications: - Much NHS-funded research is also funded by other bodies. Clinician survey response rates were between 38 per cent and 47 per cent. The analysis could be extended to other specialties.\nPractical implications: - Papers published in the few journals in each specialty that are viewed as important by clinicians could be given additional credit in assessments.\nOriginality/value: - This paper describes outputs from NHS research and shows how assessment could be extended.\n",
"role": "assistant"
}
] |
The potential use of online tools for scientific collaboration by biology researchers
|
[
"Web 2.0",
"Scientific collaboration",
"Information behaviour",
"Usability testing",
"Online research",
"Virtual Research Environment"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Collaboration in science is of great importance and has been addressed by many authors including Olson and Olson (2000), Olson et al. (2002), Hara et al. (2003) and Birnholtz (2007). Scientific collaboration is defined by Sonnenwald (2007) as \"human behaviour among two or more scientists that facilitates the sharing of meaning and completion of tasks with respect to a mutually-shared superordinate goal and which takes place in a social context.\"\n2. Background: 2.1 Scientific collaboration and the adoption of new technologies\n3. Research questions: The aim of this study was to investigate the current research practices and requirements of biology researchers in UCD and to assess the overall suitability of the OJAX++ Virtual Research Environment for these researchers. The research was divided into four sub research questions:\n4. Methodology: 4.1 Research methods\n5. Results: 5.1 Heuristic evaluation and navigation stress test\n6. Discussion: This section discusses the results in relation to the four research questions of the study.\n7. Conclusions and further research: The aim of this paper was to examine the current research practices and requirements of biology researchers in UCD and to evaluate the overall suitability of the OJAX++ Virtual Research Environment for these researchers.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to discover the research practices of biology researchers and to assess the suitability of the OJAX++ Virtual Research Environment (VRE) for these researchers.\nDesign/methodology/approach: Usability testing was used to evaluate the usability of OJAX++ in relation to biology researchers. Interviews with biology researchers in a large Irish university were conducted to investigate their research information behaviour, to establish user requirements in their discipline and to evaluate the feasibility of using OJAX++ in their research.\nFindings: The results show that biology researchers used online tools extensively in their research but do not use social networking tools. E-mail and phone conversations are the preferred methods of collaborating with colleagues. The biology researchers found that OJAX++ was easy to use, intuitive and professionally presented but in its present format, OJAX++ does not fit in with current research practices as they do not use Web 2.0 tools that facilitate tagging. A list of requirements of a VRE for biology researchers is presented.\nOriginality/value: The findings of the study will assist developers of VREs and other web tools to better understand how researchers, in particular biologists, collaborate during the research process and what they require from online research tools. This study gives an important insight into the information behaviour of life science researchers.\n",
"role": "assistant"
}
] |
Psychological and environmental antecedents of impulse buying tendency in the multichannel shopping context
|
[
"Teleshopping",
"Buying behaviour",
"Consumer psychology",
"United States of America"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.T-commerce is a revolution in retailing. Television shopping programs, one of the major t-commerce, have become an alternative shopping method for traditional retail store shoppers. T-commerce, which is open 24 hours a day seven days a week, may be considered as a convenient shopping method since it is more accessible, compared to retail stores that are only open during regular business hours. With other various benefits to customers such as saving time and transportation cost for shopping trips, t-commerce has experienced a huge growth over time (Harden, 1996). Quality Value Convenience (QVC) and Home Shopping Network (HSN) lead the t-commerce market in the USA. QVC leads with $4.4 billion in sales in 2002 (Gross, 2003). Moreover, the total sales revenue of QVC in 2002 exceeded that of other selling channels such as Bloomingdale's Inc. ($2.0 billion) and Amazon.com ($3.9 billion) (Gross, 2003). HSN also reported $2 billion in retail sales in 2002 by reaching more than 140 million households worldwide. HSN shipped 90,000 to 130,000 packages per day (HSN.com, 2003). HSN also distributes the products via more than one retail channel such as the internet and catalog (HSN Company Profile, 2003). Although sales from television shopping programs are increasing, many purchases are returned. Yankelovich Partners, a market research firm, speculates that television shoppers' inability for physical product inspection and experience may result in a higher return rate, which is estimated between 20 percent and 40 percent. This could be an obstacle for television shopping.\nConceptual framework: Impulse buying\nMethod: Sample\nResults: Preliminary analysis\nDiscussion and implications: Out of six, five causal relationships proposed in this study were confirmed through structural equation modeling. The results showed that the proposed model for impulse buying and interaction tendencies in multichannel settings generally fit the data. Multichannel shoppers who had more interactions with salespeople purchased more products on impulse in the retail setting and in turn, purchased products on impulse in the television setting. In addition, customers who had interactions with salespeople in the retail setting had greater parasocial interaction with the hosts via television shopping programs and in turn, purchased more products on impulse in the television setting. Focusing on the television setting, parasocial interaction mediated the relationship between TV shopping program browsing duration and impulse buying from television shopping programs. This implies that people who are more exposed to television shopping programs are likely to have greater parasocial interactions with the show hosts and in turn, purchase more products on impulse via television shopping programs. These findings supported the past literature (e.g. Babin et al., 1995; Grant et al., 1991; Stephens et al., 1996).We expected a positive relationship between TV shopping program browsing duration and impulse buying tendency in the television setting. However, there was no direct relationship between the amount of time television shopping programs were watched and impulse buying among multichannel shoppers. This is not consistent with the previous findings (e.g. Jarboe and McDaniel, 1987) that browsing via television shopping programs positively influences impulse buying. In these data, browsing via television only affected impulse buying from television indirectly. It is possible that personal traits such as parasocial interaction and impulse buying tendency have strong and direct influence on impulse buying tendency in television setting. Personal traits may be consistent across different shopping contexts. Once a person develops personal traits relevant to shopping in the traditional shopping context, it will Based on the sequential relationship from browsing intention via parasocial interaction to impulse buying tendency via television shopping programs, parasocial interaction seems to possibly mediate between browsing duration and impulse buying tendency. Shopper who develop parasocial interaction from watching more television shopping programs may purchase more products on impulse.This study adds valuable empirical findings to the literature on the distribution channel relationships by examining buying behavior of multichannel customers who shop both via television shopping programs as well as at the traditional retail channels. The results revealed that customer-salesperson interaction and impulse buying tendency in the retail setting influenced parasocial interaction and impulse buying tendency in the television setting. This implies that each buying setting is not separate; rather multiple buying environments may interact and influence each other. Established buying tendency from one channel may influence purchasing from another channel. Impulse buying tendency may be consistent across different buying channels. This suggests support for an impulsive buying trait.In addition, some theoretical implications can be generated for impulse buying. In the past, impulse buying tendency was examined within a limited buying setting (e.g. traditional retail channels). However, reflecting new retailing and buying trends, impulse buying tendency as a personal trait in one shopping setting can be seen to influence buying behavior in other shopping settings such as televisions, catalogs, and the internet. Thus, impulse buying tendency as both cause and effect needs to be considered and tested in multiple shopping environments for external validity for impulse buying-related theories (e.g. Dholakia, 2000). In addition, Impulse buying tendency (M=5.88) in the television setting was slightly higher than that in the retail setting (M=5.59). This implies that the inherent nature of impulsive condition in television shopping settings can intensify impulse buying tendency. Although consistent impulse buying tendency was found in two different settings, certain shopping contexts can possibly strengthen or weaken the degree of impulse buying tendency as a personal trait. Thus, in future research, impulse buying tendency can be examined in other multiple combined buying settings (e.g. traditional channel - catalog - internet). Even though the internet is considered as more planned environment compared to catalogs and televisions, impulse buying tendency as a personal trait in traditional channels may positively influence buying behavior in the internet.Customers who are likely to interact with salespeople in the store may have greater potential for building parasocial relationships with the show hosts when they watch television shopping programs. Interaction tendencies in both settings may then positively influence impulse buying in the television setting. Thus, interaction tendency seems to be consistent across different buying channels. Similar to impulse buying tendency, it is also possible that interaction tendency may be another personal trait affecting buying behavior. Customers who enjoy interactions with the salesperson in the traditional channel may also like to build an interactive relationship and increase communications with the show host in the television setting.This study also contributes to knowledge about customer relationship marketing (CRM) for t-commerce. Multichannel customers who shop via television shopping programs as well as at retail stores may be willing to interact with salespeople either for entertainment or to alleviate loneliness. Thus, developing a close relationship between customers and salespeople in both television and retail settings may be beneficial to generate sales for t-commerce. If direct (physical interaction in the retail setting) or indirect (parasocial interaction in the television setting) positive customer-salesperson relationships are built, customers may purchase more. T-commerce marketers may encourage show hosts to approach the viewers (potential buyers) through friendly and entertaining comments that can evoke emotional responses for purchasing. This also may create a synergy effect that leads to more parasocial interactions with the host and in turn, increase the amount of purchases. In addition, direct communications over the phone during the show may be helpful to facilitate more interactions. Constant reminders of asking shoppers to stay tuned on the shopping channels may increase the amount of browsing via television shopping programs. However, t-commerce marketers should be aware of negative consequences of impulse purchases such as credit card debt or non-use of product purchased. Merchandisers should select guaranteed quality of merchandise and vendors to reduce negative consequences of impulse purchases for television shoppers. Easy process of returns may minimize returning effort and help customers to return products back to the distribution center. Special instructions for merchandise return can be announced and emphasized by the show hosts.This study has some limitations. Beyond low response rate, some responses may not fully reflect actual shopper behavior. Respondents may rely on their memories to recall their past shopping experience, but often recall inaccurate information. This may even more true for older respondents. Thus, browsing duration may be possibly inaccurate in the process where respondents might need to mentally generalize or average their viewing time.In addition, because multichannel shoppers were selected without temporal limitation, store patronage difference in two shopping contexts was not captured. Some respondents shop more often from television shopping programs, whereas others shop more frequently at traditional retail stores. For example, about 20 percent of respondents did not spend money for apparel last month in the retail stores. Although they reported that they have shopped at the traditional retail stores, they may not shop often. The similar observation can apply to television shopping context.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this study is to examine the effect of psychological traits and shopping environmental factors on impulse buying tendency via television shopping programs and to reflect the inherent nature of the impulsive television shopping environment in the USA as well as the traditional retail channel.\nDesign/methodology/approach: - A total of 154 questionnaires were returned from multichannel customers who purchased apparel from television shopping programs and traditional retail stores.\nFindings: - Five causal relationships among impulse buying and interaction tendencies in both television and retail settings and TV shopping program browsing duration proposed in this study were confirmed through structural equation modeling.\nResearch limitations/implications: - This study adds valuable empirical findings to the literature on the distribution channel relationship by examining buying behavior of multichannel customers as well as some theoretical implications for impulse buying-related theories.\nOriginality/value: - This study provides insights for customer impulse buying behavior in the multiple shopping environments.\n",
"role": "assistant"
}
] |
Getting the attention the facilities manager deserves
|
[
"Knowledge management",
"Facilities",
"Scotland",
"Computer applications"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Knowledge management is permeating every aspect of the facilities management role. Details on contractor performance, asset condition, people moves and benchmarking all reflect a concerted attempt to translate data into useful information. Computer-aided facilities management (CAFM) continues to play a key part in supporting knowledge management initiatives. But, simply providing more information appears not to be the answer. The messages communicated by facilities managers to senior executives and business leaders, though important, often go unheard. This arises from the excessive volume of information and data that is generated by the activity of facilities management. The clear lesson here is that \"less is more\". In this paper we consider the problem of \"acute attention deficit disorder\" as described by Davenport and Beck (2000). The evidence suggests that senior managers are not paying attention to the long-term or short-term goals of an organisation. This is a particular problem for the facilities management team. In the latter part of the paper, the results of an analysis of current practice in the use of computer-aided facilities management (CAFM) amongst practitioners in Scotland is explained. It considers the range of tools that are currently being used and the extent to which such systems are seen as supporting organisational goals.Knowledge management is a very human activity. Ultimately, the aim of any knowledge management undertaking is to influence the behaviour and decisions taken by senior executives or employees. However, it would be a mistake to believe that the Board of Directors act on purely rational grounds, in response to a facilities manager's proposals. As Davenport and Beck (2000) note, managers in the USA are subjected to over 100 emails and voicemails each day, in an environment characterised by \"attention deficit disorder\". In the era of the lean organisation, this means more information has to be consumed by ever fewer people. In the words of Davenport and Beck (2000):Leaders today need to pay more attention to attention because it is widely misunderstood and widely mismanaged.It is through this lens that this paper considers the adoption and use of computer-aided facilities management solutions. Attention is a scarce resource. Indeed it is perhaps the most scarce resource that the facilities manager has to consider in what is increasingly called the \"attention economy\".\nAttention economics applied to facilities management: Various authors have identified the significance of the \"attention economy\" and ways in which information and communication technology (ICT) needs to be modified to address this increasingly limiting resource amongst knowledge workers and managers. Most notable amongst these are the works of Castell and Jenson (2004), Davenport and Beck (2000), Davenport and Beck (2001), Falkinger (2007) and Roda and Nabeth (2006)In the context of knowledge management applied to FM it is worth considering some of the principles that apply to the attention economy as identified by Davenport and Beck (2000):* Attention as a scarce resource. This rule states that there is only so much attention to go around. Web designers and advertisers recognise this fact when designing web portals and search engines. Invariably loyalty to specific search engines such as Google or Yahoo arises not because users find they have access to many sites, but because they provide efficient access to the few sites that interest them. A key implication of this rule is that facilities management teams should consider limiting the number of initiatives and programmes competing for the attention of decision makers and building users. This even extends to the issue of safety and signage in an environment cluttered by superfluous information.* People are sensitised to factors affecting their survival. People invariably are sensitised to factors that are seen as threatening to their survival. In the business landscape, anything that threatens the survival of an organisation is likely to capture people's attention. Knowledge conveyed in this manner is likely to have more dramatic impacts simply because of the psychological disposition of human nature. The lesson for the facilities manager here is that \"scaring\" decision makers and employees may provide an effective tool for attracting their attention, but as Davenport and Beck (2000) point out, the threat must be genuine - otherwise employees and managers will \"stop believing the threat is real\".* Minimise distractabiliy. Given the number of competing sources of information with varying levels of immediacy and appeal, facilities managers need to minimise distractibility. Employees can become overwhelmed by change initiatives in the workplace. Added to this is the operation of the informal culture which becomes highly active during times of uncertainty: the gossip culture and the fear of job losses.* Engage people. Getting people to adopt ownership of some part of a knowledge base is a powerful mechanism for securing commitment. This builds on the idea of \"co-creation\" whereby individuals and customers of FM solutions can contribute to online discussions and so contribute to the knowledge base.Each of these points highlight the increasing importance of human behaviour in the knowledge management discussion in FM. Even with a growing wealth of information made possible through help-desk reporting, field data-capture and environmental scanning, the challenge still remains to capture the attention of those in a position to act.In the remaining part of this paper, the authors analyse the nature of the computer-aided facilities management usage in an intensively knowledge based economy (Scotland) and reflects on the extent to which CAFM is meeting the demands of a knowledge management environment.\nCAFM systems in support of knowledge management: The origins of CAFM are seated in the realm of information systems rather than knowledge management systems. The need for industry tailored IT tools in the FM sector resulted in the development of CAFM systems as early as the 1970s. Most of these early systems focused on the space planning problem (Fullenwinder and Reeder, 1973; Mitchell, 1974). Space planning systems use \"a digital computer to aid the designer in the planning of architectural, urban and regional spaces\" (Teicholz, 1975). Modern CAFM information systems are highly integrated, collating information about a built facility and can include modules to integrate computer aided design (CAD) drawings, geographic information systems (GIS) data and maintenance schedules amongst many others. Traditionally CAFM systems have provided a functional device designed to automate the labour intensive process of mass data collection, storage and retrieval.Abel and Lennerts (2002) stated that the main purpose of a CAFM system is to support operational and strategic facility management (i.e. all of the activities associated with administrative, technical and infrastructural FM that arise when a building is in use) as well as underlying processes. CAFM tools are popular as a means of improving facilities control and customer service, monitoring planned expenditure against actual performance and recording vital information in a single database. According to Fraser (2004) the future of CAFM is web-enabled systems, allowing multiple users to access the information and input new and updated data in real time, immediately accessible to all who rely on accurate system information. Applied to activities such as condition based maintenance, Finch (2002) demonstrated the significant reduced operating cost of wireless online monitoring which removes the need for labour costs associated with wired systems.In many smaller organisations the use of facilities management IT tools may be limited to an electronic room booking programme whereas in a large global corporation with multiple sites, the system may include elaborate building sensor arrangements linked to computerised management systems, graphic based software tools, online information databases, project extranets, direct business-to-business and business-to-customer procurement services. The second scenario is sometimes referred to as e-FM whereby the monitoring and correction of facilities can be carried out remotely, multiple locations can be linked via centralised management control and service requests can be actioned and tracked with minimal human intervention.The increased proliferation of IT in facilities management has changed the profile of data importance. Information influences the organisation's tactical use of facilities while acting as a direction setting mechanism to help the facilities manager to be proactive and to predict and foresee change and hence optimise strategic decisions to meet the aims and objectives of the business's long-term goals. FM automation has had a significant impact on the tactical and strategic performance of the business organisation (Lunn and Stephenson, 2000).However, what of the relationship between CAFM as an information system (IS) and that of a knowledge management (KM) system? Knowledge management, unlike conventional information management, is concerned with a variety of practices used within an organization to identify, create, represent, distribute and enable the incorporation of ideas and experiences. Such knowledge includes that possessed by individuals as well as that embedded in organizations. One of the challenges in knowledge management is to capture much more than the \"explicit\" knowledge which can be readily codified and stored on information systems such as a CAFM solution. The challenge of knowledge management is to go beyond simple \"information systems\" and capture the less tangible information contained in people's heads or embedded within the operation of an organization. This type of knowledge is termed \"tacit knowledge\" - a term coined by Polanyi (1966) who observed that \"We know more than we can tell.\"Modern-day CAFM increasingly addresses the possibilities of encoding \"tacit knowledge\". Indeed the central role played by \"help-desk systems\" in supporting customer reporting and feedback, now makes CAFM an indispensable part of any large facilities management operation. Added to this are the online communities that are used to support internal customers, suppliers and clients. Information of this kind is particularly valuable as it relies on providing context for people, places, ideas and experiences: providing tacit knowledge.\nPurpose of the study: The purpose of this research was to examine the significance of information technology in the delivery of facilities management services to the built environment and business organisations in Scotland. Secondly, the study sought to gather the views of facilities management professionals regarding the usefulness of these tools in improving the effectiveness of FM in relation to knowledge management.\nMethodology: To generate accurate and reliable data, specific information was sought from a stratified sample of facilities management practitioners. Access to the study sample was through the membership of the British Institute of Facilities Management (BIFM). The main body of data was generated using a structured questionnaire. Data were collected using an online questionnaire system which was subsequently analysed in a statistical database.The quantitative assessment provided a summary of the FM industries current practices, as well as the perceived effectiveness of IT in the delivery of FM services. Conventional descriptive statistics were then used to establish the usage patterns in the facilities management industry within Scotland:* to gauge the level of understanding with regard to CAFM tools amongst organisations;* to establish how well received CAFM has been within those organisations who have implemented it;* to examine the profile of those organisations most likely to implement CAFM systems; and* to establish the extent to which CAFM systems support aspirations of routinizing knowledge management.An initial questionnaire was piloted to establish the functionality of the tool and to generate feedback relating to the questionnaire. Following completion of the pilot, the results were analysed and some amendments made to the questionnaire design following feedback.In total, 395 questionnaires were sent to members of the BIFM Scotland region in 2007. This resulted in 110 usable responses representing a response rate of 27.8 per cent.\nResults: The findings were examined under four separate headings:1. Profile of respondents.2. Organisations using CAFM in service delivery.3. Organisations not using or considering CAFM.4. FM challenges and awareness of ICT applications.\nResearch findings: Profile of respondents\nConclusions: The use of CAFM technologies in the delivery of facilities management across the built environment in Scotland presents a mixed picture, with a multitude of factors influencing the FM professionals' attitudes. It can be seen that many of the application areas sought in CAFM go beyond conventional information systems, supporting the identification, creation, representation and distribution of knowledge: as such they can be described as knowledge management tools. Specifically, uses such as increased customer satisfaction and improved communication are likely to play an ever increasing role - with the support of online communities and structured knowledge capture using the semantic web. One can expect to see a sea-change in CAFM applications, with an ever increasing circle of stakeholders that are able to contribute to and abstract from CAFM information systems. This will facilitate an inevitable trend towards co-creation. Such a trend will engage users, suppliers, clients and operatives in a much more inclusive framework. The burgeoning of input technologies, including SMS (short message service), auto ID and field data entry devices will further enrich the data capture and retrieval process. The challenge will be to increase the transparency of such systems and foster \"stickiness\" so that users become willing users. To this end, many lessons can be learnt from current practice now in evidence amongst advertisers and web designers who seek to attract the attention of users. Inevitably the challenge will be to overcome the \"attention deficit\" of invaluable stakeholders. These include building users whose tacit knowledge provides the key to improved customer service; and senior executives whose attention needs to be drawn to critical facilities issues.An encouraging aspect of this survey for CAFM developers was that more than half of those FM organisations surveyed indicated that they were already using CAFM or were considering its implementation. Software developers would also realise from the findings that they still have a large untapped market in the FM world with 40 percent of companies surveyed not yet enjoying the benefits of CAFM, demonstrating that there is room for growth in the market, provided they can target their efforts in the right areas.The research shows that the use of information technology in the form of CAFM is more likely amongst dedicated FM service providers than organisations with in-house FM delivery and that medium-sized enterprises (between 51 and 250 employees) are least likely to be using CAFM. In terms of size, the organisation most likely to be using CAFM is one with more than 500 employees and working in the public sector. Those considering the use of CAFM are predominantly client organisations with in-house FM provision. This suggests that they are following the lead of the dedicated FM service providers who have been at the forefront of adopting CAFM and other information and communication technology tools to improve their delivery in what has become an increasingly competitive market over the last five years.The survey findings showed that of those facilities managers not using CAFM, many consider that the scale of their service delivery does not warrant the use of a CAFM system. This is surprising given the now relatively low starting costs for entry level CAFM software available on the market and most providers offering packages tailored to the individual needs of the organisation regardless of size, with specific products targeting the smaller organisation. This finding demonstrates that CAFM software providers need to target the right individuals; that the investment in CAFM will provide efficiency savings and improvements in service delivery.Above all, the message seems to be that CAFM provides the capacity to deliver vast amounts of data. Whether this translates into useful information will depend on a keen understanding of \"attentionscapes.\" The ultimate users of CAFM systems are not facilities managers: they are the senior directors who have to recognise the significance of the information; as well as the building users who wish to engage in the cocreation of a sustainable working environment.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to assess the role of computer-aided facilities management (CAFM) in the delivery of knowledge management solutions. It aims to highlight the increasing problem of \"attention\" as a limited resource amongst facilities stakeholders. The paper attempts to describe the findings of a study to establish the types of CAFM functionality sought by facilities management practitioners.\nDesign/methodology/approach: - The study involved a structured questionnaire study of 110 respondents in Scotland (from 395 administered questionnaires) using a stratified sample. The objective was to obtain a profile of early-adopters and to identify barriers to further adoption.\nFindings: - The findings indicate that the use of information technology in the form of CAFM is more likely amongst dedicated FM service providers than organisations with in-house FM delivery and that medium-sized enterprises (between 51 and 250 employees) are least likely to be using CAFM. A limiting factor in continued adoption appears to be the diminishing returns resulting from information overload. \"Attention\" appears to be the limiting resource in an increasingly complex and information hungry environment.\nResearch limitations/implications: - The findings are based on a national survey in Scotland and may have limited reproducibility outside of the UK. Nevertheless the general findings are relevant to the international context.\nPractical implications: - This paper provides an indicative analysis of current trends in the computer-aided facilities management (CAFM) arena. It identifies an increasing need to capture \"tacit\" as well as \"explicit\" knowledge in such systems.\nOriginality/value: - This work develops the idea of the \"attention economy\" in the context of facilities management, highlighting the current impasse for information systems aspiring to become knowledge management systems.\n",
"role": "assistant"
}
] |
Infrastructure and core quality management practices: how do they affect quality?
|
[
"Quality management",
"Quality",
"Performance management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Quality management (QM) is defined and measured in empirical studies as practices of organizations that implement principles such as customer focus, continuous improvement, and teamwork to improve product and service quality (Dean and Bowen, 1994; Prajogo and McDermott, 2005; Sousa and Voss, 2002). Much has been written on the impact of QM practices on performance. However, previous studies presented conflicting results on how different QM practices, specifically, infrastructure QM practices and core QM practices, affect performance (Sousa and Voss, 2002). A few studies reported that only the infrastructure QM practices (e.g. executive commitment, employee empowerment, and customer focus) contribute to quality improvement, but the core QM practices (e.g. information and analysis, process improvement, benchmarking, and use of advanced manufacturing technologies) do not (Dow et al., 1999; Powell, 1995; Samson and Terziovski, 1999). On the contrary, other studies found positive relationships between the core QM practices and performance (Forza and Flippini, 1998; Rahman and Bullock, 2005; Sanchez-Rodriguez and Martinez-Lorente, 2004), and some studies found that the infrastructure QM practices such as the role of top management, employee relations (Motwani et al. 1994), employee involvement, and employee selection and development (Adam et al., 1997) were not significantly related to performance.The aforementioned mixed findings could be due to three reasons. First, the categorizations of infrastructure and core QM practices vary among the studies (Ho et al., 2001; Taylor and Wright, 2006). For example, Dow et al. (1999) consider supplier relations as one of core QM practices, whereas Rahman and Bullock (2005) categorize supplier relations as an infrastructure QM practice. Second, the studies measure different levels of performance (Kaynak, 2003). Some studies like Powell (1995) measure financial performance, while others focus on quality performance as Forza and Flippini (1998) and Dow et al. (1999) do; and still others such as Samson and Terziovski (1999) operationalize performance measures to include quality performance, productivity, and employee morale. Third, the analytical methods used to investigate relationships between QM practices and performances differ (Kaynak, 2003; Sousa and Voss, 2002). The studies concluding that the infrastructure QM practices could produce performance even without the core QM practices based their conclusion on a correlation analysis (e.g. Dow et al., 1999; Powell, 1995) or multiple regression analysis (e.g. Adam et al., 1997; Samson and Terziovski, 1999). These studies did not identify the direct and indirect effects of QM practices on performance and hence are not adequate to draw any definite conclusions about the QM practices-performance relationships (Kaynak, 2003; Sousa and Voss, 2002).This study seeks to unravel the confusions in the literature by investigating the interrelationship between infrastructure and core QM practices and their effects on quality performance. We adopt Flynn et al.'s (1995) framework of infrastructure and core QM practices. This framework has been shown to effectively examine the complicated relationships between QM practices and performance (Sousa and Voss, 2002). In this study, we measure performance at the level of quality performance, as the foremost purpose of implementing QM practices in organizations is to improve product and service quality. Then, based on the Socio-Technical Systems (STS) theory and research related to QM implementation and performance, a research model is developed to test the relationship between infrastructure and core QM practices and their direct and indirect effects on quality performance. The structural equation modelling (SEM) method is utilized to analyze the model using survey data obtained from 226 manufacturing plants in the US. The SEM method is able to test the hypothesized relationships of latent constructs simultaneously (Byrne, 1998), which distinguishes the direct and indirect relationships between the constructs and thus is desirable for this study to fully explore the relationships between different QM practices and quality performance. Results of this study help to address the inconsistencies in the literature and provide a better understanding of how different types of QM practices affect quality performance.The remainder of this paper is organized as follows. The next section reviews the literature and proposes hypotheses. The following section discusses data collection procedure, sample, and data analysis methods. Finally, discussion of the findings and conclusion of this study are presented.\nLiterature review and hypotheses: Infrastructure and core quality management practices\nResearch method: Survey instrument\nDiscussion: Although the application of scientific methods in quality improvement is the most distinctive feature of QM, firms often put greater emphasis on group-process techniques and interpersonal skills than on scientific methods (Hackman and Wageman, 1995). There is doubt about whether the technique-focused core QM practices are necessary for achieving quality improvement. The findings of this study, however, provide strong evidence of the need to establish the core QM practices in quality improvement. In the structural model, the core QM factor is found to directly affect quality performance, and it completely mediates the relationship between the infrastructure QM factor and quality performance. These findings corroborate the results of prior studies such as Flynn et al. (1995), Ho et al. (2001), Kaynak (2003), and Rahman and Bullock (2005) that the core QM practices like quality information, product/service design, and process management have significant impacts on quality performance.Recently, a structured QM method, Six Sigma is gaining popularity in industry. Six Sigma emphasizes using statistical tools and the scientific method to reduce product defects and process variability (Linderman et al., 2003). By rigorously applying QM tools and techniques and Six Sigma metrics in improvement projects, companies such as Motorola, General Electric, and Honeywell have obtained substantial benefits from Six Sigma improvement efforts (Pande et al., 2000; Snee and Hoerl, 2003; Watson, 2003). The revival of the technique- and methodology-focused QM method like Six Sigma in industry indicates that QM tools and techniques cannot be ignored in the process of QM implementation, and effective application of those tools and techniques in continuous improvement has the potential to create significant benefits.QM should be implemented as an integrated approach of different practices (Flynn et al., 1995; Kaynak, 2003; Yeung et al., 2005). The interdependence nature of QM practices suggests that firms need to implement all the practices effectively in order to obtain superior business outcomes. The STS theory suggests that in an organization, people (the social subsystem) use tools, techniques and knowledge (the technical subsystem) to produce goods or services valued by customers who are part of the organization's external environment (Pasmore, 1988). The social subsystem and the technical subsystem are interdependent on each other (Ho et al., 2001). How well the social and technical subsystems are designed with respect to one another and with respect to the demands of the external environment greatly determines organizational effectiveness (Pasmore, 1988). Thus, to be effective, both subsystems must integrate and be jointly optimized (Ho et al., 2001). This study found that the infrastructure QM practices influence to a substantially large extent the implementation level of the core QM practices, which in turn directly lead to quality improvement. These findings substantiate the STS perspective. This study suggests that organizations must effectively implement both the core and infrastructure QM practices in order to achieve maximum benefit from the QM implementation. A learning and cooperative environment within the organization and with external customers and suppliers is important to promote employees' involvement in tactical improvement tasks and to provide opportunities for employees to apply QM tools and techniques in their jobs, which will transfer the knowledge and put QM principles into effect.This study is subject to several limitations. First, the data of this study were obtained through survey and relied on the perceptions of the respondents. Second, it should be noted that the literature has shown that the relationships between multiple QM practices are more complicated when they are examined as individual practices. This study grouped the practices into two integrated factors (i.e. the infrastructure QM practices and the core QM practices) because the focus of this study was on resolving the confusion about how these two types of QM practices affect performance, other than the interrelationships among the individual practices. Discussion of the relationships between individual QM practices can be found in studies such as Flynn et al. (1995), Kaynak (2003), Sila and Ebrahimpour (2005), and Yeung et al. (2005).\nConclusion: This study contributes to the QM literature by empirically investigating the roles of infrastructure QM practices and core QM practices in determining quality performance. The resulted structural model reveals the direct and indirect effects of infrastructure QM practices and core QM practices on quality performance, and thus helps to clarify the conflicting findings in the literature regarding how these two types of QM practices influence quality performance. The findings of this study have important implications to managers. This study suggests that the core QM practices are essential for quality improvement, and the infrastructure QM practices support the application of the core practices. Thus, as the managers develop and maintain their organization's QM system, adequate resources need to be allocated to both types of practices in order to achieve the effectiveness of the whole QM system.\n",
"role": "user"
},
{
"content": "Purpose: - This empirical study seeks to resolve the conflicting findings in the quality management (QM) literature about how different QM practices, specifically, infrastructure QM practices and core QM practices, affect quality performance.\nDesign/methodology/approach: - Based on the Socio-Technical Systems theory and research related to QM implementation and performance, the study proposes a research model of the relationship between infrastructure and core QM practices and their direct and indirect effects on quality performance. The empirical data were drawn from 226 manufacturing plants in the USA. The research model was tested using structural equation modelling (SEM) technique.\nFindings: - In the structural model, two integrated factors were used to represent the two types of QM practices: the infrastructure QM includes top management support, customer relationship, and supplier relationship, and workforce management; and the core QM consists of quality information, product/service design, and process management. The analysis of the structure model shows that the core QM directly leads to improved quality performance, and the infrastructure QM contributes to quality performance by supporting the core QM.\nResearch limitations/implications: - The study examines the roles of infrastructure QM practices and core QM practices in improving quality performance. It confirms that QM should be implemented as an integrated approach of different practices.\nPractical implications: - The major implication of the study is that both core and infrastructure QM practices are important in improving quality. It is important that companies allocate resources to establish both types of QM practices in order to achieve the effectiveness of the whole QM system.\nOriginality/value: - The study utilized the SEM technique to empirically investigate the direct and indirect effects of infrastructure QM practices and core QM practices on quality performance. The SEM results help to clarify the mixed findings in the literature regarding the pattern of the QM practices-performance relationships.\n",
"role": "assistant"
}
] |
Internal branding and sustainability: investigating perceptions of employees
|
[
"Sustainability",
"Internal marketing",
"Internal branding",
"Transition economy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The increasing attention toward sustainability issues arising in society is stimulating the need for managers to work more effectively with sustainability initiatives and policies within their companies. A global survey involving executives and managers across a variety of industries confirms that 70 per cent of respondents' companies have added sustainability to their managerial agendas in the past six years (Kiron et al., 2012). Despite the widespread adoption of sustainability practices, managers face difficulties in achieving strong sustainability performance (Szekely and Knirsch, 2005). The failed attempts of many companies to integrate sustainability in their business activities have resulted in skepticism among managers, in particular, about the role of such initiatives for achieving a competitive advantage (Simpson et al., 2004). Among the main obstacles to successful sustainability implementation, previous research highlights the insufficient management of internal stakeholder relationships and especially the ethical misalignment between employee perceptions and the company's corporate identity (Powell, 2011).\nTheoretical framework: Internal branding and internal brand commitment\nMethodology: The study was conducted among employees of industrial companies operating in Belarus, which collaborated in a trainee program developed by the Belarusian State Technological University. The data were collected in 2014. The questionnaire consisted of the four parts focusing on the background of the respondents, internal branding, corporate sustainability and work environment. Scales from prior research were used for measuring internal brand commitment and its three determinants, namely, brand orientation, internal brand knowledge and internal brand involvement (Baumgarth and Schmidt, 2010).\nAnalysis and results: The majority of the respondents worked for companies in the forestry (34.0 per cent) and construction (30.3 per cent) industries. The sample also included employees from printing (16.8 per cent), chemical (15.5 per cent), cosmetics (1.7 per cent) and food (1.7 per cent) sectors. The respondents held following positions in their companies: CEO (2.1 per cent), head of department (6.9 per cent), economist (46.6 per cent), technologist (18.9 per cent), engineer (15.9 per cent) and other specialists (9.6 per cent). The majority of the respondents had worked for their current employer for less than 10 years: less than 5 years (34.8 per cent), 6-10 years (27.1 per cent), 11-15 years (19,5 per cent) and more than 16 years (18.6 per cent).\nDiscussion and conclusions: The findings of this study demonstrate the relevance of internal branding for enhancing the awareness of employees about sustainability importance and facilitating their positive evaluations of the company's sustainability performance. By drawing on the results of prior research on brand citizenship behavior (Baumgarth and Schmidt, 2010; Burmann and Zeplin, 2005; Burmann et al., 2009), the study shows that effective internal branding strategies can potentially support the successful implementation of sustainability programs within the company. In line with the findings of sustainability research applying the social identity theory (Farooq et al., 2014; Kim et al., 2010; Lamm et al., 2015; Turker, 2009), this study highlights that successfully implemented sustainability programs can reinforce the perceived organizational identification of employees and increase their internal brand commitment. The results provide empirical evidence that employees, who are more knowledgeable, involved and committed to their corporate brand, also perceive sustainability importance and sustainability performance as being high for their companies.\nManagerial implications: The outcomes of the study encourage marketing managers to utilize internal branding for supporting the implementation of certain policies and programs within their companies, especially those focusing on CSR and sustainability. The internal communication and activities should be focused on providing detailed information about these initiatives and clarifying their importance for the company. These efforts will decrease the lack of knowledge among employees and possibly reduce the misalignment between their perceptions, corporate values and performance. Also, the high relevance of certain initiatives for employees can strengthen their identification with the company and its corporate brand, increase their commitment and motivate them to more supportive behavior. We advise marketing managers to utilize sustainability initiatives and programs for enhancing the strength of their brand within the company.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate perceptual differences between employees characterized by various levels of internal brand commitment and sustainability importance. The identified clusters of employees are compared considering the main determinants of internal brand commitment and their evaluations of a company's sustainability performance.\nDesign/methodology/approach: The study was conducted among employees of industrial companies operating in Belarus. Two hundred thirty-eight responses were analyzed by using the K-means cluster analysis. The analysis of variance was applied for evaluating differences between the three identified clusters.\nFindings: Based on internal brand commitment and sustainability importance, the paper identifies three alternative clusters of employees: uncommitted sustainability laggards, committed sustainability followers and committed sustainability leaders. The results show that the three clusters differ in regard to the main determinants of internal brand commitment, which are brand orientation, internal brand knowledge and internal brand involvement. The findings demonstrate significant perceptual differences between the three clusters regarding their evaluations of a company's sustainability performance. The assessment of sustainability performance focuses on sustainability objectives, sustainability policies and sustainability decision-making and disclosure.\nResearch limitations/implications: The study was conducted in the context of a transition economy. Future studies are recommended to develop a cross-cultural comparison of relationships between employees' perceptions about sustainability performance, different determinants and the outcomes of internal brand commitment.\nOriginality/value: The paper makes a theoretical contribution to research on internal branding and sustainability by examining the potential interrelationships between internal branding strategies and sustainability performance.\n",
"role": "assistant"
}
] |
Social policies and R&D subsidies: Impact on inequality, unemployment, growth and the tax burden
|
[
"Economic growth",
"Pay",
"Research and development",
"Unemployment",
"Subsidies"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Since the early 1980s, Europe and the USA have experienced rather diverging developments in terms of inequality, unemployment and growth. In the mid-1990s, the \"transatlantic consensus\" (Atkinson, 2001) considered that the higher demand for skilled labour induced by skill-biased technological change[1] (SBTC) resulted in an \"inequality versus unemployment\" trade off. By maintaining social protection, Europe had opted for equality at the expense of employment, whereas the USA had chosen full employment and let inequality rise (Krugman, 1994). However, such an explanation falls far short of encompassing every experience that the different countries have had. Within Europe, low unemployment and higher growth characterise both non-egalitarian (the UK, Ireland) and highly egalitarian (Sweden, Denmark) countries, whereas higher unemployment and lower growth may be observed in the countries in between (Belgium, France, Germany, Italy).This paper analyses the impact of different public policies on inequality, unemployment, growth and the tax burden. A model is constructed where growth is driven by endogenous technical progress. So as to comply with the SBTC explanation for growing inequality, the characteristics of technical progress are tailored to represent computerisation. We introduce and combine several policies (minimum wage, redistribution, R&D subsidies) that have an impact on both inequality and the rhythm of technological change. The public funds provided for redistribution, subsidies and unemployment compensation are financed by an income tax. Eight scenarios (policy patterns) are then analysed and simulated.We show that all policies except pure redistribution are better than non-intervention (pure competition) in terms of growth. In particular, while creating unemployment, the setting of a minimum wage also boosts technical progress and growth, unlike a redistribution policy. In the long term, it thereby improves welfare. We then distinguish three major policy patterns. The Anglo-Saxon model is characterised by high growth, high inequality, low unemployment and a low tax burden. The Nordic model combines high growth, low inequality and low unemployment, and a high tax burden. The Continental European model puts together medium inequality and a medium tax burden, a slowdown in short-term growth, and higher long-term growth paid for by high unemployment.The general framework of the model is constructed in section 2. Section 3 defines the growth path in different configurations (policy patterns) and analyses its main features. In section 4, calibration exercises are implemented. The final section discusses the model's main findings and defines the three stylised policy patterns.\n2. The model: There are three factors of production: skilled labour (H) and unskilled labour (L) are exogenously supplied, whereas the capital good (K) is produced. The wage of unskilled labour is equal to unity. As a consequence, the price (wage) of skilled labour, denoted w, is also the skill premium, and the price of capital, denoted r, is also a relative price.There is one final good that is produced on a perfectly competitive market.The capital good (K) is produced by a monopoly on a contestable market. This monopoly also carries out an R&D activity that aims at increasing total factor productivity in the production of capital.2.1 The final goods industry\n3. General equilibrium: At each time, the model determines one full employment equilibrium characterised by the skill premium, the productions and prices of capital and of final goods, the distribution of factors across activities, the real wages of skilled and unskilled workers, the nominal and real total income, the amount of researchers HA, and the rate of growth of TFP in sector K, a.. The latter determines the model dynamics, each period being characterised by an increase in technological knowledge, and therefore in TFP in the capital good sector and in the production of the final good.3.1 Equilibrium without public policy\n4. Simulations: We now implement numerical simulations of the different public policies described above so as to analyse their impact on inequality, unemployment, growth and the tax burden. Two indicators of the tax burden are described. The first is tax rate t, which measures the weight of levies for skilled workers. The second is total tax rate t=T/(L+wH-), i.e. the ratio of total collected taxes on total income.We shall first define the eight analysed scenarios, namely non-intervention and seven different policies. The parameters used for calibration will then be introduced. Finally, the main results of the different scenarios are highlighted.4.2 Eight scenarios\n5. Discussion: We first compare the different scenarios according to their outcomes and we give some indications about their impact in terms of welfare. We finally make a distinction between three policy frameworks, namely an Anglo-Saxon model, a Nordic model, and a Continental European model, which are characterised by different features in terms of the tools utilised and outcomes obtained.5.1 Comparisons\n6. Conclusion: The model developed in this paper aims at analysing the impacts of several public policies on inequality, unemployment, growth and the tax burden. Except for pure redistribution, all policies result in higher growth than no intervention. It is the case of the setting of a minimum wage which boosts R&D and growth because it lowers the cost of skilled labour, and is thus better for growth than a pure redistributive policy. Moreover, the model leads to the distinction of three configurations which capture the mains features of the developments of Anglo-Saxon countries, Scandinavian countries, and continental European countries in the 1990s. It thereby provides a general framework to analyse and compare these experiences.\n",
"role": "user"
},
{
"content": "Purpose: - The paper seeks to analyse the impact of different public policies on inequality, unemployment, growth and the tax burden.\nDesign/methodology/approach: - A dynamic general equilibrium model is built, in which growth is driven by endogenous technical progress, to analyse the impacts of several policies (minimum wage, redistribution and R&D subsidies financed by an income tax).\nFindings: - All policies except pure redistribution are better than non-intervention in terms of growth. The authors distinguish three major policy patterns. The Anglo-Saxon model is characterised by high growth, high inequality, low unemployment and a low tax burden. The Nordic model combines high growth, low inequality and low unemployment, and a high tax burden. The Continental European model puts together medium inequality and a medium tax burden, and higher long-term growth is paid for by high unemployment.\nResearch limitations/implications: - The model could be extended by the introduction of educational policy.\nOriginality/value: - The paper distinguishes three configurations that capture the main features of the developments in Anglo-Saxon countries, Scandinavian countries, and Continental European countries in the 1990s. It thereby provides a general framework to analyse and compare these experiences.\n",
"role": "assistant"
}
] |
Gift exchange: benefits sought by the recipient
|
[
"Gifts",
"Benefits"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Billions of dollars are spent on gifts each year (Ruth et al., 1999), and gift exchange forms an important social bond between individuals (Belk, 1979; Parsons, 2002a). Many of the studies in this area have attempted to make the understanding of gift exchange practical to the retailer or marketer, and past research has linked the product benefits gained in the exchange to the associated benefits of brands (Andrus et al., 1986; Laroche et al., 2000; Parsons, 2002a), as brands are often the focus for the choice of gifts because of the associated meanings (Heeler et al., 1978; Andrus et al., 1986; Keller, 1993; Low and Lamb, 2000).The anxiety with which gifts are given has been acknowledged in the literature (Wooten, 2000), with most gift-giving literature focusing on the gift giver's choices (Caron and Ward, 1975; Otnes et al., 1993), messages they wish to attach to gift giving (Cheal, 1987; Wolfinbarger, 1990), and the exchange frameworks in which gifts are given and consumed (Sherry, 1983; Giesler, 2006). To date, however, the recipient's evaluation of their gifts has not received much attention.However, with the exception of some examination of recipients who are difficult to select a gift for (Otnes et al., 1993), there have been no direct studies of the gift recipient's preferences in terms of the benefits sought from the gift they receive. Indeed, Flynn and Adams (2009) show that there is considerable doubt as to whether gift givers can draw on their personal experience as gift recipients in order to identify meaningful gifts for others. Given this gap in the literature, the purpose of this study is to examine what, if any, are the preferred benefits sought by a gift recipient? To achieve this aim, this study uses the Parsons (2002a) framework by combining Keller's (1993) brand associated benefits with Belk's (1979) characteristics of gifts to examine the symbolic, functional and experiential benefits of gifts for gift recipients. These three benefits will be explained and discussed in the following section. Moreover, acknowledging that alternative mediators may explain differences in recipient preferences, this study also examines how the importance of these benefits may differ depending upon the characteristics of the gift recipient and the nature of their relationship with the giver.Before moving into our discussion of gifts and the exchange process, it is worth noting that we are viewing the gift as something acknowledged and recognized by both the giver and the recipient, in an exchange of meaning (discussed below). However, there are contexts argued where altruistic giving, the radical separation of gift and counter gift or the elimination of reciprocity, or even the over gifting which results in a breakdown of the exchange, can occur. Derrida (1992) has perhaps clarified this best in stating that a genuine gift requires anonymity of the giver such that there is no accrued benefit in giving. This utopian position is an extreme, which Strehle catalogues alongside those of Marx, Bourdieu, and Caille, in his discussion of the sociology of gift (Strehle, 2009), and one that is difficult to envisage in a Western culture. Komter (2007) argues that human gift giving is too layered and too complex to be incarcerated in either the utilitarian or anti-utilitarian approach encapsulated by these arguments. Nevertheless, Metcalfe and Game (2008) put forward a compelling argument that meetings held in good grace can impart a gift that is closest to this pure gift of Derrida (1992), by allowing one to have a gift without subjects and without exchange. Others discuss the gift as part of the social order, from the sacrificial connection with god(s) (Evans-Pritchard, 1954) through to the role of the gift in marriage ceremonies (Evans-Pritchard, 1948), with the perspective that gifting is a total social phenomenon (Mauss, 1954; Komter, 2007). Our conceptualization of the gift, though, is the widely used sociological exchange view, involving in our case the exchange of brand benefits.1.1 Gifts\n2 Methodology: We propose that gift recipients will distinguish between the combinations of benefit levels sought from different gift givers, because they are expecting benefits that best match the interaction intended for the relationship. We characterize the gift giver by the sociological type of relationship, the length of the relationship held with the gift giver, and the strength of the relationship held with the gift giver. We recognize, however, that earlier studies of gift giving raise the possibility that the demographic characteristics of the recipient may influence the observed differences in benefits sought, so we also examine the proposition that any observed differences are not due to giver differences, but are due to recipient differences, specifically:* age;* gender; and* income.Data were collected as part of a larger survey of consumption behavior. Using a simple random sampling approach, 250 respondents from Auckland, New Zealand, were interviewed by telephone. A total of 623 telephone calls were made, giving a total response rate of 40.1 per cent. Household respondents were selected as the person over 18 who last had a birthday and were present in the house. A total of 13 scale items were used to measure the likely benefits sought when receiving a gift. These were derived from Keller's (1993) brand associated benefits and Belk's (1979) characteristics of gifts, as discussed previously. These benefits were used to determine the symbolic, functional, and experiential benefits of gifts (Park et al., 1986). The items were: symbolic (fashionable, exclusive, well known, conservative, traditional, recognized, and prestigious); experiential (fun, unusual, and exciting); and functional (useful, functional, and features). Each item was given equal weighting, with the respondent being asked questions like \"When receiving a gift from my close relative, I should get one that is more fashionable than I would normally receive as a gift\", with responses to each statement being made using a five-point Likert scale that was anchored \"strongly agree\" to \"strongly disagree\". Thus, a respondent will not be stating that they want a functional gift from this person, or a symbolic gift from that person; rather they will be revealing a desired level for each benefit when receiving a gift from the identified giver. Equal weighting was used because it is the relative level of functional/symbolic/experiential benefit within the gift, rather than between gifts that is desired.Respondents were asked to think of one of the followings as the person giving them a gift (the proportion selecting that person is shown in parentheses): close relative (15.6 percent), distant relative (22.8 percent), partner (13.6 percent), friend (16.0 percent), colleague (18.0 percent), and acquaintance (14.0 percent). The definitions for each type of giver were given to the respondent, have previously been used in gift research (Parsons, 2002a; Saad and Gill, 2003) and are derived from the sociology literature (Caplow, 1982). Respondents were also asked how strong their relationship with the giver was (using a ten-point scale that was anchored \"weak\" to \"strong\"), and how long their relationship with the giver had existed (in years, with a one recorded for anything up to and including one year). This strength and length of the relationship also effectively allows for a hierarchy effect if there is one present (e.g. a colleague you have known and worked with for a long time versus a recent colleague with whom you have not worked closely).Finally, to examine the alternative proposition that recipient demographic characteristics were the reason for any observed differences in gift benefits sought, data were collected on the age, gender, and household income for each respondent.Given the above, the three main relationship variables included in this study were: relationship type, relationship strength, and the length of the relationship between the giver/recipient. The three alternate relationship variables were recipient age, recipient gender, and recipient household income. The three main benefit variables were: symbolic, functional, and experiential benefits. The analysis undertaken to examine the associations between variables is discussed alongside the results in the next section.The theoretical perspective taken in this research is positivist. In this, the knowledge collected here is then objectively interpreted to discover the meaning already inherent in the data. That meaning is seen as being independent of the researchers and the research process (Crotty, 1998), and in as much, the findings are seen as stable and verifiable (Neuman, 1997). The study is also quantitative in nature for two reasons; first, we want to be able to look at the gift recipient in terms of the brand associations derived from Keller (1993) using Belk's (1979) characteristics, using validated measures, which were available from the Parsons (2002a) and Saad and Gill (2003) gift studies and the Caplow (1982) sociology study, each of which follows a robust quantitative approach that allows us to establish the existence of relationships; and second, because the aim of the research is to examine and identify quantitatively measureable associations between the combination/level of gift benefits preferred by the recipient and the nature of the relationship between the gift giver and recipient.It is necessary to understand that we are looking at this gift process through the lens of a business or economic context. We believe that commercially derived or facilitated exchange forms a major component of much modern social interaction, whether through gifts (Parsons, 2002a, and this present study); identity and self-expression to others through brands and consumption (Holt, 2002; Kozinets, 2001); physical-social interaction as motivations of shopping (Tauber, 1972; Parsons, 2002b), or even the exposure to marketing communications through various media.Prior to examining the association between the nature of the relationship between the gift giver and recipient, and the level of benefits preferred by the recipient of a gift, the age, income, and gender of the recipient was explored. Otnes et al. (1993) suggest that age may be a factor in gift choice, while Garner and Wagner (1991) raise the possibility of household income influencing gift choice. Research also suggests that gender may play a part in the gift purchased (Caplow, 1982; Webster and Nottingham, 2000), though this may be driven more by context than gender. Thus, it seemed reasonable that the same variables may affect recipients. Age was measured using three age categories (under 30, 30-60, and over 60) - note that the under 30 category from Parsons (2002a) was actually 18-30 for this study; these being the broad market segments towards which gifts are often targeted (Parsons, 2002a). Household income was also measured using three categories (under $30,000, $30,000-60,000, and over $60,000). Sex was simply \"male\" or \"female\".\n3 Results: A one-way, between-groups MANOVA was performed to investigate any age, income, and gender differences in the gift benefits preferred by the receivers of gifts. There were no significant differences between under 30, 30-60 and over 60-year olds (F=1.51, Wilks' l=0.96), between gift recipients with incomes under $30,000, between $30,000-60,000 and over $60,000 (F=0.91, Wilks' l=0.98), nor between males and females (F=1.69, Wilks' l=0.98).These findings are particularly noteworthy given that past research has shown both the product type and benefit levels chosen by the gift giver are significantly influenced by each of these characteristics. This might be due to the nature of gift giving versus gift receiving. For example, income is likely to have a greater effect on the type of gift a giver can afford to buy, but much less effect on what a receiver might prefer to receive because they are not paying for it. Age may affect the giver through experience, altering their selections, whereas age may act as more of a constant for the recipient groups. For example, as someone grows older, a gift giver may become more adept at choosing the appropriate gift, whatever the relationship. However, a gift recipient may maintain preferences based upon the relationship, even as they age.3.1 Relationship between gift benefits and relationship type\n4 Discussion, limitations, and future research: Applying the above findings, this study shows people prefer gifts with a greater symbolic meaning from those that are close to them. Anecdotally, this occurs when jewellery is given to a spouse, a car is given to a child for their 16th birthday, or a \"world's best dad\" mug/t-shirt is given. The symbolism of these types of gifts is well known and associated with love and commitment (Cheal, 1987; Otnes et al., 1994), coming of age (Schwartz, 1967) and recognition (Wolfinbarger, 1990). Second, this study finds that people who have not known each other for very long, but have a strong emotional connection prefer gifts that are primarily functional. This may speak to the level of embarrassment that can occur if the symbolism present in a gift is misinterpreted or the experience is unwanted (Wooten, 2000). This could weaken or even sever the relationship (Ruth et al., 1999). Because of the relatively short time of the relationship, shared interpretive norms have not been established (Scott, 1994) and so \"safer\" gift options are chosen. This would be the case when golf accessories are brought for a new boyfriend because the giver knows they play golf as opposed to a ring being brought for the boyfriend. If a ring was bought and the boyfriend misinterpreted it as overcommitment to the relationship, there is potential for the relationship to be weakened or broken.4.1 Theoretical development\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to investigate the recipient side of gift exchange by establishing the combination/level of gift benefits preferred by the recipient. It investigates the association between these benefits and the nature of the relationship between the gift giver and recipient.\nDesign/methodology/approach: - A total of 250 people were surveyed about a recent gift receipt experience. Levels of symbolic, experiential, and functional benefits sought from the gift were established. Canonical correlation was used to analyse the interrelationships of relational variables (relationship type, strength, and length) and the preferred gift benefits.\nFindings: - People prefer gifts with a greater symbolic meaning (see with lower levels of functional and experiential benefits) from people close to them. People who have not known each other for long, but have a strong emotional connection, prefer gifts that are primarily functional. It shows that benefit associations are significant to gift recipients, the type of relationship the recipient is in has a significant effect on the desired combined level of benefit associations, and the symbolism of meaning for gifts can be expressed through benefit associations.\nOriginality/value: - Studies of gift exchange focus on the gift giver and the benefits gained from the act of giving. The results show that the exchange process used by sociologists and consumer behaviorists to describe and explain gift-giving activity cannot assume the same levels of benefits associated with the gift occurring on each side of the exchange. It is also shown for the first time that the type of giver, and the relationship the recipient has with the giver, will modify recipient preferences.\n",
"role": "assistant"
}
] |
The effects of authentic leadership and organizational commitment on turnover intention
|
[
"Turnover intention",
"Organizational commitment",
"Authentic leadership",
"Authentic leadership development"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Authentic leadership is defined as behavior that displays and encourages positive mental dimensions and a principled climate that cultivates self-awareness (SA), an internalized moral perspective (IMP), balanced processing of information (BPI), and relational transparency (Walumbwa et al., 2008). Authentic leadership research has conceptually and empirically evolved as a focus of interest that enriches studies on organizational commitment (OC) (Avolio et al., 2004; Jensen and Luthans, 2006). Previous research also indicates authentic leadership is related to critical organizational outcomes, such as empowerment, job satisfaction, job performance, and overall profitability (Avolio, 2010; Clapp-Smith et al., 2009; Giallonardo et al., 2010; Walumbwa et al., 2008, 2010).\nLiterature review: Theoretical foundations\nMethods: Sample and data collection\nResults: Structure equation modeling in EQS statistical Software was used to test the hypotheses of this study. As recommended by Anderson and Gerbing (1988), the measurement model was assessed using confirmatory factor analysis (CFA) first. After examining the reliability, discriminant validity, and convergent validity of the model constructs, a second-order factor structure that contains two layers of latent constructs was conducted (Hair et al., 2006) in which the higher order latent variable, AL, cause lower order latent variables RT, IMP, BPI, and SA. The hypothesized structural model was then tested and parameters were estimated. To test the model fit, several fit indexes were used. Hu and Bentler (1999) suggest a comparative fit index (CFI) value of 0.95 or higher and a root mean square error of approximation (RMSEA) value of 0.06 or less as indicators of adequate model fit. Generally, kh2-to-df ratio less than 2, CFI value of 0.90 or higher, and RMSEA value at or below 0.08 demonstrate good model fit (Browne and Cudeck, 1989).\nDiscussion: The results of this study suggest a variety of significant theoretical contributions as well as critical leadership and organizational implications. This study was an attempt to understand and empirically test hypothesized effects of authentic leadership on employees' OC and the effects of that OC on TI. Our model proposes that the effect of authentic leadership on TI is mediated by OC.\nImplications: Research implications\nLimitations and future studies: There are some limitations of the present study. The first limitation of this research is convenience sampling. Convenience sampling is a sampling method used to select a naturally occurring group within a population. It is possible that this convenience sampling influenced the sample profile as 75 percent of the respondents with less than two years or less of tenure. We expected such a relatively low tenure rate based on the US Bureau of Labor (2014) that reported the median years of employee tenure in the accommodation and food services industry was 2.1 years, the lowest of all industry categories in 2014. To more accurately represent the profile of target population, future studies should consider restating this proposed model by random sampling employees with large organizations or industry associations. Second, same-source participants from our study rated the antecedent, mediating, and outcome variables in our study. Same-source data can be a possible limitation since they present the chance that these results can be attributed to common-method variance (Podsakoff et al., 2003). Common-method bias can be abridged by implementing several preemptive strategies (such as collecting data from different sources) (Lindell and Whitney 2001; Podsakoff et al., 2003). Therefore, future research should recruit subjects from different sources or at different periods of time to reexamine the relationships presented in the current study.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore whether authentic leadership in hospitality is composed of four distinctive but related substantive components (i.e. self-awareness, relational transparency, balanced processing, and internalized moral); the impact of authentic leadership on employees' organizational commitment (OC); the impact of employees' OC on their turnover intention (TI); and the indirect effect of authentic leadership on employees' TI via OC.\nDesign/methodology/approach: The authors tested a sample of 236 students working as employees in hospitality in the USA, on the idea that authentic leadership increases OC which in turn decreases TI. The participants were asked to rate the manager's leadership style and the frequency of their leadership behavior.\nFindings: Results of structural equation modeling provide support for the positive effect of authentic leadership on OC in the hospitality industry, and suggest that OC mediates reduced TI.\nPractical implications: The findings in the present study are extremely useful to managers, human resource managers, and organizations as a whole. Practitioners looking to increase employee OC and decrease TI can do so by augmenting the authentic leadership qualities of managers.\nOriginality/value: The results of this study suggests a variety of significant theoretical contributions as well as critical leadership and organizational implications. The effects of authentic leadership were empirically tested on employees' OC and the effects of that OC on TI.\n",
"role": "assistant"
}
] |
The Emperor's new tweets
|
[
"Social media",
"Twitter",
"Facebook",
"CEOs",
"Personality",
"Metrics",
"Frequency",
"Passion",
"Perspective",
"Senior executives"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: By turns exhilarated, exhausted and confused, marketers and corporate communicators have wrestled for a number of years with the risks and opportunities represented by the digital universe in general and social media in particular. Brands, we are told, are in constant danger of being blown up by suddenly visible and angry consumers buying companyxsucks.com URLs and launching derisory hashtags on Twitter to excoriate the latest target of the public's seemingly unlimited supply of venom. A recent column by Andrew Hill, writing in the Financial Times' Special Report on \"Global Brands\" (\nWhich platforms: the Willie Sutton principle: Although he never actually said it, legendary US bank robber Willie Sutton will forever be associated with the remark a reporter attributed to him about why he robbed banks, \"it's where the money is\" (Sutton, 1974). The Willie Sutton principle applies equally well to the issue of where business executives should be seeking to create online communities. It should be wherever his or her target audience likes to spend time. The answer to this question will depend in large part on who the target audience is. For one type of CEO, that target is investors; for another, it might be customers and prospects in the tech space; and for yet another, potential recruits in high demand categories. Some CEOs might even want to reach existing employees in public forums. It all depends on the individual company's circumstances and business strategy.\nDesired behaviors: The second critical question is what kind of behaviors, either in the online context or in the real world, are we trying to promote? Some senior executives believe or claim they believe that they are polishing the reputation of their organization by their activities in social media. While this may be true in certain circumstances, we believe that it is important to be much more specific about the desired behavior change in the target audience to justify the time and resource investment. We think that the important questions to ask are:\nThe content: Here the subject of inquiry becomes somewhat perilous because social media place a distinct burden not only on content but also on tone. Freed of the ballast of conventional forms such as the keynote speech, the white paper and the press release, executive social media content can sometimes seem lost or out of joint. If social media are, as we have been told, a critical forum for leaders to exhibit authenticity, where does that leave individual executives whose management style is more oracular than intimate? The result, too often, is a strange lurching between the high and low - a blog post about trends in the supply chain, followed by a tweeted picture of the family at the beach. It is also important to stick to age- and role-appropriate language and, it should not need pointing out, icons and acronyms. Real-world rules apply here, but it is sometimes surprising to see LOLs in posts from CEOs who would never dream of opening the annual shareholder meeting with the word \"wassup?\"\nFrequency and timing: The question of frequency is perhaps the easiest to answer, as the different social media platforms have established consumption cultures of their own that, to some degree, dictate the optimal frequency of content refreshment. On Twitter, for example, the brevity of the medium itself (140 characters) and the way that readers interact with it has driven brand users to multiple intra-day posts. Citing analysis by SocialTrack, Kevan Lee, writing in Fast Company, points out that rates of engagement on Twitter correlate strongly with the total volume of Tweets, with spikes of increasing engagement appearing along the volume curve (Lee, 2014). While we believe that very few senior executives would want to establish a frequency of posts this high, we can assume that Tweets that only appeared once a month would make it difficult for the author to establish credibility in this medium. Each of the other main platforms, Facebook, LinkedIn, Google+, among others, has a rhythm that seems to work best. The most important point about frequency, however, is that there should consistency and predictability, whether that is three times a day or once a month.\nPutting the plan into practice: Having established answers or at least theories to test answers to the preceding questions, it becomes time to actual put the CEO's social media plan into action. Most commentators suggest that it is important to start by creating a bank of potential posts and content that will cover two to -three months of social media outings. Each executive will need to determine how much of that content should literally originate with him or her and how much can be written by others and approved. Each individual leader, planning a social media strategy, should also create a personal guide for their posts to ensure that they do not inadvertently divulge confidential or material information. Much of this is common sense for senior executives, but there are some pitfalls unique to social media. Many smart devices that can take photographs or shoot videos have geo-tagging enabled. Sharing a photo in a blog post or Tweet could provide a competitor with information about the executive's whereabouts, unintentionally providing clues to a secret business meeting such as one involving potential merger discussions. Whatever the individual guidelines are they should provide sufficient protection against faux pas without handcuffing the content so completely that it renders the purpose of a social media presence useless.\nConclusion: In conclusion, we believe that CEOs and senior executives can use social media effectively, as long as they have a clearly defined purpose and have established ways of measuring whether that purpose is being fulfilled using social media. It is critical that those metrics do not simply measure the online activity but reach out to perception and behavior change that can be directly measured, whether that be investor activity or new business opportunities. On a final note, we believe that CEOs and other senior executives have a wide range of communications vehicles open to them perhaps more than any other single group of individuals. For us, this means that once the itch to be au courant by having a social presence has faded and competitive one-upmanship regarding social media between CEOs ceases to be a factor, the choice to be social or not social in media will once again become a matter of individual taste and preference.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to identify decision tools for determining how to structure a social media presence for senior executives.\nDesign/methodology/approach: The desired decision tools emerge from a review of the opportunities and challenges facing senior executives in social media.\nFindings: By following a structured inquiry into the business purposes of an executive's social media presence, it is possible to determine what that presence should consist of.\nOriginality/value: In the current literature, there are few attempts to establish practical criteria for social activities by senior executives and how to measure its value.\n",
"role": "assistant"
}
] |
Optimal promotional strategy for intra-category cross-selling: An application to culinary products in Taiwan
|
[
"Budget allocation",
"Fast-moving consumer goods",
"Promotional strategy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Fundamental changes are continuing to take place in the \"fast-moving consumer goods\" (FMCG) industry, as both retailers and manufacturers begin to embrace the current \"category management\" approach to product assortment processing (Basuroy et al., 2001; Du et al., 2007). Pondering the wide variety of shopping options, and recognizing extreme price competition amongst retailers, today's shopper is increasingly likely to discriminately seek out products amongst various sources, and amongst various products and brand extensions. Furthermore, with the natural maturing and lapsing of the market towards the declining stage of the life cycle of any product, manufacturers and retailers are faced with the need to improve revenue levels, when at the same time, they also need to invest heavily in pricing, promotion and merchandising activities.\nMethodology: Data collections and sample\nResults: The results of this study on promotional frequency, with regard to different sales promotion vehicles and packages, are summarized in Table II. If=0, then this would imply that the previous promotional investment strategy had led to virtually no increase in sales revenue.\nConclusion: In the fiercely competitive fast-moving consumer goods (FMCG) market, the effects of sales promotions are far too complex to explain, and indeed, such promotions can no longer guarantee stimulation of sales. Furthermore, manufacturers have difficulties making promotional decisions, a challenge which leads to the effects of promotions appearing uncertain, yet promotional activities continue to receive a considerable share of promotional budgets. We have analyzed the optimal sales promotional strategy for intra-category cross-selling, which includes budget allocation, promotional frequency and the evaluation of promotional performance. Interestingly, our results reveal that neither sales promotions nor the moderating effect exhibit any real measure of success.\n",
"role": "user"
},
{
"content": "Purpose: The aim of this paper is to investigate an optimal promotional strategy of intra-category cross-selling on culinary products for the fiercely competitive, fast-moving consumer goods (FMCG) industry.\nDesign/methodology/approach: A linear regression model and a Markov switching autoregressive model is used, that incorporates a retailing demand process to capture a nonlinear structure among promotional budget allocation, and evaluate promotional performance, and optimal promotional frequency within a given time span. Three product categories are applied with 39 months of time-series data from a multinational packaged food company in Taiwan.\nFindings: The result shows that most previous decisions on promotional budget allocation are non-optimal - most promotional investments were either extended too long or allocated too low in stimulating sales.\nResearch limitations/implications: This study suggests implications for the brand or category manager in removing such non-optimal promotional policies.\nOriginality/value: Previous promotional investment is evaluated by comparing the changes in promotional budget allocation. Markov's switching feedback rules are then applied to determine the proper length of equilibrium state with and/or without promotion. Finally, effective decision rules on magnitude, duration, and frequency of intra-category promotional strategy are induced.\n",
"role": "assistant"
}
] |
Preparedness: the state of the art and future prospects
|
[
"Resilience",
"State of preparedness",
"Emergency planning",
"Emergency measures",
"Disasters",
"Business continuity",
"Resilience engineering"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Since the start of the twenty-first century, events described as disasters have strengthened our attention toward communities' state of preparedness. Whether considering the terrorist attacks of September 11, 2001, the 2004 tsunami, Hurricane Katrina or the Haitian earthquake, to name just a few, these events have raised concerns about communities' response capacities and, more generally, our state of preparedness. At its Hyogo Conference, the United Nations (UN) formulated a strategy to make our communities more resilient.This framework, combined with the many studies of preparedness (in the broadest sense) undertaken by different countries, leads us to examine the development of the concept and the new perspectives that resilience has introduced into the field of disaster management. In recent years, the trend seems to be to approach the concept of preparedness in two very different, but complementary, directions. The first sees preparedness from the perspective of response capability. The second integrates management measures designed to maintain operational continuity.This article presents the changes in the concept of preparedness in the context of emergency measures and describes the new needs that have appeared in recent years. Preparedness should no longer be seen as just the capacity to respond to an event but as anticipating the different possible ways of addressing the event with a resilient perspective. First, several definitions of preparedness will be presented, followed by a discussion of the two different approaches to emergency preparedness that were mentioned earlier in this paper. Finally, the concept of resilience engineering will be discussed. The concepts presented in this article are part of an all-hazards approach focussed on consequence management (Robert et al., 2007). This is directly related to the work of the authorities responsible for emergency measures management that need to control the consequences of various events, mainly impact on population.\n2. Preparedness: changing definitions: 2.1 Brief background\n3. Method for evaluating preparedness[1]: Stakeholders have been evaluating preparedness for some years now. The first evaluations were done by means of exercises and simulations, as well as operating experience feedback. This methodology, which came from the military, is still widely used today (Conseil pour la reduction des accidents industriels majeurs (CRAIM), 2007). Simulations (or exercises) allow stakeholders to be immersed in a hypothetical situation (but one that is as close as possible to reality), so their reactions and the outcomes of the various actions that are taken can be analyzed. This procedure gives users an operational vision of preparation, but it does not provide any indicators enabling managers to regularly evaluate the preparedness of their companies. As well, since preparedness does not depend on operating staff alone, indicators that allow one to combine all functions present in a company have become necessary to expand the concept.Thereafter, the various scientific studies presented above made it possible to integrate the concept of management of preparedness into the evaluation of that state. With his criteria for disaster planning and management, Quarantelli (1998) opened the door to numerous studies, including those by David Simpson.The scientific literature contains numerous indicators (see Table III for some examples) that measure various aspects of preparedness, as well as methodologies for accumulating everything into an index. These indicators focus both on intervention and on planning. Nevertheless, there do not seem to be any universally accepted indicators or methodologies that would discriminate among indicators and allow the most appropriate ones to be chosen. Each country and organization wanting to measure its state of preparedness adapt the different indicators and methodologies available to their context. Nothing suggests that a method and universal indicators are desirable. The adaptation that an organization can undergo allows a better integration of the concept. The context also assumes that the environment in which the state of preparedness is measured is not the same for all (different vulnerabilities and different contingencies). Despite this, we can generalize a common indicator and, subsequently, a specific indicator, for certain hazards or certain type of consequences.Moreover, such indicators are no guarantee of the success of an intervention. A process of reflection must be initiated to find so-called performance indicators that will actually allow causes to be linked to effects.Simpson (2004) proposes a methodology for evaluating a municipality's (local administration) preparedness for disasters. This methodology includes an extremely exhaustive set of indicators that measure the preparatory means that have been put in place (see Table III for some indicators example). The methodology is then refined in his Working Paper No. 06-03 (Simpson, 2006), leading to the creation of a disaster preparedness index. Simpson eventually published an article that presents his work as a whole and enables his methodology to be implemented (Simpson, 2008). This research was conducted for two communities in the USA (Sikeston, MO and Carbondale, IL). These indicators were developed to assist decision makers in assessing their state of preparedness.O'Leary (2004) presents a practical guide for developing indicators to measure disaster preparation. This guide first explains how communities prepare for disasters, then sets out a technical framework for indicators and performance measurement. The guide follows in Simpson's footsteps, in the sense that it proposes a statistical approach to combine a set of indicators into an index. The theoretical framework presented in this work may be applicable in other fields since it is focussed on indicators in general.McEntire and Myers (2004) approach preparedness from the angle of a process. They define ten steps needed to achieve a state of maximum preparedness. Their approach is similar to that of Quarantelli (1998) because they propose not a direct measure but a work methodology intended to foster enhanced preparedness. It is meant to fit into a broader preparation program, necessitating leadership, professionalism, effective public communications and interagency links, all of which depend on modern technologies (McEntire and Myers, 2004).In their report, Sutton and Tierney (2006) present the concepts underlying disaster preparedness. They have separated preparedness, and its measurement, into three units of analysis: household preparedness, business preparedness and public sector/government agency preparedness. For each unit, different research instruments are presented. There are also common metrics that can be used in the questionnaires they present, culminating in common measures to guide preparation. The metrics presented in this report are generic, that is, they provide general questions to ask with a view to being better prepared, without specifying any concrete indicators. For example, Sutton and Tierney mention that a business needs indicators concerning knowledge of risks, but they do not propose any concrete measure. Their report follows along the same lines as those discussed above (McEntire and Myers, 2004; Quarantelli, 1998) in that it suggests business practices that should lead to better preparation and proposes avenues that should be explored for this purpose.3.1 North American government policies\n4. Resilience: the emergence of a new concept: In light of the various disasters (both natural and anthropic) that have affected communities around the world, the notions of adaptability and maintenance of activities have become more important. In this regard, the Hyogo Conference laid the foundations for expanding the concept of emergency measures management to include community resilience. This resilience is reflected in \"the capacity of a system, community or society potentially exposed to hazards to adapt, by resisting or changing in order to reach and maintain an acceptable level of functioning and structure. This is determined by the degree to which the social system is capable of organizing itself to increase this capacity for learning from past disasters for better future protection and to improve risk reduction measures\" (UN, 2005, p. 9).4.1 Resilience: definition and context\n5. Conclusion: Ever since the Hyogo Conference in 2005, the resilience of communities and, by extension, of organizations has increasingly come to replace the concept of disaster preparedness. Many governments have initiated work in this regard; researchers too have been working to understand and develop the concept and provide tools that will help organizations be more resilient.The evolution of the concept of preparedness to the state of resilience in the context of engineering resilience will help organizations better understand their operation and the environment in which they operate. Preparing emergency measures in an all hazard approach assumes all risk is best represented by the study of the consequences that must be managed. The organization must learn to adapt to a changing environment that generates impacts at different levels, both positive and negative.Recent research in this field is also studying cooperation and the interrelation of different organizations in order to be more resilient. The interdependency of our critical infrastructures and of different government agencies is now well known. The resilience approach, combined with adaptability, should be applied on the ground by the organizations that manage emergency measures for the population, in other words, municipalities working together with the owners of critical infrastructures.Three key challenges were presented in the previous section:(1) evaluating an organization's adaptability;(2) gaining in-depth knowledge of a system and choosing the right information; and(3) continuously measuring resilience.By addressing these challenges, it should be possible to develop resilience indicators that will enable communities to better handle the various disruptive events that may arise and thus to better protect the people exposed to such events.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to show the evolution of the concept \"state of preparedness\" into \"state of resilience\" in the context of emergency management, and the implications raised by this new concept.\nDesign/methodology/approach: - This paper is a literature review (scientific and governmental) of the most important articles in the field of state of preparedness evaluation.\nFindings: - This article presents two trends in the state of preparedness evaluation: response capability and preparation management. These two trends contribute to the evolution of the concept \"state of preparedness\" into \"state of resilience\", a state that is defined as the ability of a system to maintain or restore an acceptable level of functioning despite disruptions and failures.\nOriginality/value: - This literature review helps define the concept of \"state of preparedness\" (in terms of both management and response capability) as the new trend of resilience.\n",
"role": "assistant"
}
] |
Linking affective commitment to supervisor to work outcomes
|
[
"Affective commitment to the supervisor",
"Work engagement",
"Innovative work behaviour",
"Feedback seeking",
"Error reporting",
"Individual behaviour",
"Employees attitudes",
"Interpersonal relations"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The concept of commitment has been defined as a \"force that binds an individual to a course of action of relevance to one or more targets\" (Meyer and Herscovitch, 2001, p. 301). Meyer and Allen (1991) proposed a three-component model of commitment that differentiates between affective, normative and continuous commitment. This study however, exclusively focuses on affective commitment because \"it is the most widely studied, the most generalizable across targets, and the most predictive of employee behaviour\" (Morin et al., 2011, p. 718). Affective commitment to a given target can be broadly defined as an attachment to, identification with and involvement in that particular target (Meyer and Allen, 1997; Meyer and Herscovitch, 2001).\nTheory development and hypotheses: Affective commitment to supervisor and work engagement\nMethod: Sample and procedure\nResults: Preliminary analysis\nDiscussion: Theoretical implications\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to explore the relationship between affective commitment to the supervisor and three work outcomes: innovative work behaviour and two types of learning behaviours, namely, feedback seeking for self-improvement and error reporting. Additionally, it seeks to examine the mediating role of work engagement in these relationships.\nDesign/methodology/approach: - In total, 192 research scientists from six science research centres in Ireland completed self-reported questionnaires. Structural equation modelling was used to test the research hypotheses.\nFindings: - Results indicated that as hypothesised, work engagement fully mediated the link between affective commitment to the supervisor and the three employee outcomes included in this study.\nResearch limitations/implications: - The cross sectional design of this study does not permit causal inferences. Additionally, all data were self-reported and therefore common method variance may be an issue. Despite these limitations, the results suggest that affective commitment to the supervisor is likely to enhance employees' work engagement, innovativeness and learning. These findings imply that building employees' commitment to their supervisors can be a potent strategy for increasing individual and organizational effectiveness.\nOriginality/value: - This is the first study which has linked supervisory commitment to innovative work behaviour, feedback seeking and error reporting. Furthermore, it highlights one potential mechanism in the form of work engagement through which supervisory commitment relates to these work outcomes.\n",
"role": "assistant"
}
] |
Assessing hospital quality management systems: evidence from Iran
|
[
"Iran",
"Hospital quality improvement",
"Hospital quality management system"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Improved healthcare has become an important and critical issue, and many hospitals managers are working to improve healthcare quality (Hammer et al., 2013; Makai et al., 2009). The pressure for quality improvement (QI) from government and insurance organizations, increasing competition in the healthcare market, enhanced customer positions and increasing awareness about patient safety make the health systems and institutions more complex, which has led to adopting hospital quality management systems (QMS) (Wardhani et al., 2009).\nMethod: Study design\nResults: Our findings showed that most hospitals were non-teaching (n=17, 60.7 percent), public-owned (n=16, 57.1 percent) and small (n=15, 53.6 percent). The QMS mean score ranged from 1.53 (benchmarking) to 2.71 (data used from internal audits to evaluate and adjust care processes). The mean score for all QMS dimensions ranged from 1.96 for \"Quality policy documents\" to 2.14 for \"Quality monitoring by the board.\" The mean QMSI score was 18.4; the lowest was 10.3 and the highest was 25.1. The QMSI score was 15.3 for public and 20.9 for non-public hospitals (p0.001) (Table I). Generally, the QMSI mean score for non-public was higher than for public hospitals. The difference between public and non-public hospitals was significant for all nine QMSI dimensions (p0.001). The largest gap was observed in the \"Evaluating results\" dimension (-1.01) and the smallest was observed in the \"Analyzing professional performance\" dimension (-0.46).\nDiscussion: The QMSI mean score in our study was 18.5 (68 percent), which is relatively good, which shows the hospital QMS maturity levels and the steps taken to improve hospital services quality. Previous studies, all conducted in Europe, reported a score between 19 and 19.7 (Hammer et al., 2013; Kristensen et al., 2015; Wagner, Groene, Thompson, Klazinga, Dersarkissian, Arah and Kringos, 2014; Wagner, Groene, Thompson, Dersarkissian, Klazinga, Arah and Kringos, 2014), which is in line with our findings. The first QMSI dimension examines how hospital management board members develop and implement their quality strategy and goals. In our study, the lowest average score was related to the quality policy document. Unlike our study, one Netherlands evaluation found that most hospitals have quality policy documents, such as mission statements and annual quality reports (Wagner et al., 2006). In a Lithuanian study, the quality policy and goals definition was the least important challenge to implementing a QMS (Buciuniene et al., 2006). We can attribute weaknesses in this area to contextual and structural factors. None quality-oriented culture, poor planning, no justification for quality management and weak QI structure are factors behind Iranian healthcare quality management failure (Mosadeghrad, 2014b). Quality should be recognized as a strategic issue and be included in hospital mission and vision statements (Mosadeghrad, 2014a). Additionally, a quality structure should be created in hospitals and supported by enough full-time staff with specific job descriptions and budgets to implement QI activities (Wardhani et al., 2009).\nLimitations: Our study has limitations. Results are based on hospital quality director perceptions and, therefore, may feature response bias, which was the case in similar studies (Makai et al., 2009). However, hospital quality directors, as the key persons in the QMS implementation, have comprehensive knowledge about their organizations. In previous research, the study population was hospital quality directors (Hammer et al., 2013; Wagner, Groene, Thompson, Klazinga, Dersarkissian, Arah and Kringos, 2014; Wagner, Groene, Thompson, Dersarkissian, Klazinga, Arah and Kringos, 2014; Xiong et al., 2016). Another limitation is that hospital managers participating in the study may have a better QMS status than non-participant hospitals. Therefore, our results may be influenced by selection bias and should be interpreted cautiously.\nConclusions and recommendations: Our study provides detailed results regarding QMS implementation and improvement areas in Iranian hospitals. Most QMSs in these hospitals were relatively good, but non-public hospitals (private and charity) were better than public hospitals. It seems that governmental legislation (including national accreditation standards) must be fulfilled to achieve continuity in hospital activities, which played an important role in deploying the QMS. A higher QMSI score in non-public hospitals can be attributed to various issues, such as the structure and management model that governs private and non-public sectors. These factors brought QMS to the health sector much earlier than the public sector and have led to effective steps in this area. Areas requiring improvement include: defining specific and clear policies and strategies for improving quality; analyzing patient experiences, feedback and complaints; establishing an effective system for collecting and analyzing performance data; using data in QI activities; evidence-based decision making; and designing a system for analyzing clinical practices.\n",
"role": "user"
},
{
"content": "Purpose: A quality management system (QMS) is defined as interacting activities, methods and procedures used to monitor, control and improve service quality. The purpose of this paper is to describe the QMS status using the Quality Management System Index (QMSI) in hospitals affiliated to Shahid Beheshti Medical Sciences University in Tehran, Iran.\nDesign/methodology/approach: In this cross-sectional study, 28 hospitals were investigated. A validated 46-item questionnaire was used for data collection. Data were analyzed using descriptive statistics, Pearson correlation, independent student's t-test and regression analysis.\nFindings: The mean QMSI score was 18.4: 15.3 for public and 20.9 for non-public hospitals (p=0.001). The lowest (1.96) and the highest (2.14) scores related to \"Quality policy documents\" and \"Quality monitoring by the board,\" respectively. The difference between public and non-public hospitals was significant in all nine QMSI dimensions (p=0.001). The QMSI score was higher in non-public and small hospitals than in public and large ones (p=0.05).\nOriginality/value: Most QMS studies come from developed countries, and there is no systematic information about the mechanisms and processes involved in implementing QMS in developing countries like Iran. This is the first study on Iranian hospital QMS using a newly developed tool (QMSI), and results showed that QMS maturity in these hospitals was relatively good, but the non-public hospitals status (private and charity) was far better than public hospitals.\n",
"role": "assistant"
}
] |
Criteria for the indoor environment for energy performance of buildings: A new European standard
|
[
"Standards",
"Heating and ventilation services",
"Temperature",
"Lighting systems",
"Noise control"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: European Parliament and Council approved, in December 2002, a directive on the energy performance of buildings. The directive requires member countries to:* develop a comprehensive methodology for calculation of integrated energy performance of buildings and HVAC systems including heating, cooling, ventilation and lighting;* set minimum requirement of energy performance of new buildings;* apply requirements in existing buildings;* develop energy certification system for buildings; and* have heating and air-conditioning systems inspected regularly.Energy consumption of buildings depends significantly on the used criteria for the indoor environment, which also affects health, productivity and comfort of the occupants. An energy declaration without a declaration related to the indoor environment makes no sense. The indoor environment is mentioned several times in the directive. First of all, energy saving measures should not sacrifice peoples' comfort and health. Secondly, it is recommended besides the energy certificate and actual values for the energy consumption also to display in the building the design values for the indoor environment and indicators for the environmental comfort. There is, therefore, a need to for specifying criteria for the indoor environment for design, energy calculations, performance evaluation and display of operation conditions.This paper describes how design criteria of indoor environment are set for dimensioning of systems and for energy calculations the draft standard (CEN prEN15251, 2005). It also highlights some of the new principles in the standard such as three categories of indoor environment, difference between target values for dimensioning and energy calculations, principles of defining the ventilation rates, and evaluation of indoor environment. For dimensioning and energy calculations different approaches are introduces for mechanical cooled buildings and buildings without mechanical cooling. Finally, this paper will discuss some of the issues pointed out during the public review.\nScope of the standard: The European draft standard specifies the parameters of impact and/or criteria for indoor environment and how they are used to meet the intent in the EPBD. It:* Specifies the indoor environment parameters, which have an impact on energy performance of buildings and specifies how to establish indoor environmental input parameters for the building system design and energy performance calculations.* Specifies methods for long-term evaluation of the obtained indoor environment as a result of calculations or measurements.* Specifies criteria for measurements, which may be used if required by inspection.* Identify parameters to be used by monitoring and displaying the indoor environment in existing buildings.* Is applicable mainly in the non-industrial buildings where the criteria for indoor environment are set by human occupancy and where the production or process does not have a major impact on indoor environment. The standard is thus applicable to the following building types: single family houses, apartment buildings, offices, educational buildings, hospitals, hotels and restaurants, sports facilities, wholesale and retail trade service buildings.The parameters and criteria are based on existing standards and guidelines (such as ISO DIS16814, 2005; EN ISO 7730, 2005; CEN prEN13799, 2005; CEN CR 1752, 2002, and existing national standards).\nDesign criteria for dimensioning: For design of building and dimensioning of HVAC systems the thermal comfort criteria (minimum room temperature in winter, maximum room temperature in summer) shall be used as input for heating load (EN12831, 2002) and cooling load (CEN prEN15423, 2005) calculations. This will guarantee that a minimum-maximum room temperature can be obtained at design outdoor conditions and design internal loads. Values are presented for the sizing and dimensioning of the systems and well as for design of buildings without mechanical cooling.In general, national specified criteria for design and dimensioning of systems must be used, but the standard gives in case of no national regulation recommended design values in informative annexes. The recommended criteria are given for three classes (categories). Using a higher class with stricter criteria will result in higher calculated design loads and then may result in larger systems and equipment. The designer shall document design criteria for the indoor environment. As an example thermal design criteria for different type of spaces are given in Table I for buildings with mechanical cooling.Especially, the criteria for kindergarten and department stores are under discussion. In these type of buildings activity will not be uniform as some people will be sedentary, children playing and people walking. Also the clothing may vary between the different groups of occupancy (sales people, customers). Other type of buildings like hospitals, restaurants, sports facilities, and warehouses will have similar problems regarding variations in activity and clothing between different types of occupants. As a basis the temperature ranges are based on a general comfort criteria using the PMV-PPD index (EN ISO 7730) of +-0.5 for category A (+-0.2 for category A and +-0.7 for category C).Buildings without mechanical cooling\nIndoor environment parameters for energy calculations: Standardised input values for the energy calculations are needed for calculations specified in article 3 and in the annex of EPBD. To perform a yearly energy calculation (CEN prEN13790, 2005) criteria for the indoor environment must be specified and documented.Thermal environment in mechanical cooled buildings\nEvaluation of indoor environment: As the loads of the building vary spatially and temporally the designed system may not be able to fulfil the design intent in all rooms during all hours. There is a need to evaluate the long-term performance of building in respect of indoor environment. This evaluation is necessary for the display of the climatic factors (indoor environment) in the energy performance certificate (a requirement of directive in articles 6 and 7). The standard also presents indicators for such evaluation and their use. The evaluation of indoor environment of a building is done by evaluating the indoor environment of typical rooms representing different zones in the building. Evaluation can be based on design, measurements or calculations. Evaluation includes:* thermal criteria for winter;* thermal criteria for summer;* air quality and ventilation criteria;* lighting criteria;* acoustic criteria; and* draft criteria.As the criteria are based on instantaneous values, it values outside the recommended range should be acceptable for short periods during a day. Therefore, it is recommended that 3-5 per cent of the time (working hours) the calculated or measured values can be outside the range. The 3-5 per cent is to be used for daily (meaning 15-25 minutes during a working day), weakly (24-120 working minutes) and yearly (50-100 working hours).Finally, the standard are discussing methods to give an overall evaluation of the indoor environment, which in the future could be used for an indoor environmental certificate. Unfortunately, only a few research results are available on combined effects of the indoor environmental parameters.\n",
"role": "user"
},
{
"content": "Purpose: - To provide input data to design and energy performance calculations of buildings and ventilation, heating, cooling and lighting systems.\nDesign/methodology/approach: - European directive for energy performance of buildings was approved in the beginning of 2003. The transition period is 3-6 years depending on the article. European Standardisation Organisation (CEN) has drafted several standards to help the member countries implementing the directive. One of these is the \"Criteria for the indoor environment including thermal, indoor air quality (ventilation) light and noise.\" The standard has been developed based on existing international standards and guidelines for the indoor environment taken into account the latest results from published research.\nFindings: - The standard specifies design values of indoor environment, values to be used in energy calculations, and methods how to verify the specified indoor environment in the buildings. The paper describes some of the principles used in standards, and gives examples presented in the standard. The standard covers all building types but the paper is focuses on the non-residential buildings, numeric examples are given only for offices.\nResearch limitations/implications: - The draft standard is under international review process during writing this paper, and subject to changes. The standard give default criteria for the indoor environmental parameters, which can be used if no national requirements are available.\nOriginality/value: - This paper describes the indoor environmental parameters, which are important for people's health, comfort and energy consumption of buildings. This will help users to select more uniform input data for energy calculations.\n",
"role": "assistant"
}
] |
Making sense of mixed-embeddedness in migrant informal enterprising: The role of community and capital
|
[
"Immigrants",
"Entrepreneurship",
"Ethnic groups",
"Enterprise"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The notion of mixed-embeddedness is a crucial theme in migrant entrepreneurship research (Kloosterman et al., 1999; Kloosterman, 2010). Mixed-embeddedness aims at reconciling changes in socio-cultural frameworks with transformations in the urban economy as it encompasses \"the crucial interplay between the social, economic and institutional contexts\" (Kloosterman et al., 1999, p. 257). To understand the choices and behaviours of migrant entrepreneurs, mixed-embeddedness requires to consider their level of embeddedness in both social networks and market opportunities (Kloosterman, 2010).\nEnterprising migrants and community: The notion of community is central to a variety of disciplines (Bulmer, 1985). Studies on geography (Cater and Jones, 1989); regional and urban development (Masurel et al., 2002); religion (Fitzgerald, 2003); media (Howley, 2005); politics (Frazer, 1999); and sociology (Tonnies, 1887/2001; Gusfield, 1975) all debated its determinants and core elements. Various attempts to conceptualise community only renewed the confusion and confirm the elusiveness of a shared definition (Delanty, 2003). Inspired by the experiences of the UNESCO Institute for Lifelong learning, Lumpkin et al. (2018) attempted to organise these perspectives into four major conceptualisations that reflect different organisational manifestations of community. The authors suggest that community can be conceptualised around geographical demarcations; identity affiliations; shared interests; or common intentions. Affiliations, interests, and intentions are essentially relational exercises as they refer to the nature and quality of relationships within a particular location (Gusfield, 1975). Tonnies (1887/2001) stressed the importance of such human connections in his initial conceptualisation of Gemeinschaft. This encompasses a web of human relationships at times interpreted as commonality of goals and objectives (Somerville, 2016); shared learning process in a particular field (Wenger, 1998); exchange of information and mutual support (Cater and Jones, 1989); issues of identity and belonging, especially in terms of kinship and networks (Coleman, 1990); cultural alignment (Portes, 1997); political mobilisation (Delanty, 2003); religious association (Fitzgerald, 2003).\nEnterprising migrants and capital in the informal economy: Previous research widely discussed how migrant enterprising is likely to emerge in informal contexts (Ram et al., 2007; Williams, 2007; Ram and Jones, 2008). Informal enterprising remains under-researched in mature economies, as the phenomenon is traditionally associated with developing or emerging economies (Webb et al., 2014; Williams, 2015). Dedicated research stressed the importance of informal enterprising also in mature contexts (Frith and McElwee, 2009). This is especially important as informal ventures often incubate future formal businesses (Williams and Martinez, 2014).\nMethodology and research design: The paper adopts an interpretative stance to explore how migrant entrepreneurs interpret mixed-embeddedness. It draws on the empirical evidence from a cross-sectional sample of 26 asylum seekers that engaged with enterprising activities in the city of Nottingham in the UK. A qualitative analysis of the accounts helped to develop theoretical insights on how these particular migrants use understanding of community and notions of capital to make sense of their mixed-embeddedness.\nAnalysis and findings: The data analysis involved five rounds. The first and second round observed the relations with social networks (Engelen, 2001) and market opportunities (Ram and Jones, 2008), respectively. A recursive hermeneutic process guided the iterative readings of the data and of the related theory to ensure the process did not exclude any relevant concept (Drakopoulou-Dodd et al., 2018). The progressive emergence of discourses of community and value-creation invited the research team to go back to the literature to explore these notions and incorporate them to frame the analysis. The third and fourth round of analysis hence explored the geographical and relational aspects of community, respectively (Gusfield, 1975). Bourdieu's (1986, 1989) notions of capital proved useful to interpret what guided the migrants' entrepreneurial choices in the final step of the analysis.\nDiscussion: Mixed-embeddedness remains a key framework for understanding migrant entrepreneurship, especially in informal contexts. Being an interactionist approach, it links agency and structure in a meaningful way (Kloosterman, 2010). However, the concept remains fuzzy, difficult to operationalise and to verify empirically (Razin, 2002). Although it reconciles different levels of analysis of the migrant entrepreneurship phenomenon, it falls short in explaining the interpretative positions of migrant entrepreneurs. Mixed-embeddedness sees agents as reflexive actors who can to some extent inform their own choices (Kloosterman, 2010). Nevertheless, little research exists on 'how' these entrepreneurs enact their reflection.\nImplications for research and practice: In terms of implications for future research, the framework proposed in this paper can form the basis for allowing scholars to address empirically how migrant entrepreneurs reconcile their embeddedness in both social and economic contexts. In terms of designing new research, scholars can use the framework to focus on aspects other than ethnicity in investigating informal migrant enterprising. Furthermore, the framework justifies the investment in potential longitudinal studies that will allow researchers to better observe these transformations over time. Finally, future research could explore in detail each of the emerging forms of enterprising to investigate under-researched issues in informal migrant enterprising. For example, research on mentoring-based forms of enterprising might contribute to shed light to yet unearthed enterprising dynamics in terms of talent management in the context of the informal economy.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to discuss how informal migrant entrepreneurs with different legal statuses interpret their mixed-embeddedness in social and economic contexts. Legal status represents a key determinant in shaping accessible social networks and market opportunities that in turn influence entrepreneurial choices.\nDesign/methodology/approach: The paper adopts an interpretative stance to explore how migrant entrepreneurs interpret mixed-embeddedness. It draws on the empirical evidence from a cross-sectional sample of 26 asylum seekers that engaged with enterprising activities in the city of Nottingham in the UK. A recursive hermeneutic process guided the iterative readings of the accounts to develop theoretical insights on how these agents reinvent their relationship with structure.\nFindings: A novel theoretical framework emerges from the data analysis to present how these particular migrants use understandings of community and notions of capital to make sense of their mixed-embeddedness. The main theoretical contribution of the framework is to illustrate how groups with different legal statuses produce unique interpretations of mixed-embeddedness. This, in turn, reflects onto specific forms of enterprising and innovative entrepreneurial choices. The framework also produces an empirical contribution as it re-centres the analysis of mixed-embeddedness around the migrant entrepreneur from previous meso- and macro-level perspectives that dominated recent research.\nResearch limitations/implications: The paper expands knowledge on the notion of mixed-embeddedness by providing insights on how informal migrant entrepreneurs make sense of it. This can form the basis for allowing scholars to address empirically how migrant entrepreneurs reconcile their embeddedness in both social and economic contexts. In terms of practical implications, the paper paves the way for policy-makers to re-evaluate the current approach to the right of asylum seekers to pursue entrepreneurial activities.\nOriginality/value: The notion of mixed-embeddedness is central to research on informal migrant enterprising. Nevertheless, the concept remains fuzzy and difficult to operationalise. The paper offers an opportunity to understand how migrant entrepreneurs make sense of mixed-embeddedness so that future scholars can better explore how mixed-embeddedness reconciles agency and structure.\n",
"role": "assistant"
}
] |
Searching and deleting: youth, impression management and online traces of search
|
[
"Search engines",
"Goffman",
"Ethnography",
"Youth",
"Impression management",
"Online traces"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper investigates the role of traces of online search in young people's everyday lives. More specifically, it addresses, through a qualitative study, how Swedish teenagers reflect on as well as deal with the information that becomes stored as they use various search engines on their mobile digital devices. Such devices are a feature of everyday life for most teenagers in Sweden. Most of them have their own smartphone and also have access to other digital devices such as computers, tablets or laptops. The statistics from Sweden show similar patterns to that of the USA and the EU in regards to ownership and use of various mobile devices among teenagers (Anderson and Jiang, 2018; Davidsson et al., 2018; Livingstone et al., 2017). These devices facilitate all kinds of activities from online search to the use of social media. The activities have in common that they leave traces; some in a way that makes it visible to the user, such as through a search log, while others are more difficult to detect, such as third-party tracking. Pangrazio and Selwyn (2019, p. 420) note that \"as digital data become more ubiquitous to everyday life, it is also becoming increasingly difficult for non-specialists to define and understand\". This poses new challenges for internet users and points towards a development where the systems concerned are so complex that it is hard to grasp the consequences of online activity. It is therefore difficult to make informed decisions in relation to online activity and personal data, from social media to online search. Given these premises, how do young people navigate and make decisions in relation to their online activity?\nPrevious research: Young people and online traces\nMethod and material: The material for the study was produced through an ethnographic study between December 2014 to May 2016 wherein focus groups, class room observations, interviews, and go-alongs with pupils were carried out. The ethnographic approach stems from Anthropology but has been taken up by many disciplines. While different methods can be used within ethnographic work, unlike statistics that give a broad picture; \"ethnography is better at representing the nuances and complexities of everyday life\" (McGuigan, 2009, p. 93). The ethnographic approach was well suited for the project since the aim was to understand the role of search engines in the everyday life of teenagers. Understandings of nuances and complexities in everyday life are traditionally achieved through longer stretches of fieldwork; like the one in this study (Ehn et al., 2016). In the current paper, material from interviews, go-alongs and observations in two schools during the spring of 2016 was analysed. The participants were pupils aged 15-16 years. There were nine participants in total - five from School A and four from School B. Previous studies from the ethnographic project, addressed online search in school and also how the participants use and reflect on using Google (see also Andersson, 2017a, b). The question of online traces of search was not addressed previously. The research was carried out in accordance with ethical research guidelines (Swedish Research Council, 2002). The participants were informed that participation was voluntary and that they may stop participating at any point. Approval was sought from both participants and their parents. Prior to interviews, the participants were given a written statement about the project that was to be signed by their parents or legal guardians in order for the interviews to be carried out. All interviews were recorded and transcribed by me. The schools and participants have been anonymised in the material. As is common practice when using ethnographic methods; the article is deliberately written in first person.\nConceptual framework: Impression management and presentation of self\nFindings: Searching and deleting\nDiscussion of findings and implications: This paper can be viewed as a novel exploration of the role of online traces of search among young people, as well as an exploration into using audience and impression management as analytical concepts. The present study shows that traces of search can influence the activity of searching as well as where it is carried out and on which device. For the participants in the study the topic for the query could influence which device that they use. The findings show that while mobile devices make it possible to search for information almost anywhere at any time, context still matters. The participants' knowledge of how to remove traces of search, and thereby protect their impression management, can influence which device they prefer to use. Bowler et al. (2017) suggest that questions such as where data lives and who has access to it, are not yet part of young people's thought process. In contrast, the findings here show that the participants had an awareness of the fact that their online activities left traces; yet they did not fully grasp how data flows, and for the most part they viewed their online traces as residing in the devices. The findings point towards that it is easier for young people to conceptualise the flow of information when it might have consequences for them. One of the main reasons for removing traces is that they do not want certain people - an audience - to view items on their search history beyond their control. While the participants were not asked about audience as such it was clear that who might see their search history was an important component in their removal of online search traces. The concept can thus be beneficial to the researcher when constructing questions to participants but also for the analytical process. Furthermore, when teaching about data flows, the audience concept can be useful in order to illuminate how networks function. Carrying out this type of analysis could also be used to illuminate issues of power. Although power was not in focus for the paper, implicitly it still is as the participants usually explain wanting to keep certain information away from teachers and parents, i.e., people who are in a position of power in relation to them, because of the potential consequences otherwise.\nConclusion: This paper addressed the ways young people conceptualise and deal with their online traces of search. The way they deal with their online presence, I would argue, can be viewed as them taking control and agency in a digitally saturated and complicated situation. This resonates with Buchanan et al. (2017, p. 280) and their discussion on agency: \"When examining children's online behaviour, we consider children's agency as being evident in the ways [...] that they mediate their online actions based on past experiences, present contingencies and an understanding that their actions have future consequences\". In a similar manner, the participants in this study dealt with their online traces as best they could. The future consequences of leaving traces were usually understood as an adult or peer gaining access to them. However, this did not seem to make them refrain from online search, but instead compelled them to erase their traces, by searching and deleting. The concept of audience has been helpful in elucidating how the participants conceptualised who might see their online traces. The participants were for the most part more concerned with parents, friends and teachers viewing their traces of online activity and less concerned about corporations mining their data. Their strategies in relation to online traces consisted of deleting individual entries in their search history, or deleting it altogether. Most of the participants had several mobile devices and tried to maintain a division of labour between them, e.g. so that school-related activities were carried out on the school laptop. Yet this division became blurred as the various devices move around with the user. This mobility sometimes contributed to an understanding of the fact that data flows, such as through using the school Wi-Fi on a private smartphone. One consequence of their conceptualisation of online traces is that their strategies were based on removing that which they could see, rather than a strategy to act beforehand in regards to certain information. Another consequence is that they might believe that they have been successful in not sharing any data if they could not see it in their search log. The findings of this study highlight that as we equip young people with laptops in school, we should also equip them with an understanding of data and online traces, both in relation to what data the schools are recording and to a wider context. Learning to become information literate is not only, or even mainly, a question of learning to use devices and platforms in a narrow and instrumental way. Literacy in relation to data, information and media should also include a critical awareness of data flows between devices and platforms in a wider sense.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore and analyse how young people conceptualise online traces of search and what strategies they have for dealing with them.\nDesign/methodology/approach: The topic was investigated through a qualitative and ethnographic approach. Interviews, go-alongs and observations in schools were carried out with nine participants in Sweden.\nFindings: The findings show that the participants' main strategy in relation to online traces was to remove items from their search logs. Search logs were a tangible way of conceptualising online traces of search. The participants removed items in relation to an imagined audience which, in the present material, primarily consisted of parents and teachers. The findings also showed that the participants had some awareness of online traces but had difficulties in understanding the way that data flow and the persistence of data. Their strategies were more reactive than proactive in relation to online traces.\nOriginality/value: The present study contributes with a novel exploration of understandings of online traces of search. Further, it contribute theoretically by investigating the topic through the lens of audience and impression management.\n",
"role": "assistant"
}
] |
Fear and loathing in cooperative collection development
|
[
"Interlibrary loan",
"Cooperative collection development",
"Consortial borrowing",
"Data repository",
"Libraries",
"Collections management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Shared collection development, particularly among large research libraries with a culture of ownership, is a scary and unwelcome prospect. Libraries are afraid that they will not be able to provide needed research material promptly to their users and the burden of cooperative collection development appears to be so much more difficult and time consuming than sole ownership, despite the potential savings. This article will evaluate the Borrow Direct partnership as the foundation for sharing physical resources to overcome the fear of cooperative collecting. Borrow Direct is a partnership of nine academic research libraries: Brown University, Columbia University, Cornell University, Dartmouth College, Harvard University, Massachusetts Institute of Technology, Princeton University, University of Pennsylvania, and Yale University. This article will envision a data collection and storage model that could address the loathing inherent in cooperative collecting.\nThe founding of Borrow Direct: a \"Kumbaya\" moment: Borrow Direct offers library users a searchable collective catalog of all the partner libraries\nBorrow Direct policies and procedures: less is more: It is tempting to try to imagine every scenario, document every variable, and build functionality to address every possible event, but the rejection of this philosophy contributed greatly to the ease of running Borrow Direct at the partner libraries. The emphasis from the start was on access and speed. There was no staff mediation of requests, meaning there was no control over the number of requests placed per person or even the number of times the same person requested the same item. Mediation of all transactions in order to catch the small percent of problem transactions was deemed not to be worth the staff time or the patron\nSystem integration: let the computer do the heavy lifting: Two of the most labor intensive tasks in ILL are getting the patron\nThe Borrow Direct experience: show me the numbers: Clever design can be entertaining to discuss but actual performance in the workplace is the true measure of any innovation. While anecdotal feedback has been very positive, user feedback is a self-selective measure. The real question is: has Borrow Direct met the objective of providing fast, reliable service that could allow librarians across the partnership to begin considering cooperative collecting? In other words, is it enough that a copy of a book is available through Borrow Direct or must my library own a copy to ensure access?\nCooperative collecting: a hard nut to crack: No resource sharing platform, no matter how good, can erase all of the barriers to cooperative collecting, particularly among autonomous institutions with an historical reputation for high-quality, comprehensive collections. Even with subjective considerations aside, national rankings such as the Association of Research Libraries (ARL)[2] rely on the finite measure of ownership, not access. In this environment, no single library wants to be the first to appear to be ceding their collecting duties to outside entities, even (or perhaps especially) peer institutions. Overcoming this taboo requires collaboration, communication, and information.\nA comprehensive data repository: you snooze, you lose: Documenting potential questions can inform what data a library should collect, but it is simply impossible to know today what questions will be asked next year or five years from now. If you do not have the data, it will be too late. Facebook knows this, Google knows this; it is accepted wisdom in the marketplace. Gather everything you can, even if you do not know what you need it for. Just as e-commerce informed the evolution from a high-maintenance, mediated ILL environment to an unmediated, assembly line workflow, comprehensive data gathering will drive the evolution of cooperative collecting. Information is power and librarians should be the first to recognize this.\nMetridoc: one ring to bind them: There may be a multitude of ways to gather data for collection assessment, but Metridoc[3] is one example of a database project that is designed to extract data from user actions and link that data to communities, funding sources, and activities. It is an open source project at the University of Pennsylvania, sponsored by the Institute of Museum and Library Services. The underlying philosophy of Metridoc is to capture data at the event level. In layman\nConclusion: Several years of debate and development took place between the first planning meeting for Borrow Direct and its implementation. Borrow Direct has continued to evolve and grow over the years. The e-commerce model that Borrow Direct was built on is old news in the marketplace, but for the most part, libraries have lagged behind when it comes to streamlining resource sharing operations and putting the power in the hands of the user. As the numbers prove, removing the impediment of mediation and forging cooperative agreements among peer institutions will transform ILL and make cooperative collecting a real and distinct possibility. In order to seize the new collecting opportunities, libraries must be in a position to make informed decisions. It is now time to recognize and adopt the marketplace trends in tracking and storing use data. To gather only enough data to answer today\n",
"role": "user"
},
{
"content": "Purpose: The paper\nDesign/methodology/approach: The author uses data from the Borrow Direct cooperative to illustrate how efficient and effective ILL services have become, and identifies a number of data types that might be collected in a data repository and used to inform the construction of a successful cooperative collection development program.\nFindings: A data repository consisting of the following data types will prove useful in outlining the parameters of a cooperative collection development program: ILL and circulation transactions, database usage statistics, online catalog searches, electronic resource access logs, courseware usage and resources, user demographics course offerings and enrollment, and funding sources.\nOriginality/value: Library managers and administrators interested in levels of library cooperation much deeper than the traditional ILL and consortial borrowing operations will find this article especially useful. The discussion of Metridoc and its application in building data repositories is also helpful.\n",
"role": "assistant"
}
] |
Professional identity and role transitions in clinical managers
|
[
"Nurse",
"Identity",
"Clinical manager",
"Doctor",
"Hybrid management",
"Role transition"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Clinicians who take a role as clinical manager have to balance and mediate between different worlds - the managerial and the professional - with different and often conflicting demands and logics (Gray and Harrison, 2004; Llewellyn, 2001; Witman et al., 2011). While some authors suggest that clinicians can occupy such roles easily (Llewellyn, 2001, p. 593), others argue that clinicians may fail to meet their own and other's expectations (Fulop, 2012, p. 579; see also Kippist and Fitzgerald, 2009). There is a need for more research on how clinicians conceptualize and \"do\" management in order to better understand why expectations might not be met (Witman et al., 2011). Clinicians' professional background influences their identity and might have an impact on how they perceive and transition into the managerial role (Fitzgerald et al., 2006).\nMethods: Participants\nFindings: We found that doctors experienced difficulties in reconciling the role as health professional with the role as manager. They maintained a health professional identity and reported to find meaning and satisfaction from clinical work. Doctors also emphasized clinical work as a way of gaining legitimacy and respect from medical colleagues. Nurses recounted a faster and more positive transition into the manager role, and were more fully engaged in the managerial aspects of the role. We elaborate further on these findings below. We have organized the findings in three sections: transitioning into the managerial role, doing clinical management and views on being a clinical manager.\nDiscussion: Professional identity\nMethodological considerations: Witman et al. (2011) argue that there is a systematic bias in the literature on clinical managers, as most of the research is based on interviews, with little emphasis on observations. The fact that we were able to observe participants in meetings with staff and other professionals gave us an opportunity to observe possible discrepancies in verbalizations and behaviors. It also enabled us to analyze views in a more nuanced manner. Observations validated the accounts that participants gave in interviews about their experiences and enactment of the manager role. Another strength of our study is that we explored both doctors and nurses' views and experiences in the same organizations, which allowed us to pinpoint differences between managers with different professional backgrounds. Although there will be individual exceptions, we believe that our findings are transferable to other countries, as clinicians and especially doctors, often have similar socializations into their professional role. The difference in education and status between doctors and nurses is also similar. Although we have suggested that need satisfaction could play an important part in successful role transition, we were not able within the design of the present study to test this hypothesis. Future studies could quantitatively compare the degree of need satisfaction experienced by clinicians in manager roles versus their respective professional roles. Researchers could also apply a longitudinal approach by following the same individuals over several years and exploring factors that influence transitions in identity. Lastly, we also found that doctors recounted a sense of loss involved in terminating clinical commitments. Coupled with results from other studies (e.g. Ham et al., 2011), this appears to constitute a general pattern among medical managers and can be understood from the strong sense of meaning and purpose derived from doing patient related work (Pratt et al., 2006). Although the sense of loss did not constitute a general theme among the nurses in our study, similar experiences might be involved when other health professionals take on managerial roles. This could be studied further.\nImplications: Management programs aimed at clinicians who take on managerial responsibilities have traditionally emphasized the development of specific competencies, skills and techniques. Our findings suggest that health care organizations should acknowledge the importance of need fulfillment in the managerial role. Need fulfillment may be associated with effective identity building and role transition. The role as manager needs to become part of one's self-identity in order to sustain interest for the years required to develop and practice complex management skills. If clinicians are to develop an interest in management, and subsequently develop a complex identity, taking on management responsibility should be seen as contributing to need satisfaction, rather than thwarting it. Management programs should also acknowledge the sense of meaning and purpose imbedded in the professional role, and that this could create a feeling of loss when clinicians move into management positions.\nConclusion: Our study suggests that the inclusion of aspects from identity and need satisfaction literature could enrich studies of clinicians' transition into managerial roles. Our study suggests that need satisfaction might have a part in sustaining clinicians' motivation to engage and identify with a managerial role, and thus developing a complex identity. Organizational efforts to engage clinicians in management need to acknowledge the professional identity they bring into the managerial role, and the specific barriers that need to be addressed in order to support the transition into a new identity. Our results indicate that aspects related to re-socializing and identity should be included in management development programs.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate how clinicians' professional background influences their transition into the managerial role and identity as clinical managers.\nDesign/methodology/approach: The authors interviewed and observed 30 clinicians in managerial positions in Norwegian hospitals.\nFindings: A central finding was that doctors experienced difficulties in reconciling the role as health professional with the role as manager. They maintained a health professional identity and reported to find meaning and satisfaction from clinical work. Doctors also emphasized clinical work as a way of gaining legitimacy and respect from medical colleagues. Nurses recounted a faster and more positive transition into the manager role, and were more fully engaged in the managerial aspects of the role.\nPractical implications: The authors advance that health care organizations need to focus on role, identity and need satisfaction when recruiting and developing clinicians to become clinical managers.\nOriginality/value: The study suggests that the inclusion of aspects from identity and need satisfaction literature expands on and enriches the study of clinical managers.\n",
"role": "assistant"
}
] |
Study on the cultural communication preferences of teachers
|
[
"Communication",
"Culture",
"Teachers",
"Turkey"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Communication is a basic dimension shaping human relations. The issue of \"communication\" attracts the attention of many groups and has been subjected to many studies due to its various components and types. A very broad definition of communication can be \"mutual interaction.\"Another communication definition is: to ensure understanding between people through verbal and nonverbal ways to achieve desired objectives and to affect behaviors (Can, 1994).In today's globalizing world, there are many opportunities to establish communication between people from different cultures. While only a few people preserved the monopoly of \"establishing communication between people from different cultures\" in the past, this has gradually turned out to be a problem of all humanity today. Increase in the number of communication tools and the increase in international travel have had a huge impact on this phenomenon. To obtain an effective result from communication established between people from two different cultures, it is important for both sides to know the cultural background of the other. While encompassing everything about humanity, culture shapes the communication styles of people as well. Social scientists are interested in the problems arising during communication processes between people from different cultures and which cultural features affect the communication and in which way.By analyzing the relationship between communication and culture, Edward Hall classifies cultures as high- and low-context cultures from the point of communication. People in high-context cultures do not convey everything they want to say while establishing communication with others. Thus, it is difficult for the receiver to completely understand the intention of the speaker by just looking at the message. Other environmental factors are affective in the communication as well as the message. These factors are \"how the message is conveyed\"; \"gestures and mimics of the person conveying the message\"; \"words, signs, etc. used and stresses made by the same person\"; and other contextual factors (Morden, 1999; Hofstede, 2001; Erdem, 2003). In high-context communication, indirect expressions are more prominent than open and clear expressions. Generally Middle Eastern countries, China, and Japan can be listed as examples of high-context cultures (Boru, 2007).In a high-context cultural environment there will be a differentiation between the formal aspect and real aspect of communication. Formal aspect forms nonfunctional and rasped side of communication in the cultural environments where high-context communication is dominant. It is necessary to use informal channels to be able to understand people in the real sense. In high-context based institutions, it is important to learn the data flow in the informal communication channels. Formal-informal differentiation may be said to be lower in low-context cultures.In low-context cultures, on the contrary to high-context cultures, it is indicated to be open and direct in communication. In such kinds of cultures, everything intended to be conveyed is given with the message. There is no need for any factors other than the message (Sumer, 2000; Tayeb, 2001; Erdem, 2006b). According to Hofstede (2001), the classification made by Hall as \"high-context and low-context cultures\" is also a distinction of \"traditionalism and modernism.\" High-context communication is generally a characteristic of traditional societies while low-context communication is a characteristic of modern societies.Detection of which communication tendency is prevailing in education institutions is of great importance to understand how the system functions.\nMethodological issues: Aim\nFindings of the study: Distribution of the participant teachers on the basis of specific independent variables is given in Table I.The age average of teachers has been calculated as 36.32+-8.85 (year), the average employment period as 13.64+-8.88 (year), and the income average as 1444.04+-567.17 (YTL).Cronbach a coefficients of the dimensions in KIO-TUR scale, arithmetic averages and standard deviations of the scores taken by teachers in each dimension and the statistical results showing inter-dimensional comparisons are given in Table II.As can be seen in Table II, teachers scores above 4 (which can be accepted as midpoint) in both communication dimensions. This reflects that teachers bear the traces of both communication tendencies. To find out which communication tendency is more dominant, dimensions have been compared through paired t-test and statistically meaningful differences have been recorded between these two dimensions (t=4.409, p=0.000). High-context communication has been found to be more dominant than the low-context communication in teachers included in the sampling group of the study.Cultural communication tendencies have been compared on the basis of the teachers' demographic features (Table III).When teachers' cultural communication tendencies are compared on the basis of sex (Mann-Whitney U test), no statistically meaningful difference has been detected between males and females in terms of high-context dimension (Z=-1.104, p=0.270) and low-context dimension (Z=-0.567, p=0.570). In the comparison made on the basis of marital status, while no meaningful difference has been revealed between the groups in the high-context dimension (Z=-1.650, p=0.099), a statistically meaningful difference has been observed between two groups in low-context dimension (Z=-1.968, p=0.049). It has been found that married teachers establish low-context communication more when compared with single teachers. It is necessary that such a result must be examined in detail.Variance analysis has been used to compare teachers' communication tendencies on the basis of their age groups. Meaningful differences have been observed in high-context communication dimension in terms of teachers' age groups (F=6.459, p=0.002). Turkey's-b test points out that the difference in the results stems from that fact the high-context communication scores of \"those 30 years or less\" and \"those 31-40 years old\" are higher than the corresponding score of \"those teachers who are 41 or more years old.\" It can be concluded that high-context communication tendencies decrease as teachers get older. Meaningful differences have also been detected in the low-context communication tendencies in terms of teachers' age groups (F=8.575, p=0.000). Turkey's-b test conducted has pointed out that the given difference results from the fact that the score of teachers who are in the \"30 years old or less\" age group is lower than the scores of other groups. An increase is observed in the low-context communication tendencies of teachers as they get older.When teachers' communication tendencies are compared on the basis of employment periods (Kruskal-Wallis variance analysis), significant differences have been detected between the groups in terms of high-context communication (X2=7.048, p=0.029). Paired comparisons have revealed that this situation results from the fact that the score of teachers belonging to \"20 or more years of employment\" group is lower than the scores of \"10-19 years of employment\" group and \"0-9 years of employment\" group. The tendency of teachers to establish high-context communication decreases as the employment period increases. When teachers are compared in a low-context communication dimension on the basis of their employment period, a significant difference has been detected between the groups (X2=10.752, p=0.005). Pair comparisons have revealed that this difference results from the fact that the low-context communication score of \"0-9 years of employment\" group is lower than the score of \"20 or more years of employment\" group. The tendency of teachers to establish low-context communication increases as the employment period increases.When cultural communication tendencies are compared on the basis of educational levels, no meaningful difference has been observed between the groups in terms of a high-context communication dimension (X2=0.692, p=0.707) while a meaningful difference has been found between the groups in terms of a low-context communication dimension (X2=7.337, p=0.026). Pair comparisons have revealed that this difference results from the fact that low-context communication scores of \"pre-license graduate\" teachers are higher than the corresponding scores of \"license graduate\" teachers.When teachers' communication tendencies are compared on the basis of income levels, no meaningful difference has been detected between groups in terms of a high-context communication dimension (X2=4.941, p=1.174) while a meaningful difference has been found between groups in terms of low-context communication dimension (X2=6.240, p=1.044). Pair comparisons have revealed that this difference results from the given scores of \"999 YTL or below\" income group which is higher than the corresponding scores of other income groups.\nResults and evaluation: In the scope of this study, teachers' cultural communication tendencies have been evaluated in terms of high- and low-context distinction and, comparisons have been made on the basis of demographic variables. Teachers have attained scores above 4 (midpoint of the scale) in both dimensions. In the light of this finding, it is possible to conclude that teachers bear traces of both high- and low-context communication tendencies. However, when dimensions are compared with each other, it is seen that high-context communication is more dominant than low-context communication in teachers.Hofmann (1999) states collectivist cultures prefer high-context communication, while individualistic cultures prefer low-contextual communication. In the studies conducted on the individualism-collectivism distinction in sampling groups selected in Turkey, it is concluded that collectivism tendency is more dominant than individualism (Wasti, 1995; Aycan and Kanungo, 2000; Sargut, 2001; Erdem, 2003). In the scope of this study it is also observed that teachers' high-context communication tendencies are more dominant.In addition, in the studies conducted by Erdem (2005, 2006b) who has developed the KIO-TUR scale used in this study, high-context communication has been found to be more dominant than low-context communication in Turkey.Prevalence of high-context communication tendencies in teachers shows that teachers prefer to use indirect and implicit expressions rather than clear and direct expressions in their messages. It is also concluded that sex (a demographic variable) has not lead to any meaningful difference in both communication dimensions. Marital status, education level and income level, on the other hand, are found to result in changes only in a low-context dimension, while age and employment period variables are found to result in changes in both dimensions.Low-context communication tendencies of married teachers are seen to be higher when compared with single teachers. An increase has also been observed in low-context communication tendencies with the increase in age and employment period. Low-context communication tendencies increase as life experience increases.According to research results, it can be said that an implicit and indirect communication is prevailing among teachers, merely looking at the message is not enough to communicate effectively, it is necessary to pay attention to other media factors, too (speaking manner, gestures and mimics, intonation, etc.).Detection of teachers' communication tendencies is of great importance to understand how the education system functions. The relationship between communication tendencies and organizational variables can be discussed in future studies. This study has been conducted with teachers employed in three neighboring provinces of Turkey. Future studies can be conducted on many sampling groups from difference regions of Turkey.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to look at the communication preferences of teachers on the basis of high- and low-context distinction.\nDesign/methodology/approach: - In total, 275 teachers employed in primary schools in Elazig, Malatya, and Diyarbakir provinces were chosen as the sample group. Cultural communication scale-Turkish scale developed by Erdem was applied to measure communication tendencies of teachers.\nFindings: - At the end of the paper, during which high- and low-context dimensions are treated separately, high-context communication tendencies of teachers are found to be more dominant than their low-context communication tendencies. Communication tendencies have also been compared in terms of demographic variables and it is evident that variables such as marital status, education level, income level, age, and employment period have resulted in differences in communication tendencies.\nOriginality/value: - The paper presents a study on the cultural communication preferences of teachers from a Turkish perspective.\n",
"role": "assistant"
}
] |
Food supply chain integrity: the need to go beyond certification
|
[
"Halal",
"Supply chain integration",
"Food integrity",
"Food supply chain integrity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Prior to the last few decades, the food supply chain (SC) and predominantly localised and mostly consisted of the small- to medium-sized independent local businesses (Roth et al., 2008). The business concentration enables good visibility and control along the SC with lesser tiers of suppliers (Tse and Tan, 2011). The deliberation of food SC for only the targeted nearby market is due to the absence of technological infrastructures (i.e. transportation, cold warehousing, and food technologies) which restricts food from travelling longer distances. As a result of this, the earlier SC is considered as the more effective SC network (Hearnshaw and Wilson, 2013). For example, the smaller marketplace permits higher attention to each specific product; hence the quality is always maintained as the production is largely based on more accurate sales forecasts. In addition, the SC concentration enables direct interaction between the sellers and the buyers, which facilitates the control on the food products' quality. A relatively more controlled marketplace minimises the risk of the members of SC to be opportunistic and violate the norms of the shared values and trust. Lyles et al. (2008) have argued that under conditions of trust, there is an expectation that the partners of the SC are capable of performing the tasks expected which is gained from the transparency of the business process by the food manufacturer.\n2. Literature review: 2.1 Halal food SC challenges\n3. Research framework: Our conceptual model (as shown in Figure 1) includes SC integration, halal food SC integrity, and firm performance. In particular, SC integration is consists of internal integration, supplier integration, and customer integration (Wong et al., 2011). Meanwhile, the halal food SC integrity is conceptualised as a hierarchical component model that includes four dimensions: raw materials integrity, production integrity, service integrity, and information integrity. The conceptualisation primarily covers the food integrity definitions that was highlighted by Elliott (2013) comprising the form of the way food was sourced, procured, distributed, and being honest about the elements to consumers from the context of SC.\n4. Empirical study: 4.1 Survey instrument development and deployment\n5. Discussion: The major aim of this study is to establish a conceptual model adopted from a SSP paradigm for investigating the relationships between SC integration and halal food SC integrity and the impact of halal food SC integrity on firm performance in a Malaysia context. It was purported that the strategy had a direct effect on the structure, and the structure on the performance. This means that the positive impact between the SC integration and halal food SC integrity indicated the feasibility of implementing the strategy to the structure, and when the halal food SC integration was linked with firm performance, the positive connotations indicated that the practices led to increased firm performance. The first hypothesis set up in this section was concerned with the association between the SC integration and halal food SC integrity, followed by that of halal food SC integrity and firm performance. We conceptualise the halal food SC integrity as formative-formative hierarchical component model consisting raw materials integrity, production integrity, service integrity and information integrity.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to establish a conceptual model adopted from a strategy-structure-performance paradigm for investigating the fit between the supply chain integration and halal food supply chain integrity and the impact of halal food supply chain integrity on firms' performance in a Malaysian context.\nDesign/methodology/approach: This study comprises a sample of a halal manufacturing firm in Malaysia. A cross-sectional research design was used in this study. Data were gathered based on mailed and personally administered questionnaires. SmartPLS was used to analyse the 254 valid responses.\nFindings: The research findings indicate that internal integration and strategy have positive impact on halal food supply chain integrity. The study results confirmed that customer integration and supplier integration contributes to halal food supply chain integrity. It also finds that halal food supply chain integrity has a significantly positive impact.\nResearch limitations/implications: The results suggested that a strategic collaboration with the supplier pivoted around the quality and integrity of the raw materials should be undertaken.\nPractical implications: The results from this study supports that the managers should adopt all halal food supply chain integrity components to achieve a superior performance. Even though some of the components did not yield significant results in terms of their relationships with firms' performance, these dimensions were generally related to the standardised industry requirements, such as certifications.\nOriginality/value: The findings are original and unique and are based on established theories from the literature on supply chain management practices. The research findings are useful to academics and policymakers interested in fostering a halal supply chain in Malaysia.\n",
"role": "assistant"
}
] |
Malaysian consumers' credit card usage behavior
|
[
"Consumer behaviour",
"Debts",
"Psychographics",
"Credit cards",
"Malaysia",
"Developing countries"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The spread of credit card ownership and usage across developing Asia Pacific countries has been overwhelming. A review of literature on credit card reveals that most studies have been undertaken in developed country settings. Credit cards were first issued in the USA in the early twentieth century. Since then, they have become a major system for exchange of transactions (or payments) that stimulates household and personal spending even in many developing countries of the world (Watkins, 2000).In Malaysia, cards were first introduced in the mid-1970s (Loke, 2007). At the early stage, credit cards were only issued to professionals or those considered successful businesspersons by card issuing companies. By the end of 1970s, an estimated 20,000 cards were issued. During that time, owning a credit card was considered a symbol of prestige. However, with the passage of time, eligibility criteria for obtaining credit cards have been increasingly relaxed. As a result, the number of cardholders reached to about three million by the turn of the last century.The proliferation of credit cardholders has brought an indiscriminate spending by users with many side effects. Many Malaysian consumers seem to display excessive buying behavior, commonly known as \"compulsive buying.\" Easy availability of credit and compulsive buying has led to many adverse consequences such as addiction to shopping and excessive debt. Another noteworthy feature of card users in Malaysia is that they use the card to grant loans for themselves. It is easier to obtain credit cards in Malaysia rather than applying for personal loans, which requires the applicant to comply with more formalities like providing guarantors or collaterals. It was reported that outstanding debts from credit card holders amounted to RM15.719 billion by the year March 2009 (RM or Ringitt Malaysia is the local currency, 3.40 RM was approximately equal to 1 USD at the time of study). By the year 2009 outstanding credit card debts accounted for 1.35 percent of the total loans outstanding or 11.41 percent of the total consumer credit (Bank Negara Malaysia, 2009). On a more serious note, 6.43 percent of the outstanding debts had to be converted to non-performing loans. An alarming increase in the number of credit card holders seeking bankruptcy proceedings over the years was also reported.The objectives of this research are to study the effect of personal attributes on the attitude of credit card usage, as well as to measure the effect of attitudes toward average credit card debt. This study proposes a model that may lead to a better understanding of average debt of credit cardholders in Malaysia, which in turn is influenced by cardholders' attitude and behavior toward usage of credit cards. While, past studies have revealed that social and psychographic characteristics affect consumers' purchasing behavior (Gatignon and Robertson, 1991; Rogers, 1995), findings regarding the effects of these characteristic variables have not been consistent across studies (e.g. Rogers and Shoemaker, 1971; Robertson, 1985; Steenkamp et al., 1999). The present study is transpired by these contrary findings that examine consumer behavior theories in predicting and explaining such behavior. The causal path of the model is that psychographic variables lead to attitude formation that in turn influence behavior. A positive attitude toward credit card is likely to ensure that consumers are cautious in spending.\nLiterature review: Consumer credit\nConceptual framework: Consistent with the conceptualization of this study, a model was developed specifically to study the relationship between psychographic variables, attitude development and credit card usage by cardholders in Malaysia. Figure 1 shows a model of the constructs and hypothesized relationships investigated in the study. Based on the foregoing review of literature, the following hypotheses are postulated:H1. There is a significant association between average credit card debt and cardholders attitude toward credit cards usage across Malaysia.H2. There is a positive relationship between the psychographic variables and the attitude of cardholders in Malaysia.\nResearch methodology: Survey instrument\nAnalysis and discussion: Profile of survey respondents\nResults: To test the first hypothesis, correlation (Pearson's) test was used to test the relationship between average credit card debt and cardholders attitude toward credit cards in Malaysia. From the results of the test (see Table III), it can be seen that there was no significant relationship between the two variables tested. This hypothesis does not find support. A possible explanation for this might be that respondents are very cautious in their spending using credit card or it is also probable that most of use credit card for convenience and promptly settle credit card bills. Further research may be undertaken in future to study the underlying motives.To test the second hypothesis, a multiple regression procedure was used. Attitude as a dependent variable was regressed against the independent variables: life style, self-esteem, time consciousness, peer group pressure, exposure to advertisement and gregariousness. The results of the multiple regression analyses are displayed in Table IV.The results indicate that four of the six relationships are in predicted positive direction but only the estimate for consumers' lifestyle and attitude was significant. Thus, only this variable was supported, while others were not. This suggests that for credit cardholders in Malaysia \"life style,\" was the only variable that influenced the attitude toward credit card spending. The coefficient of multiple determination (R2) was 0.138. This indicates that about 13.8 percent of the total variation in attitude can be explained by the six independent variables. The low R2 implies that the independent variables chosen can account for only a small fraction of the variation in attitude.In addition, due to a high intercorrelation among the independent variables, collinearity diagnostics statistics were performed on each of the variables. Collinearity refers to the situation, where there is a high multiple correlation between two varaibles (Norusis, 1993). In this study, multicollinearity was not shown. The variance inflation factor (VIF) is below 10 for all the independent variables. When VIF is high, there is multicollinearity and instability of b and beta coefficients, which does not apply in the present instance. For diagnostic purpose, the Durbin-Watson value of 1.527 falls within the acceptable range. It shows no autocorrelation problem.\nDiscussion, conclusion and marketing implications: This paper undertakes a study of the factors that are believed to influence attitude toward spending using credit card and the attitude formation toward credit card debt. A model was developed based on review of literature, to identify the factors that affect the attitude of consumers toward credit card spending. However, the results from this study indicate that only about 14 percent of the independent variables under study account for the variation in attitudes. The conclusions that can be drawn are against this backdrop.First, only the dimension of life style appeared to influence attitude while other factors like self-esteem, time consciousness, peer group pressure, exposure to advertisement and gregariousness were not. In another study, time consciousness and gregariousness were found to have an impact on level of card usage (Sharaf, 1999). To this extent, the findings of the present study were different because of the socio-cultural differences across Malaysia. With the improvement in standard of living among Malaysians, the life style also changed. In large cities, there is a proliferation of branded products. Consumers across Malaysia are becoming more brand-conscious and are seeking these products that give them products suiting their lifestyle and thereby satisfaction.A second important finding of this study is that surprisingly self-esteem did not have any impact on the formation of attitude. Past studies on self-esteem observed that consumers' attitude toward spending was influenced by their level of self-esteem. The findings of this study are contrary to the popular belief. Self-esteem did not have any impact on the attitude could be due to the fact that credit card is no longer considered a prestige item to posses. It is now a necessity. Users need credit card for convenience.Finally, behavior was not significantly associated with attitude. There was no significant impact of attitude on credit card debts. A possible explanation for this is that consistent to other findings, credit cards across Malaysia as a convenience product has led to compulsive shopping behavior causing addiction to shopping and heavy debt.Limitations and future directions of research\n",
"role": "user"
},
{
"content": "Purpose: - Despite the spread in usage and ownership of credit cards, few studies have examined its effect on consumer debt in developing nations. The main purpose of this paper is to understand consumers' attitude and spending behavior using credit cards.\nDesign/methodology/approach: - This study is based on a questionnaire survey conducted in Malaysia. Based on an extensive review of literature, a model is developed to identify the psychographic factors that influence the consumer attitudes toward using credit cards.\nFindings: - The paper found support for some of the theoretical expectations and lends support to some of the earlier deviations reported in the literature.\nPractical implications: - The findings are likely to be important to banks and financial institutions issuing credit cards, as they help managers to have a better understanding of cardholders in Malaysia and their attitude and behavior toward usage of credit cards.\nOriginality/value: - This paper makes a valuable contribution given the fact that there is a dearth of empirical studies of this nature focusing on Malaysia.\n",
"role": "assistant"
}
] |
Managing knowledge boundaries for open innovation - lessons from the automotive industry
|
[
"Open innovation",
"Case study",
"Automotive industry",
"Knowledge boundaries"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Firms that pursue an open innovation strategy tend to achieve higher innovative performance (Laursen and Salter, 2006; Parida et al., 2012). Central to open innovation is the emphasis on a search for ideas and knowledge from external actors, which are brought into the firm and integrated into internal innovation processes (Chesbrough, 2003, 2006). There is an unchallenged assumption in the open innovation literature that organizational boundaries become \"porous\" (Laursen and Salter, 2006) or even \"fade\" (Dittrich and Duysters, 2007) simply as a result of adopting this strategy. This raises questions of what organizational boundaries exactly are and how they can be \"opened up\" so that valuable knowledge from the outside can be identified and integrated.\n2. Theoretical background: 2.1 Open innovation in the automotive industry\n3. Methodology: Case studies are particularly strong for studying abstract concepts that are not directly observable, such as the concept of knowledge boundaries in this study. Empirically capturing knowledge boundaries is challenging and requires studying the context of the phenomenon as well as directly interacting with actors who create it (see also Carlile, 2002, 2004). Our ontological stance is that such boundaries only exist because of their construction through powerful actors who shape and reproduce such boundaries through their actions (Giddens, 1984). As such we follow an interpretive stance, and see phenomena such as knowledge transfer across organizational boundaries as socially constructed rather than as objective characteristics with universal application.\n4. Managing knowledge boundaries for open innovation: The open innovation group within the AIN was founded out of the growing recognition of carmakers that new technological impulses are less likely to stem from the traditional boundaries of the automotive supply chain:If the innovation is from our suppliers, I will get it anyway at some point. I have yearly meetings with Bosch, Siemens and so on [...]. They don't need to tell me here what is new because long before it becomes public they have already told me(meeting minutes, Innovation Manager CAR3).\n5. Discussion: By setting up an open innovation competition with a broad search scope that targeted \"distant\" knowledge sources outside the existing automotive supply chain, the members of the open innovation initiative set up by carmakers in Germany tried to reach into unfamiliar knowledge domains. Our study demonstrates the difficulties of integrating external actors outside the automobile supply chain and the distant knowledge that they bring into carmakers' innovation processes. The car manufacturers aimed to have a truly open innovation process (Ili et al., 2010), but found that unacknowledged boundaries to absorbing and using foreign knowledge can frustrate this objective. In this paper, we provide more insight into how such knowledge boundaries function, but also into how gatekeepers such as the innovation managers of different carmakers active in the open innovation initiative straddle these organizational knowledge boundaries. We summarize our main findings with three working propositions that specify the different roles of gatekeepers when they (potentially) mediate between the structural conditions of knowledge boundaries, and the structural consequences in terms of an increased permeability of knowledge boundaries in an open innovation context. The rendering of structural conditions and action in a reciprocal relationship is based in a perspective which conceives of social structure and human agency as interdependent constructs (Giddens, 1984).\n6. Implications for practice: Our study holds some valuable lessons for managers and policy makers. First, it has been found that it is above all interpretative (and only resulting from that political) boundaries, which lead to problems of competently evaluating distant knowledge. One approach to overcome this shortcoming would be to create more interaction opportunities between carmaker representatives and inventors before or shortly after they submit their ideas. The lack of networking opportunities before the actual submission deadline could be seen as a general problem of this innovation competition, as it is also conceivable that inventors could benefit from active interaction with each other, as it is often good-practice in online web communities and idea competitions. Recent research on the supply side of the ideation process suggests that idea creators have an increased need for support in the idea elaboration stage. After an idea is generated it needs to be further refined and developed through constructive feedback and suggestions to help ideators identify ways to improve and expand their ideas (Perry-Smith and Mannucci, 2017).\n7. Conclusion and limitations: For managers as well as for researchers our study suggests that the perception of distance between the knowledge domains of the players involved in knowledge exchange can be altered (to some degree). Allowing distant foreign knowledge to enter an organization and be successfully used requires that the outside knowledge is actively connected to what inside knowledge is available inside an organization. While we offer insights into how knowledge or interpretation boundaries work, and formulate suggestions on how they can be better managed in the open innovation process to increase their permeability, our study has some limitations: first, we were not able to \"objectively\" assess the quality of the submissions for the innovation process. Over the years, however, the number of ideas submitted increased substantially as the competition became better known. Carmakers in Germany have a strong reputation for quality and fairness in dealing with firms they collaborate with and outside parties would be highly motivated to participate (cf. Boudreau et al., 2011). A share of the ideas submitted through the competition will have had the potential to be further developed into useful contributions for carmakers. The likelihood that potentially valuable knowledge was actually not recognized is non-negligible.\n",
"role": "user"
},
{
"content": "Purpose: The rising need to innovate and obtain knowledge from more distant knowledge sources calls for new innovation strategies and a better integration of other external actors who lie outside the traditional automotive supply chain. Such an open innovation strategy challenges organizational boundaries both on the firm and supply chain level, yet our understanding of the functioning of such boundaries and how they can be managed to allow for purposive knowledge flows is limited. The paper aims to discuss these issues.\nDesign/methodology/approach: In a longitudinal case study, the authors trace the development of the first open innovation network in the German automotive industry over a period of five years based on archival data, semi-structured interviews, and field observations.\nFindings: While the automotive industry is advanced in collaborating with suppliers for innovation, routines for assessing and integrating ideas from sources outside the supply chain are still underdeveloped. The authors show which current knowledge boundaries pose obstacles for open innovation initiatives in this industry, and how they could be mediated through the involvement of gatekeepers.\nOriginality/value: The authors challenge and clarify the notion of the \"permeability of organizational boundaries\" in the open innovation literature and investigate the role of gatekeepers for open innovation.\n",
"role": "assistant"
}
] |
Clustering Spanish alcoholic beverage shoppers to focus marketing strategies
|
[
"Alcoholic drinks",
"Consumer behaviour",
"Clustering",
"Spending behaviour",
"Alcoholic beverage shoppers",
"Beverage industry"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Over the past few decades, drinks companies have enjoyed unprecedented commercial success. However, factors such as the changing climatic conditions, economic recession and consumers becoming health-conscious, are forcing the beverage industry to rethink their business methods in order to strengthen their brands in a market that has become even more demanding and globalised (Kao, 2015).\n2. Literature review: This study begins with using an approach based on that by the Association for Consumer Research (ACR, 2018), specifically, a multidisciplinary focus of consumer research (McFadden, 1986; Ashok et al., 2002), in which the areas of marketing, psychology and sociology are coherently integrated for integral assessment of consumer behaviour.\n3. Methodology: 3.1 Participants and procedure\n4. Results: 4.1 Results obtained in the hypothesis contrasting\n5. Discussion: In the field of alcoholic beverages, different factors may influence customers' buying decisions; factors that companies have to take into account when designing their marketing strategies. Therefore, taking into account the significant results obtained in our research and the cluster analysis done, we show below the characteristics of the seven groups of buyers:\n6. Conclusions, implications, limitations and further research: Taking the results of this research into account, we can conclude that Spanish shoppers buy alcoholic drinks depending on certain individual and environmental factors.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is twofold: to analyse the influence of both individual and environmental factors in order to explain alcoholic spending and to identify different groups of alcoholic beverage shoppers.\nDesign/methodology/approach: For the research, an online questionnaire was distributed among 1,023 Spanish participants of age more than 18 years. Multivariance techniques were used for analysis, and the results show significant influence of specific motivational domains, brand awareness, perceived culture, socio-demographic variables and purchasing environment. With these results, a cluster analysis was carried out identifying seven groups of alcoholic beverage shoppers.\nFindings: This research confirmed the influence of both individual and environmental factors, and the authors have identified seven different groups of alcoholic beverage shoppers: prosperous, social, non-conformist, postmodern, communities, controllers and planners. This led to the consideration of various management implications, which would boost the success of the beverage industry.\nOriginality/value: The originality of this research focuses on identifying different cluster of shoppers who purchase alcoholic beverages and on revealing the characteristics of each identified cluster. Based on the exact profile proposed, the alcoholic beverage industry should design more appropriate marketing strategies to achieve competitive advantage and to reinforce purchasing because in today's markets, there are fewer and fewer situations where a mass marketing approach is feasible.\n",
"role": "assistant"
}
] |
Ethnic identity, consumption of cultural apparel, and self-perceptions of ethnic consumers
|
[
"Ethnic groups",
"Consumer behaviour",
"National cultures",
"Clothing",
"United States of America"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The US Bureau of Census (2000a, b) population projections reveal that the racial and ethnic composition of the USA will change significantly by 2050. Currently, Hispanics and African Americans comprise over a quarter of the total population of the USA, and this number is predicted to increase substantially in the future. The Hispanic population is expected to double from 11.8 percent of the total population in the year 2000, to 24.3 percent of the total population in the year 2050 (US Bureau of Census, 2000a, b). Similarly, the African-American population is expected to increase from 12.8 percent of the total population in the year 2000, to 14.7 percent of the total population in the year 2050. The Asian and Pacific Islander population in the US is also expected to triple from 11.3 million (4.1 percent of the population) in 2000, to 37.6 million (9.3 percent of the total population) in 2050. The changing pattern in the compositional structure of projected US population is a concern and challenge to retailers, since it has important implications for understanding the core needs of different ethnic subcultures, and the design of appropriate product and marketing strategies targeted to meet these needs.Apart from the population figures, the purchasing power of the different ethnic subcultures forms another important consideration for retailers and marketers. The combined buying power of Hispanics, African Americans, and Asians, currently exceeds one-trillion dollars, and is expected to rise further (Hammer and Skolnicki, 2005). Retailers including JCPenney, Sears Roebuck (now acquired by Kmart), Target, Walmart and Kmart are now focusing on increasing their market share by implementing strategies targeted to ethnic subcultures, such as targeting merchandise assortments in stores, introducing ethnic brands, and offering specialty products such as apparel and jewelry for cultural celebrations. For example, in 2003, Kmart launched a Hispanic-directed apparel brand, called Thalia Sodi Collection, which generated more that $17 million in 39 weeks (Alexander, 2005). Sears is similarly offering apparel product lines that are designed for specific ethnic subcultures, such as Azucar Bella, a Hispanic-directed brand, and Russell Kemp, an apparel brand targeting African American consumers (Howell, 2004).In the context of apparel retail, the success of such endeavors largely depends on the retailers' ability to understand why ethnic consumers' \"choose to\" or \"choose not to\" wear apparel products associated with their ethnic culture. Specifically, a key question that has been infrequently addressed in past literature relating to ethnic consumer behavior is: \"do ethnic consumers attribute positive (or negative) emotions and meanings to the consumption of cultural (ethnic/ethnic-inspired) apparel, and what do these attributions depend upon?\" This study explores this question through survey research conducted among four ethnic subcultures in the USA. The study specifically investigates if consumers' use of cultural indicators in dress, and attribution of emotions and meanings to this usage, is predicted by their strength of ethnic identification. The study also examines whether the consumption of cultural apparel, operationally defined as the usage of ethnic/ethnic-inspired apparel, mediates the effect of strength of ethnic identification on consumers' internal attribution of emotions and meanings related to cultural apparel consumption.\nReview of literature: Individuals within the US culture are surrounded by a context that can be defined as postmodern and multicultural. Both terms signify a pluralism of meanings; a multiplicity of individual meanings in the case of the former; and a pluralism of shared meanings in the case of the latter. Kaiser (1993, p. 43) advances that clothes are rooted in successive contexts, which range on a continuum from \"micro\" to \"macro\" dimensions. Viewed from the micro perspective, which focuses on individual perceptions and behavior, an individual's attempt to manage his or her appearance involves \"negotiating identities\" through a \"cultural dialogue or struggle for meaning\" (Kaiser, 1993, p. 43).Based in the symbolic interactionist perspective, Goffman (1959) presented a compelling argument that linked individual identity, social interactions, and the physical setting through a dramaturgical performance of one's self in the front, back and outside stage settings. The actor or individual uses this performance as a medium to communicate information about the self and manage impressions that others receive based on implicit and explicit goals. This exchange of information facilitates the development of identity, and particularly social identity or \"collective representation\" in the front stage performance of the individual. In constructing this social or collective representation, the individual may use variety of verbal and non-verbal communicative cues such as dress in the process of tailoring an appropriate representation of the self to specific groups of audiences. This dialogue in the construction of identity through dress is particularly visible among individuals belonging to ethnic minority groups operating within the broader context of the mainstream American culture. These individuals often encounter two distinct groups of audiences associated with the ethnic and the mainstream cultures with differing expectations of \"appropriate self-presentation.\"An ethnic minority group has been defined as a subcultural group within a dominant culture that has a distinct cultural background, and is differentiated from the dominant culture through externally visible characteristics, which may be physical or cultural (Forney, 1981). Dress is one such externally visible characteristic, which in the case of members of ethnic subcultures can serve two functions: to visually indicate subcultural ethnic affiliation or membership; or to visually integrate the individual to the mainstream American culture, thus reflecting acculturation or homogenization of values and of external characteristics. Forney posits that individuals' retention of ethnic identity through elements of dress is related to their level of assimilation into the mainstream culture.Ethnic identity and consumer behavior\nMethod: Sample and data collection\nData analysis and results: Preliminary analyses\nDiscussion: The broad purpose of this research was to examine the relationship between ethnic identity, consumption of cultural apparel and self-attributions of emotions and meanings among consumers belonging to ethnic subcultures in the USA. Based on the review of literature, it was hypothesized that the construct of strength of ethnic identification will predict the consumption of cultural apparel among ethnic consumers (H1) and the extent to which they attribute positively/negatively valenced internal emotions and meanings to this consumption (H2 and H3). The results of this study revealed support for these hypotheses.The support of H1 is consistent with the results of other studies which have found that consumers' strength of ethnic identification significantly influences their consumption of ethnic and ethnic-inspired apparel (Forney and Rabolt, 1985-1986; Kim and Arthur, 2003), ethnic foods and drinks (Chung and Fischer, 1999; Laroche et al., 1998; Xu et al., 2004), and other ethnic products such as movies and music (Xu et al., 2004). The significant positive relationship between strength of ethnic identification and consumption of cultural apparel is also consistent with the directional effects that have been found in previous studies, where higher levels of ethnic identification have correlated with increased consumption of ethnic foods (Laroche et al., 1998; Xu et al., 2004), traditional ethnic dress (Forney and Rabolt, 1985-1986) and ethnic-inspired contemporary apparel (Kim and Arthur, 2003). Hence, this study reinforces that the construct of strength of ethnic identification is a significant predictor of culturally-related consumption patterns of ethnic consumers; and provides further empirical evidence that the construct is also useful in predicting ethnic consumers' attributional responses associated with culturally-related consumption.Based on the self-perception theory (Bem, 1972), this study hypothesized that the consumption of cultural apparel will directly predict consumers' attributions of emotions and meanings to culturally-related consumption (H4a and H5a). The support of these hypotheses demonstrates that the findings are consistent with the self-perception theory. Further, the study also proposed that consumption of cultural apparel will mediate the influence of strength of ethnic identification on the attributional responses of ethnic consumers (H4b and H5b). The test of mediation revealed that consumption of cultural apparel perfectly mediates the influence of strength of ethnic identification on consumers' attributions of emotions to culturally-related consumption; and partially mediates this influence on consumers' attributions of meanings. This finding has important implications that are discussed in the subsequent section.\nImplications: As revealed through the review of literature, few studies have examined the attributional responses of ethnic consumers, and the meanings and motivations related to their consumption of cultural products. This understanding provides valuable insights that can direct and channel product development, merchandising and marketing efforts to better meet the emotional needs and preferences of these consumers. However, it is important to mention that the sample used in this study does not consist of a representative sample drawn randomly from ethnic students enrolled at the university. Hence, caution should be applied in extending the results based on this sample to other populations.This research has revealed that the consumers' strength of ethnic identification influences their consumption of cultural apparel, and their attributions of internal emotions and meanings to this usage. Previous research has partly supported this finding in different geographical locations such as the San Francisco Bay Area (Forney and Rabolt, 1985-1986) and Hawaii (Kim and Arthur, 2003), both of which reflect areas where ethnic groups tend to concentrate. In the current study, the participants represented students at a Midwestern University, which demonstrates that the influence of strength of ethnic identification as a predictor of consumption patterns of ethnic consumers is not restricted to areas of ethnic group concentration.A recent qualitative study conducted among Scottish-Americans (Crane et al., 2004) revealed an interesting finding with respect to the influence of strength of ethnic identification on consumption of cultural apparel. In this study, the role of dress in self-symbolizing ethnic identity increased with greater salience of the Scottish identity; however, this role diminished with increased feelings of completeness in the Scottish identity. This finding suggests that the extent of completeness experienced by an individual with respect to ethnic identity may moderate the influence of strength of ethnic identification on consumption of cultural apparel. While some of the findings of the interview-based study are consistent with the current results, further research is needed to examine the moderating influence of sense of completeness in ethnic identity.The current study also demonstrated that consumption of cultural apparel mediates (perfectly and partially) the influence of ethnic identification on consumers' attributional responses. This empirical finding in a convenience sample of college students is of value to retailers since it could provide an alternative mechanism for predicting consumer responses to cultural products based on past consumption patterns for such products, since data regarding consumers' strength of ethnic identification may not be easily available. However, further research is needed to validate this finding in a larger representative sample of consumers. Future research could also benefit by investigating whether the consumption of a specific cultural product category predicts attributional responses toward other cultural product categories. For example, does consumption of cultural apparel predict consumer attributions associated with the consumption ethnic foods or vice versa? Answers to these questions can help retailers in predicting consumer response toward new cultural products by using data on patterns of consumption for existing cultural products.The sample in this study scored high in ethnic identity, attribution of internal meanings and internal emotions, but was at the midpoint of the scale on consumption of ethnic apparel. This fact could suggest that there is an untapped market for cultural apparel that may be important for product developers and retailers to consider. The results of this study also revealed that consumption of cultural apparel is greater during ethnic celebrations, as compared to casual occasions and non-ethnic special occasions. This finding could have implications for new product development efforts targeting ethnic consumers. Currently retailers such as JCPenney stock jewelry and other products that cater to specific Hispanic celebrations, such as \"Quinceaneras.\" Retailers may also benefit by directing such efforts to other ethnic subcultures based on the ethnic composition of the neighborhood. Further, festival promotions, gift coupons and discounts can also be used to attract consumers' attention to such retail offerings during specific festivals.Beyond these practical implications, this study also offers important implications related to the symbolic function of clothing consumption in an increasingly multicultural society. In a consumption-oriented historical account of Jewish acculturation, Heinze (1990) articulates that consumption of goods serves the important function of bridging gaps between cultures, contributing toward the construction of meaningful cultural identities. His narrative emphasized the role of consumption, particularly the adoption of mainstream American consumption patterns in the immigrant acculturation process, a theme that has received increasing attention in subsequent works (Penaloza, 1994; Oswald, 1999). Oswald examined the complicated nature of symbolic consumption among Haitian immigrants in the USA and its impact on the construction of the Haitian-American identity. According to the author, this identity is an unstable construct that necessitates the process of \"culture swapping,\" wherein the semiotic consumption of goods associated with both \"home and host cultures,\" plays a significant role in identity formation and negotiation. This symbolic consumption often externalizes overlapping cultural contexts and group memberships, and illustrates the concept of \"straddling the boundaries between several cultures at once\" (Oswald, p. 311). Consumption of ethnic-inspired contemporary apparel represents an example of such a negotiation.Cultural products such as apparel have also been found to resonate with the US mainstream consumer segment, called \"cultural creatives\" (Littrell and Dickson, 1999; Ray and Anderson, 2000). Cultural creatives constitute one-fourth of US adults and are mostly college-educated women (60 percent) who are in their 40s (Ray and Anderson). They are also globally concerned consumers who value the integrity and authenticity of cultural products (Ray and Anderson). While this consumer segment prefers culture-specific apparel, researchers have also identified another segment with preferences for pancultural apparel, integrating aesthetic elements from multiple cultures (Littrell et al., 1999). Future research needs to be conducted to comparatively examine the self-attributions of emotions and meanings in wearing culture-specific and pancultural apparel among both mainstream and ethnic consumers.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to investigate whether ethnic consumers' consumption of cultural apparel, and attributional responses related to their consumption, is predicted by their strength of ethnic identification. The study also examined whether the consumption of cultural apparel mediates the influence of strength of ethnic identification on consumers' attributional responses.\nDesign/methodology/approach: - Internet survey research was the chosen methodology for this study. A convenience sample of 106 research participants from four ethnic subcultures in the USA were recruited.\nFindings: - Regression analyses revealed that strength of ethnic identification was a significant predictor of cultural apparel consumption and attribution of emotions and meanings to the consumption. Further, consumption of cultural apparel perfectly mediated the influence of strength of ethnic identification on consumers' attributions of emotions, and partially mediated this influence on consumers' attributions of meanings.\nResearch limitations/implications: - The study provides valuable implications for channeling apparel product development, merchandising, and retailing to better meet the emotional needs and preferences of ethnic consumers. The main limitation of this study is the use of a non-representative sample.\nOriginality/value: - Prior research on ethnic consumers' motivations in consuming cultural products is limited. This study fills this gap in the literature at a time when large retailers are seeking to attract ethnic consumers through culturally targeted apparel products and brands.\n",
"role": "assistant"
}
] |
Managerial behavior and performance in China, the UK, and the USA
|
[
"Leadership development",
"Managerial behaviour",
"Cross-cultural studies",
"High performing managers",
"Managing across cultures",
"Multisource feedback"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: This study examines managerial behaviors displayed in the workplace and explores the relationship between these behaviors and overall managerial job performance. Multiple studies have suggested that successful managerial behaviors may not be the same in various countries (Clark et al., 2016; House et al., 2004). This study extends prior research on behaviors viewed as necessary for success in various cultural contexts, by providing a more granular examination of managerial behavior. These differences may in part be due to the phenomenon of cultural relativism, originally defined by cultural anthropologists. Cultural relativism holds that behaviors must be examined in the context of the culture in which they occur; behavior that is completely acceptable in one culture may be viewed as intolerable in another culture. Critical commentary on such relativist evaluation suggests that one must also take into account moral and ethical standards of human behavior, which may be overlooked in a simplistic application of cultural relativism (Brown, 2008). Understanding which managerial behaviors are associated with high job performance may help to focus the actions of these managers and enhance the specificity of selection, coaching and leadership development initiatives.\n2. Literature review: 2.1 Managerial behavior - job performance\n3. Methods: 3.1 Instrument: the PROFILOR(r) for managers - description\n4. Results: 4.1 Exploratory step-wise regressions\n5. Discussion: 5.1 Patterns observed in the findings\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study was determine which, if any, managerial behaviors were associated with high managerial job performance in three selected countries, China, the UK and the USA. After identifying which behaviors were associated with high managerial job performance, the study then compared the results from each country to identify behaviors that were unique to a country and those that recurred across multiple countries.\nDesign/methodology/approach: This study draws from an archival database of multisource (360deg) feedback rating managerial job performance. Supervisors' ratings on 23,877 national managers from China, the UK and the USA were examined using simultaneous stepwise regression analysis.\nFindings: This study found that there were unique behaviors associated with high managerial job performance in each country examined. Additionally, the study found that were also behaviors associated with high managerial performance shared between all three countries.\nOriginality/value: This study offers another insight to the unique-versus-universal managerial behaviors debate in leadership development literature. Existing literature offers mixed messages regarding universal or unique behaviors across countries. Understanding which key managerial behaviors are associated with perceived high managerial job performance in each country may help to focus the development of these managers and enhance the specificity of selection, coaching, and training initiatives.\n",
"role": "assistant"
}
] |
Country of origin and ethnocentrism in the context of lateral, upward and downward migration
|
[
"Country of origin",
"International marketing",
"Ethnocentrism",
"Immigrant",
"Migration pattern"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: An emerging stream of research (e.g. Davies and Fitchett, 2004; Penaloza, 1994; Ustuner and Holt, 2007) challenges the assumption of unidimensional identity in consumer ethnocentrism (CE) and country of origin (COO) literatures. This nascent literature goes beyond static and homogeneous depictions of self and identity and extends the notions of CE and COO to important, previously unheeded contexts. These include ethnic, immigrant, or otherwise bicultural populations, where self and identity possess orthogonal and co-existing anchors. For example, Zolfagharian and Sun (2010) and Zolfagharian et al. (2014) draw on first-generation immigrants to explore the multiplicity of ethnocentric tendencies (i.e. toward multiple countries) and expose situations where consumers identify with both the COO of the product and the country where it is sold. Such research is increasingly relevant to studies of marketing and consumption as the social fabric around the world is exponentially growing more diverse and multicultural (Klein et al., 1998; Mihailovich, 2006). Investigating familiar constructs in new, multicultural settings is key to keeping past and existing knowledge viable (Maheswaran and Shavitt, 2000; Miller, 1997).\nLiterature review: COO and CE\nMethod and analysis: Sampling\nResults: Factor analysis and validation\nDiscussion: On the one hand, CE and COO literatures have traditionally viewed consumer identity as unidimensional and homogeneous entities and consequently run the risk of irrelevance given the increasingly multicultural composition of twenty-first century societies (Klein et al., 1998; Mihailovich, 2006; Penaloza, 1994). The present study contributes to a critical stream of research that goes beyond static conceptualizations of the self and identity and extends the notions of CE and COO to the multicultural immigrant communities, where the consumer has access to a mixture and multiplicity of identity anchors. On the other hand, research on immigrants continues to revolve around those who migrate from developing to developed countries albeit the paucity of research on other immigration patterns (Gindling, 2009; Shrestha, 2011). The present study utilizes a sample that consists of four diverse immigrant groups with distinctive immigration patterns.\nManagerial implications: The findings of this study are in line with previous work that explains how immigrants use a variety of home country products and traditions in their host country. Oswald (1999), Mehta and Belk (1991), and Penaloza (1994) all find that immigrants, alongside their assimilation experiences, consume products and perform customs from their respective home countries in an attempt to maintain their mixed identities comprised of a blend of home and host cultures. The findings of this study are also in line with previous empirical studies from sociology (e.g. Rockquemore and Brunsma, 2002) and marketing (e.g. Zolfagharian et al., 2014; Zolfagharian and Sun, 2010) that find bicultural consumers to possess mixed identities and to draw upon these identities when evaluating different products.\nLimitations and directions for future research: Recently, the relevance of existing marketing and consumer knowledge in multicultural societies of the twenty-first century has been questioned (e.g. Maheswaran and Shavitt, 2000; Miller, 1997; Yamada and Singelis, 1999). For example, existing CE scales (e.g. Shimp and Sharma, 1987) as well as COO scales (e.g. Martin and Eroglu, 1993) limit the number of countries the respondent can identify with to only one and, therefore, are not appropriate for use with multicultural consumers (Zolfagharian et al., 2014). We used a modified version of the CETSCALE scale (Shimp and Sharma, 1987) that Zolfagharian et al. (2014) adapted to the context of immigrant consumers. A fruitful avenue for research is rooted in this limitation. Future studies should revisit the content and dimensionality of CE in multicultural settings and consider developing a new measurement scale that, from the onset, incorporates the possibility of mixed identities. As a corollary to this direction, researchers should address the issue of measurement invariance in similar future studies, which incorporate multiple immigrant groups characterized by different cultural backgrounds and investigate the factor structure of their measurement scales.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the effects of consumer ethnocentrism and country of origin across different immigrant communities.\nDesign/methodology/approach: A survey was used to collect data from immigrants in the USA and Mexico.\nFindings: For immigrants with high levels of ethnocentrism, the bias for home and host country products interacts with the country of origin effect and creates multiple scenarios where the two effects move in the same or opposite directions. For immigrants with low levels of ethnocentrism, on the other hand, the country of origin effect alone is salient.\nResearch limitations/implications: The authors used a modified version of CETSCALE. Future research should revisit the content and dimensionality of consumer ethnocentrism in immigrant and other multicultural settings.\nPractical implications: Both scholars and practitioners should exercise caution when working with ethnocentrism and country of origin as today's societies are increasingly multicultural, which requires major modifications to the theories and tools.\nSocial implications: Similar to ways in which the US Census Bureau enabled multicultural consumers to assert their mixed identities, scholarly and business circles should embrace multiculturalism and empower immigrants.\nOriginality/value: Previous studies of consumer ethnocentrism and country of origin in multicultural contexts have restricted themselves to only one pattern of migration: consumers who move from developing to developed countries. The paper addresses this limitation by investigating various patterns of migration (including lateral, upward and downward) in multiple first-generation immigrant communities in two countries.\n",
"role": "assistant"
}
] |
Estimated numbers of new TB cases in Patan, a city in Nepal
|
[
"Nepal",
"Patients",
"Targets",
"Communicable diseases",
"Health services"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Since April 1998, a public-private partnership for TB control has been operating throughout Patan, a city of 200,000 people that lies immediately to the south of Kathmandu, the capital of Nepal. The partnership provides diagnosis and treatment for TB based on the WHO- and IUATLD-recommended DOTS strategy (Enarson et al., 2000; WHO, 2003). It includes diagnosis provided by private practitioners (PPs), direct observation of treatment and late patient tracing by NGOs, and training and drugs by the Nepal National TB Programme (NTP). From the start of the public-private partnership until 2004, over 1,500 TB patients were registered in the five DOTS centres in the city. Over 90 per cent of these patients were successfully treated, while 1.2 per cent defaulted, 1.8 per cent failed to improve and 2.2 per cent died while under treatment. About 15 per cent of all these patients were directly referred by private practitioners to one of the DOTS centres. Further details of the partnership are given elsewhere (Hurtig et al., 2002; Newell et al., 2004). Awareness of TB and DOTS has been promoted in the Patan community over several years by staff and volunteers of Yala Urban Health Programme (YUHP). An external evaluation conducted in 2000 found that 54 per cent of households were aware of YUHP's activities, and that they received an average of six health education visits per year (unpublished data).Despite these encouraging results and a great deal of effort invested in advocacy and publicity about TB, DOTS treatment and the public-private partnership, the apparent case-finding rate in the city was less than 50 per cent (Newell et al., 2004). This case-finding rate was derived from an Annual Risk of Tuberculosis Infection (ARTI) of 4 per cent, the rate generally used for cities in Nepal. In contrast, the national case-finding rate in Nepal increased from 50 per cent in 1996 to 69 per cent in 2001 (Nepal National TB Programme, 2000/2001), based on a national ARTI of 1.8 per cent. The national ARTI figure was based on several tuberculin surveys, while the urban figure has its origin in two small-scale tuberculin surveys carried out in 1991 in the most deprived areas of Kathmandu and Patan (JICA, 1987-1993).In Patan in the year commencing March 2000, 192 new sputum positive patients were registered with government and NGO clinics. A survey of private practitioners suggested that during the same period they diagnosed approximately 120 patients (Newell et al., 2004), approximately 90 of who were referred to a government or NGO clinic, and presumably registered there. A survey of private laboratories who provide diagnostic services to private practitioners, and a survey of pharmacies who provide anti TB drugs for patients of private practitioners, suggested that the numbers of patients private practitioners claim to have diagnosed was approximately correct (Newell et al., 2004). An unpublished survey of non-allopathic providers (paramedics, ayuvedics, homeopaths and traditional healers) in Patan indicated that none of these attempted to treat TB, but instead referred TB suspects to the NTP.Based on the usually accepted ARTI of 4 per cent in urban areas, the expected number of new smear-positive cases in Patan's 200,000 population was approximately 400 per year. This number was not compatible with the number of patients observed: at least 400-(192+120-90)=178 patients were not accounted for.We needed to try to understand this discrepancy between observed and expected numbers of new sputum positive patients in Patan, as it threw into doubt targets for TB control in Patan, and more generally in Nepali cities and elsewhere in Asia. However, it was not feasible on cost grounds to perform either an ARTI survey or a full TB prevalence survey in Patan. We therefore decided to perform a household survey in areas of Patan to try to determine the incidence and prevalence of new sputum-positive TB there.\nMethods: In order to estimate the number of new TB cases in Patan, a large household survey was undertaken in a central area of the city. The intention was to obtain information about TB patients regardless of where, or by whom, they were diagnosed or treated. The survey was conducted in two phases.Study area\nResults: Possible and probable TB cases identified through household survey\nDiscussion: Our estimated number of new smear-positive TB cases (31 cases) is equivalent to an incidence of 85 cases per 100,000 population per year. Even the upper limit of the 95 per cent CI of this estimate (125 per 100,000) is much lower than the incidence of 200 per 100,000 that is implied by using an ARTI of 4 per cent, indicating that our low estimate is not simply a consequence of inadequate sample size. It is, however, similar to the incidence of 100 per 100,000 implied from the 2 per cent national estimate of ARTI.The CHVs who conducted the survey believed that no households had been missed. There were no independent, contemporary records against which this claim could be assessed. However, surveys conducted in the study area during 1998 (ward 19) and 2000 (wards 6, 9, 11, 12 and 17) reported a total of 6,462 households and a population of 31,797 in the six wards. The figures obtained during the present survey are 24 per cent and 16 per cent higher, respectively. Natural growth and net migration must account for some of this increase, but we believe the present survey achieved a very high coverage of target households.In the survey only four out of 41 patients identified (10 per cent) were treated privately, and none of these were smear-positive cases. The main assumption in the method used to estimate the number of new TB cases is that the level of completeness of the survey data was similar for both DOTS-treated and privately-treated TB patients. This may well not be true, since one-reason patients may choose treatment in the private sector is to conceal their disease.In assessing the validity of this assumption, we can draw on the findings of surveys in Patan designed to estimate the numbers of TB patients being managed by private practitioners. An ARTI of 4 per cent implies 74 smear positive patients in our study population of 36,918. After allowing for under-estimation during our survey, we estimated that there were actually only 31 smear positive patients in the study area. If the 4 per cent ARTI figure was correct, our study area should have contained 74-31=43 smear positive patients a year being treated by private practitioners and who were not detected by our survey. Given the severity of TB symptoms and relatively good access to treatment, it is unlikely that people are dying of TB without seeking treatment from either the public or private sector.In fact, private practitioners reported that they saw approximately ten smear-positive patients from Patan per month, of whom three-quarters were referred to DOTS centres (Newell et al., 2004). For our study population this equates to only six smear-positive patients per year being treated exclusively in the private sector. Since our study area is one of the poorer parts of Patan, even this figure may be an over-estimate. While we do not claim that this figure is very accurate, it is so much lower than the 43 smear positive patients predicted by the 4 per cent ARTI figure that it strengthens the case for questioning the validity of the 4 per cent ARTI. If one assumes that the ARTI in Patan is indeed closer to 2 per cent than 4 per cent, this may be explained by the long history of low-cost TB treatment and defaulter-tracing that has been available through Patan Hospital since 1984, and through its predecessor since the 1970s.A lot of effort has been invested in Patan to raise awareness about TB and its treatment and to reduce stigma. It was therefore disappointing that as many as 28 per cent of households with a DOTS-treated TB patient did not disclose the disease during the household interview. The respondents were usually heads of households, but if the head was absent other members of the household were accepted. We do not know how many of the respondents who denied knowledge of a TB patient in their household were keeping a family secret, and how many genuinely did not know about the patient. The TB patients from these households are being discreetly contacted to learn more about their reasons for concealing the disease.\nConclusion: Two main conclusions arise from this study. The first is that using the ARTI may lead to misleadingly high targets for urban TB control, a conclusion that concurs with other published views (Rieder, 1995; Borgdorff, 2002). Unrealistically high targets may cause TB workers to become demoralised, and useful strategies to be abandoned. However, there are no clear guidelines on how else to create realistic targets - and since such targets are key to the WHO/IUATLD DOTS strategy, further work is needed to identify better ways of setting them. One option is to base targets on up-to-date tuberculin surveys from selected urban communities. The second conclusion of this study is that using a household survey to estimate TB incidence and prevalence can create as many questions as it answers. We suggest that in the interim, before better methods are developed, targets for urban TB control in Nepal and elsewhere should be revised by experts in the field.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of the paper is to show that, despite comprehensive coverage of services for TB provided by a public-private partnership for TB control in Patan, a city in Nepal, case finding is low, compared with the target based on an Annual Risk of Tuberculosis Infection (ARTI) of 4 per cent. Doubts have been raised as to the appropriateness of the target. The objective of the study was to estimate the number of new TB cases occurring in Patan, to assess whether the target was appropriate.\nDesign/methodology/approach: - In the paper doorstep interviews were conducted with all households in the study area, followed by in-depth interviews of households with possible or probable TB cases. The survey findings were validated against the patient registers of the five DOTS centres in Patan.\nFindings: - The paper finds that, among the study population of 36,918, the household survey identified 17 smear-positive TB cases (none privately-treated) and 24 smear-negative/extra-pulmonary cases (including four privately-treated cases). Validation against the DOTS patient registers indicated that the survey was 54 per cent complete. After adjusting for incompleteness, the number of smear-positive cases in the study area was estimated as 31, equivalent to an incidence of 85 smear positives per 100,000 population and an ARTI of 1.7 per cent.\nOriginality/value: - The paper shows that using the ARTI may lead to misleadingly high targets for urban TB control. Unrealistically high targets may cause TB workers to become demoralised, and useful strategies to be abandoned. Therefore, further work is needed to identify better ways of setting targets.\n",
"role": "assistant"
}
] |
Rethinking share-of-wallet at the bottom of the pyramid: using financial diaries to observe monthly category trade-offs
|
[
"Consumer behaviour",
"Bottom of the pyramid",
"Share of wallet",
"Emerging market",
"Financial diaries"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Over four billion people constitute the \"bottom of the pyramid\" (BoP). This population segment is usually cited as those living on less than US$2.50 per day, although scholars have widely debated poverty benchmarking (Prahalad and Hart, 2004; London et al., 2010; Simanis, 2012; Chen and Ravallion, 2013). While BoP research has grown in the twenty-first century (Kolk et al., 2014), Chikweche et al. (2012) correctly stated that the study of BoP consumer behaviour is in its infancy, and most studies have focused on Latin America and Asia - largely ignoring 800 million BoP consumers in Africa. This omission is significant, as Africa is expected to home most of the world's working age population by 2050 (Melorose et al., 2015).\nLiterature review: The global BoP\nMethodology: To measure monthly SoW fluctuations, a financial diary panel was used. Although trends in poverty and inequality during the post-apartheid era have been the subject of intensive analysis in South Africa (Finn et al., 2014), the surveys from sources like the National Income and Expenditure Dynamics Survey (NIDS) and Statistics South Africa (StatsSA) have minimally observed income and expenditure patterns on a bi-annual basis (StatsSA, 2013, 2015; NIDS, 2016). Table II (below) is a summary of the comparative findings of Finn et al. (2014) on SoW over three sets of data collection - over nearly two decades.\nSample: In South Africa and other emerging economies, research is often constrained by lack of the BoP population data through which to create accurate sample frames (Scott and Vigar-Ellis, 2014; Rensburg and Botha, 2014; Ijabadeniyi et al., 2015; Nailer et al., 2015; Cant and Wild, 2016). Once the provinces and metropolitan areas were chosen, the study sample of 80 households was selected using a non-probability quota sampling technique (Malhotra, 2010). Researchers note that, when a sample frame is not available and research is exploratory, non-probability sampling is justified (Schillewaert et al., 1998; Tustin et al., 2005; Baker et al., 2013; Ortinau, 2016). However, to increase the generalisability of the findings, quota sampling was used (de Rada and Martin, 2014; Babin and Zikmund, 2015).\nResults: Income analysis\nProposing a metric for category share of spending: As SoW is traditionally measured at a firm level, the metric is useful for understanding profitability through brand extension and multi-branding strategy. The narrow definition of SoW, however, does not adequately define the true share of category from a consumer perspective. This study highlights an apparent need for a household-orientated measure of SoW that is more linked to category share of expenditure (or spending). As category trade-off was observed to play such a significant role in BoP consumer choice, a measure would help with understanding the potential implications of category SoS (CaSoS). To date, no such category-related metric exists. A proposed means to measure CaSoS (this would need further testing) would be to track average share of expenditure of a single category over time, and then measure an index of CaSoS relative to the mean:\nDiscussion and implications: From this study, several practical and theoretical implications emerge for understanding the BoP household behaviour. Theoretically, the income and substitution effects both show how shifts in income and price can have an impact on choice. This study shows that inconsistent monthly shifts in income and expenditure create unpredictable changes in expenditure patterns. Whilst this does not fully undermine the income effect in particular (because there are standing assumptions), the findings emphasise that inconsistency in the BoP household income and expenditure cannot easily be encapsulated by the microeconomic theory.\nFuture research: Whilst the methodology was rigorous for a study of this nature, sample location and size could have been increased to incorporate more data, but time and resource limitations excluded this possibility. Similarly, potential fieldworker and participant bias are common with primary data collection (Malhotra, 2010; Belk et al., 2013; Aguirre and Hyman, 2016; Roulston and Martinez, 2016). Although robust and accepted by the research community since the 1900s (Johnson and Bytheway, 2001; Gershuny, 2011), financial diaries do not have as strong a tradition as standard research questionnaires, and really only became popular in the 1990s (Kent, 1993; Collins et al., 2009; Taylor and Lynch, 2016; Morduch and Schneider, 2017a). Part of the reason for the lack of use is that diaries are costly to administer and rely on strong participation from both the researcher and household. Diaries have standard completion accuracy issues and are subject to omissions (Collins et al., 2009; Alia et al., 2015). The cost means that diary samples are often smaller and used for exploratory purposes (Taylor and Lynch, 2016). Timing and reporting bias are also a challenge, although other forms of data collection are not exempt from such bias (Collins et al., 2009; Taylor and Lynch, 2016). Finally, it has been observed that participants are more likely to under-report expenditure when assessment is more regular. This under-reporting was observed with StatsSA using a weekly financial diary methodology to assess survey samples income and expenditure over a single month (Finn et al., 2014).\n",
"role": "user"
},
{
"content": "Purpose: This study provides empirical evidence for the impact that income and expenditure fluctuations have on understanding the fundamentals of BoP household share-of-wallet in South Africa.\nDesign/methodology/approach: The study used a longitudinal financial diary methodology to record household income in 80 households (in four different geographic areas) over four monthly waves.\nFindings: The study showed the lack of reliability of aggregated population income and expenditure surveys when understanding the specific behaviour of BoP households. The study concludes that major category trade-offs occur on a monthly basis, and that these trade-offs directly impact our fundamental understanding BoP SoW.\nOriginality/value: While the BoP consumer theory is developing (especially in the last decade), most of the theory is focused on development and business strategy. Empirically based consumer theory is noticeably lacking, given that the BoP is the largest population segment on earth. In addition, research is largely absent of highly rigourous and in-depth quantification of consumer SoW behaviour. This study contributes to the BoP theory by examining monthly fluctuations in income and expenditure, a line of analysis not done before to this extent. In doing so, the study proposes a new metric for the measurement of category expenditure as an index of the total spend.\n",
"role": "assistant"
}
] |
Doctor-patient communication: a review and a rationale for using an assessment framework
|
[
"Assessment",
"Doctor–patient communication"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Ineffective doctor-patient communication (DPC) is problematic as it can lead to nonadherence and other health-related issues such as stress of paying for medical expenses (Meluch and Oglesby, 2015). Medication nonadherence is an important public health consideration, affecting health outcomes and overall healthcare costs (Aurel and McGuire, 2014). In fact, the estimated annual cost of hospital admissions for patients who had not taken their medications as prescribed reached $8.5 billion in 2012. When patient satisfaction is high, compliance with medical directives increases (Jha et al., 2008); predisposition to follow up with the provider is strengthened (Brown et al., 2003); the inclination to initiate litigation against providers tend to decrease (DuPre, 2010); and 30-day readmission rates are lower (Boulding et al., 2011). Thus, a critical factor in the effectiveness of healthcare delivery is sustaining patient centeredness through meaningful DPC.\n2. Background: 2.1 The HCAHPS survey\n3. Methods: The framework in Figure 1 as well as the method for assessing the effectiveness of DPC (see Figure 2) are based on the competing values framework (CVF) developed primarily to assess organizational culture and evaluate the effectiveness of organizational and interpersonal communication (Belasen, 2008; Belasen et al., 2015). Ovseiko and Buchan (2012) noted that the CVF is the most frequently adopted instrument in health services research, supporting Kwan and Walker's (2004) argument that the CVF has become the dominant model in organizational culture assessment in North America and elsewhere (Saame et al., 2011). An excellent review and application of the CVF as a tool for assessing healthcare organizations is provided by Looi et al. (2016). Moreover, the CVF has also been widely used for assessing and developing communication roles and personality traits (Belasen and Frank, 2008, 2010). Studies have shown that providing consumers with frameworks for understanding quality helps them value a broader range of quality indicators such as safety, effectiveness, and patient-centeredness. Further, when measures are grouped into user-friendly formats, patients can see the meaning of the measures more clearly and understand how they relate to their personal care (Hibbard and Pawlson, 2004). An adaptation of the CVF for examining the context and effectiveness of provider-patient interactions appears in Figure 1.\n4. Findings and discussion: In general, the aggregate mean in our sample of 100 residents was 4.25 with a standard deviation of 0.65, significantly lower (t=2.01; p<0.05) than the preferred \"ideal\" mean of 5.56. Among individual elements, the source of this gap stemmed from scores in \"Non-verbal Expression,\" \"Sense-making,\" \"Validation\" and \"Listening.\" The differentials in \"Verbal Inspiration,\" \"Clarification,\" \"Facilitation\" and \"Empathy\" were all statistically indifferent from the ideal means.\n5. Implications: Viewing the physician-patient encounter as a positive interpersonal context helps increase the trustworthiness of DPC. By focusing on the patient's specific issues and concerns and by giving and receiving reinforcing feedback, a trusting relationship between physicians and patients can be developed. Studies have shown that positive experience breeds higher commitment to quality care and better compliance, patient follow-ups, and overall increase in patient satisfaction (Verlinde et al., 2012). Moreover, and as discussed throughout this paper, optimizing physician-patient communication can lead to better patient health and outcomes (Duberstein et al., 2007; Jensen et al., 2010). Yet, while communication training for health providers is expanding, methods are often inconsistent, squeezed into curricular environments often packed with a focus on technical or clinical skills, and occur outside the clinical setting (Belasen et al., 2015). Moreover, linguistic barriers and cultural differences, already abundant, will only increase in coming years (Kundhal and Kundhal, 2003). Moving the training closer to the clinical setting, when feasible, offers advantages in that assessments of communication strengths and weaknesses, similar to the method discussed in this paper, can be pursued more systematically.\n6. Limitations and directions: While patient care is sometimes viewed as purely technical, there is evidence that DPC strongly predicts clinical outcomes as well as patients' overall ratings of hospitals. We agree with Quigley et al. (2014) call that since the patterns of variation are consistent with the nature of the specialty care (e.g. simple instructions may be especially important in geriatric medicine due to cognitive limitations of some older patients while interventional radiology requires complex language and extensive explanations), more research is needed to extend our understanding of the impact of the DPC on the overall HCAHPS ratings of hospitals. Following Matusitz and Spear (2014) suggestion, we think that researchers should adopt a qualitative method (e.g. content analysis) for analyzing DPC discourse.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore the extent to which improving doctor-patient communication (DPC) can address and alleviate many healthcare delivery inefficiencies.\nDesign/methodology/approach: The authors survey causes and costs of miscommunication including perceptual gaps between how physicians believe they perform their communicative duties vs how patients feel and highlight thresholds such as the Hospital Consumer Assessment of Healthcare Providers and Systems (HCAHPS) used by hospitals to identify health outcomes and improve DPC.\nFindings: The authors find that DPC correlates with better and more accurate care as well as with more satisfied patients. The authors utilize an assessment framework, doctor-patient communication assessment (DPCA), empirically measuring the effectiveness of DPC. While patient care is sometimes viewed as purely technical, there is evidence that DPC strongly predicts clinical outcomes as well as patients' overall ratings of hospitals.\nResearch limitations/implications: More research is needed to extend our understanding of the impact of the DPC on the overall HCAHPS ratings of hospitals. The authors think that researchers should adopt a qualitative method (e.g. content analysis) for analyzing DPC discourse.\nPractical implications: When a sufficient amount of DPCA training is initiated, a norming procedure could be developed and a database may be employed to demonstrate training program's efficacy, a critical factor in establishing the credibility of the measurement program and nurturing support for its use.\nOriginality/value: The authors highlight clinical and operational issues as well as costs associated with miscommunication and the need to use metrics such as HCAHPS that allow consumers to see how hospitals differ on specific characteristics.\n",
"role": "assistant"
}
] |
Heidegger among dryads: on the origin of the female work of art
|
[
"Women",
"Architecture",
"Arts",
"Heidegger",
"Gender",
"The Origin",
"Post‐structuralism",
"Disruption",
"Hidden voices"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Voice perhaps of language. Voice of voice. Unconscious of woman, female spirit, femme vital, she, elle (Irigaray, 1999 [1983]). Voice(s) or their element which could be pure desire, or air (Irigaray, 1999 [1983]) I term labyrinth, a necessarily opaque motif, by nature unthinkable, unwritable as such. Then how to explain the relevance of what follows for organisational behaviour and managed change? Assuming a reader uninterested in facile models, it suffices to recall Robert Cooper's (1989) admonishment, to overturn the \"writing of organisation\" in favour of the \"organisation of writing\". To inquire who designs the built environment does not seek a list of names, but the nature of their club.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to examine a lacuna in Heidegger's \"The Origin of the Work of Art\", that is, that it does not consider that such \"origin\" might be gendered. Originally published in a collection entitled Holzwege, which in the Cambridge (2002) edition is translated as Off the Beaten Track, Heidegger explains, \"Wood\" is an old name for the forest. In the wood there are paths, mostly overgrown, that come to an abrupt stop where the wood is untrodden. They are \"Holzwege\".\nDesign/methodology/approach: - The paper is written within a philosophical tradition and seeks to examine Derrida's lengthy deconstruction of Heidegger's essay and Irigaray's elegant gender critique, which are among works attesting to the importance and controversy The Origin has provoked among post-structuralist thinkers.\nFindings: - The paper considers The Origin as an auto-deconstructive text, through which other voices (one, or more, not Heidegger's) are heard; voices calling him, through him, as him; which he cannot identify or silence. They disrupt and confound his thinking and writing; he tries (unsuccessfully) to mend the discrepancies, assuage the violence, but ultimately leaves the essay in limbo, intuiting something even he cannot comprehend, an alien writing, that its internal contradictions be reserved for others to puzzle over.\nOriginality/value: - The paper offers a reading of The Origin against the Classical male hegemony excluding women from their proper participation in architecture, art in general, aesthetic hermeneutics and philosophy of art. These circuits of exclusion being integral to the phallocentric orthodoxy inherited from the Ancient Greeks.\n",
"role": "assistant"
}
] |
Spiritually inspired creativity in business
|
[
"Religion",
"Sustainability",
"Social economy",
"Entrepreneurship",
"Management",
"Human development"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: creativity and spirituality: Luc de Brabandere makes a distinction between innovation and creativity. Innovation is to change reality. It requires action by a team, is continuous, takes a long time, and delivers something new to the system. Its impact is measurable and certain, and requires the tools of project management. Creativity, on the other hand, changes perception and requires new thinking and feeling. It is discontinuous, and takes an instant. Its impact cannot be measured, and it requires learning methods fueled by questions, surprises and incomplete answers (de Brabandere, 2005).\n2. Spirituality in business: For over a decade efforts have been made to include spirituality in business and economics all over the world. In 2001 one of the authors organized an international conference in Szeged, Hungary, which was probably the first European workshop on \"Spirituality in Management.\" The papers of this workshop were published in the Kluwer Academic Series of Business Ethics. After this workshop the cooperation among different European centers involved in business ethics resulted in setting up the European SPES Forum in 2004 in Leuven, Belgium (SPES, the Latin word for Hope, is also an acronym for \"Spirituality in Economics and Society\"). The aim of the Forum is to promote spirituality as a public good and as a source of non-instrumental reasoning in business. The European SPES Forum was just one among other new networks in the field of applied spirituality. In the USA and Asia similar initiatives took place. Wuthnow (2000) talks about a \"quiet revolution\" and reports that 40 percent of Americans are involved in a small group that meets regularly and provides support for its participants. Business spirituality is a global discourse embedded in a broader intercultural and intercontinental development. The Palgrave Handbook of Business and Spirituality (Bouckaert and Zsolnai, 2012b) is a response to developments that simultaneously challenge the \"business as usual\" mindset.\n3. Working models of spiritually inspired businesses: We present three cases of spiritually inspired business models. The first is connected with Hindu spirituality, the second is based on Christian spirituality, while the third is influenced by Anthroposophy.\n4. Conclusion: beyond the materialism of modern mainstream business: There is growing evidence that the materialistic model of modern mainstream business does not produce true well-being for people and actually undermines well-being. \"Outmoded mental models have produced an intellectual bankruptcy: the bankruptcy of mainstream economic thought\" (Scharmerand Kaufer, 2013, p. 11). By advocating economic action on the basis of money-making, and by justifying success in terms of profits alone, the materialistic business model encourages the irresponsible behavior of economic actors, contributes to ecological destruction and disregards the interests of future generations. The presupposed and still widely used \"rational management model\" often produces non-rational outcomes for society, nature and future generations (Zsolnai, 2015).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the relation of spirituality and creativity in business context.\nDesign/methodology/approach: The paper presents practical examples of spiritual-based creative business models in different faith traditions (Hinduism, Christianity and Anthroposophy).\nFindings: Spirituality and a deep sense of connectedness are essential to enhance creativity and care in business. Spirituality creates free space and openness to allow the future to emerge organically. It creates a distance between the self and the pressures of the market and the routines of business and daily life. This distance is a necessary condition for developing creative, ethical and responsible solutions to the complex challenges around us.\nOriginality/value: Spiritually inspired creative business models overcome the instrumental rationality and materialistic orientation of today's business management which produces large scale ecological, social and ethical \"ills.\" The paper shows that alternative business management practices need a spiritual foundation to be more creative and caring.\n",
"role": "assistant"
}
] |
The role of emotions in employees' explanations for failure in the workplace
|
[
"Problem employees",
"Emotional dissonance",
"Employee behaviour"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Occasionally employees fail doing their jobs. Such failures frequently involve costs both for the organization and the failing employee. Costs of failure can include wasted working hours, product disqualification, lowered customer satisfaction, damaged reputation of the firm, harmed relationships between the failing employee and managers and/or co-workers, injuries, and even the loss of lives (Kletz, 2001; Tobias, 1999). In order to prevent repetition, organizations need to examine the reasons for failure.\nExplanations for undesirable outcomes and behaviors: In accordance with Benoit (1995) and Schoenbach (1990), we refer to four different types of tactics that people may use to explain their involvement with undesirable situations. The following typology and related definitions are also consistent with typologies suggested by others (Folkes and Whang, 2003; Hodgins et al., 1996; Hunter and McClelland, 1991; Itoi et al., 1996; Gonzales et al., 1990; Weiner et al., 1987, 1991a):Justification\nThe link between an employee's emotions and the choice of a tactic for explaining failures: As noted earlier, emotion is a powerful predictor of behavior. In failure situations, the function of emotions is often to deflect threats to the self (Haidt, 2003; Leary, 2004; Oatley, 2000; Tangney and Dearing, 2002; Tangney and Fischer, 1995; Tracy and Robins, 2004). In particular, the social emotions of shame and guilt are often associated with the experience of failure (Poulson, 2000). Another emotion that is likely to arise in response to personal failure events is fear (Cummings and Anton, 1990). Each of these emotions may have a different impact on the way an employee explains the failure. In the next section we delineate the typical characteristics of each of these emotions.Shame and guilt\nStudy 1: To test the role that guilt and shame play in determining an employee's choice of explanation for a failure, we examined employees' stories of real-life situations in which they failed in the organization and had to provide an explanation for it.Method\nStudy 2: Study 2 tested how shame, guilt and fear affect explanations for failure, using imaginary scenarios describing failing employees. Descriptions varied in the extent to which the organization was threatening and the extent to which failure was likely to induce feelings of either guilt or shame in the employee. Thus, study 2 served both as a test of the role fear has in the choice of explanations for failure, and as an attempt to replicate the findings of study 1 using a different method (fictional scenarios).Method\nGeneral discussion: The present research considered the role of a worker's emotional reaction in the way s/he explains a failure. Unique effects of three emotions, shame, guilt and fear that might be experienced by workers having to explain the undesirable outcomes to others were assessed. In general, we found that guilt was conducive to the organization's goal of getting an accurate explanation from the worker, and guilt tended to elicit from workers reactions that helped them maintain a good relationship with the organization. These effects were reflected in the fact that increased guilt was associated with an employee's willingness to take responsibility for what had happened. This was also shown both by their greater likelihood of apologizing and lower likelihood of using denials and excuses. Under these conditions participants were also more likely to offer compensation.However, this desirable effect of guilt was somewhat weaker when the organization was perceived as threatening. Under high threat, a worker's tendency to abstain from using denials was no longer related to the experienced level of guilt, although guilt was still positively related to level of compensation and willingness to apologize.Under the high threat condition, both the level of compensation offered by the employee and the likelihood that s/he would apologize were increased. Yet, given the effect that a high threat had on other types of explanations, it is not clear that these were sincere apologies or that compensation attempts were likely to restore or create a good atmosphere for the worker in the organization from that point forward. This prospect is in line with the fact that under high threat, fear was also positively correlated with the likelihood that the worker will deny his/her role in bringing about the undesirable outcome and provide an excuse. However, this assumption merits further research because it implicates that apologies and compensations may occur for different or even opposing motivations under certain conditions.Overall, these effects indicate that whereas guilt works to the benefit of the organization by encouraging workers to take responsibility for what happened, to provide an accurate account and to try to restore the situation, fear mostly works against such attempts. In addition, fear resulting from a threatening atmosphere may wipe out at least some of the beneficial effects of guilt. These effects suggest that conditions that enhance guilt feelings in failing employees work to the benefit of all parties involved when the motivation is to learn from that failure and to maintain a good relationship between the organization and the failing worker. In contrast, a threatening atmosphere not only works against such effects but also blocks some of the desirable effects that guilt might have had in the same context.While our research indicated that guilt and fear each uniquely affect a worker's choice of how to explain a failure, shame appeared not to have any special input to such a reaction. One reason for this may be that shame is highly correlated with guilt and that these two emotions are hard to distinguish in practice (Tangney et al., 1996a, b). The fact that in both studies we failed to create situations in which either of these emotions was dominant reflects the co-occurrence of these emotions and the difficulty in distinguishing between them.However, there might also be an additional reason why shame had no unique effect on the accounts and reactions in question. Although Lewis (1971) emphasized that the significant difference between guilt and shame is related to the role of the self in bringing about the event that is the antecedent for each emotion, there is yet another way in which these emotions may differ. Smith et al. (2002) have recently supported the view that shame is strongly affected by public exposure, while guilt is less sensitive to the question of whether others are or are not aware of one's actions and outcomes. Given that all the conditions examined in our studies involved an inquiry into the event (real or hypothetical) and hence involved public exposure to one degree or another, shame was likely to be similarly high under all conditions. The fact that the failures we studied all involved public exposure may also explain at least in part why shame had practically no independent impact. We assumed initially that increased shame is most likely to enhance denials and excuses, yet given that the worker's failure were known to others, we may have actually created little room for explanations by the worker that distorted reality. If this is true, shame should be expected to have its greatest impact when the failure is not known to others and the worker can decide whether or not to reveal the failure and his/her role in bringing it about. Accordingly, the chances that a worker will tell about a failure, take responsibility for it and provide an accurate account are expected to be lower as the level of shame goes higher. Finally, the involvement of the factor of publicity may also explain why in study 2 the counterfactual thinking strategy suggested by Niedenthal et al. (1994) failed to lead to distinct situations involving mainly guilt-related thoughts or mainly shame-related thoughts. It may be that the publicity of the failure situation is enough to blur the distinction between guilt and shame.Overall, although we were able to find substantial support for our assumptions, there are also some limitations concerning our methodology. We used self-reports to study emotions and explanations for recalled as well as hypothetical failures. This strategy proved effective in both fields of enquiry (Hareli and Weiner, 2000; Weiner et al., 1987). Yet, to provide further evidence for the role of emotions in explanations of failure in organizations, it is important in the future to move beyond such self-reports. This is especially necessary because self-reports often reflect participants' \"theories\" about the issue under study rather than their actual reaction in real life situations (Robinson and Clore, 2002). One possibility might be, for example, to use recorded material coming from inquiries of real failure events in organizations. Such a strategy could substantiate the findings of the present research. Another limitation associated with self-report studies is the possibility of common method variance. Common method variance refers to the fact that because a few measures come from the same source, any defect in the source contaminates all the measures (Podsakoff and Organ, 1986). Common method variance could have also been the cause of some of the correlations that emerged between our measures. The use of additional types of measures such as physiology and behavior in future research would be a good strategy for dealing with this threat.In spite of these limitations, the lesson that organizations can take from this research is that the emotions experienced by employees in the context of a failure and its inquiry have an important role in employee's decisions concerning how to explain such failures when questioned about them. While guilt feelings work in the interest of organizations by fostering explanations that help to cope with the event effectively, fear mostly hampers such goals. Moreover, guilt can help to maintain positive relations between the failing employee and others in the organization and can promote reparative attempts on the part of that employee such as the offering of compensation.Inducing fear of retaliation in workers is a strong obstacle to effective inquiries into employees' failures in the organization. In contrast, a supporting atmosphere coupled with empowerment of employees to take responsibility for their work not only increases trust (Cummings and Anton, 1990; Tzafrir et al., 2004), but is also more likely to induce emotions in failing employees that are associated with behaviors that benefit the organization. This is partly because such an atmosphere seems more likely to cause workers to experience guilt rather than fear when they fail. This reaction can be expected from the link between responsibility and guilt feelings (Weiner, 1985) - an additional advantage to the empowerment of employees.Overall, the present research is yet another indication of the importance of emotions in organizational settings and the fact that organizations ought to be more attentive to the effect that the emotions their employees experience at work have on their performance (Ashkanasy et al., 2000). Organizations need to create a climate that nurtures in the employee emotions that work to the benefit of both the employee and the organization when certain events such as failure occur. It also reinforces the idea that the study of the function of discrete emotions in organizational behavior, as opposed to general affective states such as moods, is a fruitful endeavor (Barsade et al., 2003; Lazarus and Cohen-Charash, 2001; Lord and Kanfer, 2002). Accordingly, exploration of the role that other emotions likely to arise in the context of failure, such as embarrassment and regret, might have in affecting worker response might be particularly informative.\n",
"role": "user"
},
{
"content": "Purpose: - The paper aims to study how shame, guilt and fear experienced by failing employees determine their explanation of the failure.\nDesign/methodology/approach: - Employees participated in two studies, one assessing actual personal examples of failures and another used imaginary vignettes. To manipulate the extent to which guilt or shame was the dominant emotion experienced by the failing employee, participants were asked to generate counterfactual thoughts typical of each of these feelings. Fear was manipulated by describing a threatening atmosphere in the organization. Measured was the likelihood that the employee took responsibility for what happened and provided a valid explanation. Likelihood of explaining the event by using excuses, justifications, concessions or denials was also measured.\nFindings: - Findings indicate guilt was associated with explanations that help the organization learn from the failure and assist employees in restoring their relationships with the organization and co-workers. Heightened levels of fear, however, decreased this desirable effect of guilt. Shame had no unique contribution to an employee's choice of explanations.\nResearch limitations/implications: - The use of self-reports and vignettes limits the ecological validity of the present findings. Nevertheless, it provides preliminary evidence for the importance of the factors under study.\nPractical implications: - These findings contribute to an understanding of the ways organizations can provide emotional settings conductive to constructive failure inquiries both for organizations and employees.\nOriginality/value: - The role emotions play in explanation of failures is an understudied issue both in social psychology and organizational research. The present study opens an avenue for more studies in this direction.\n",
"role": "assistant"
}
] |
A model-based approach to support privacy compliance
|
[
"Data protection",
"Privacy",
"Data processing",
"Compliance checking",
"Compliance demonstration"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Privacy is a multifaceted concept that has legal, social and political aspects. It is subjective in nature and culturally variable, not least because it is derived from societal demands, expectations and culture, which are all in turn influenced by political, social and economic factors, as well as information technology advancements (Onn et al., 2005). The complexity challenges software engineers to understand and translate abstract privacy principles and various perceptions into operational requirements (Gurses et al., 2011; Spiekermann, 2012). The variability challenges software engineers to understand and consider multiple stakeholders' expectations and concerns, which in turn, requires a specific expertise, contextual analysis and resolution of stakeholders' security and privacy interests (Gurses et al., 2011).\n2. Background and motivation: 2.1 The problem\n3. An approach to support privacy compliance: To facilitate reasoning about privacy compliance, the abstract purposes for which personal data are collected and processed need to be operationalised. The abstract purpose is a nonoperational objective to be directly achieved by the collection and processing of personal data. By operationalising it, we wish to move towards a position in which the objective is expressed in terms of personal data items, data-processing activities that consist of concrete actions and events that cause the execution of these actions and roles that define a set of responsibilities performed by different actors according to their capabilities. Further, abstract purposes are made operational through constraints that specify conditions to be satisfied before, or to be guaranteed after, the execution of corresponding operations. Accordingly, the main aspects of abstract privacy principles are specified as constraints on the abstract purpose. In particular, the purpose operationalisation consists of four separate, but related activities: refinement, conceptualisation, representation and evaluation.\n4. An example: 4.1 Overview\n5. Conclusion: Through a realistic case study, we have demonstrated the usefulness and applicability of the extension mechanisms that serve as a stepping stone to formally reason about the compliance of a requirements model with abstract privacy principles. The standard extension mechanisms have enriched the APDL profile with specific properties and constraints suitable for the domain of privacy and data protection.\n",
"role": "user"
},
{
"content": "Purpose: Concerns over data-processing activities that may lead to privacy violations or harms have motivated the development of legal frameworks and standards. Further, software engineers are increasingly expected to develop and maintain privacy-aware systems that both comply with such frameworks and standards and meet reasonable expectations of privacy. This paper aims to facilitate reasoning about privacy compliance, from legal frameworks and standards, with a view to providing necessary technical assurances.\nDesign/methodology/approach: The authors show how the standard extension mechanisms of the UML meta-model might be used to specify and represent data-processing activities in a way that is amenable to privacy compliance checking and assurance.\nFindings: The authors demonstrate the usefulness and applicability of the extension mechanisms in specifying key aspects of privacy principles as assumptions and requirements, as well as in providing criteria for the evaluation of these aspects to assess whether the model meets these requirements.\nOriginality/value: First, the authors show how key aspects of abstract privacy principles can be modelled using stereotypes and tagged values as privacy assumptions and requirements. Second, the authors show how compliance with these principles can be assured via constraints that establish rules for the evaluation of these requirements.\n",
"role": "assistant"
}
] |
The mobile phone as an argument for good governance in sub-Saharan Africa
|
[
"ICT",
"Comparative study",
"IT diffusion and adoption",
"G20",
"O38",
"O40",
"O55",
"P37"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: There are at least five main reasons behind the decision to assess the role of mobile phones[1] in promoting good governance in sub-Saharan Africa (SSA).\n2. Conceptual background and hypotheses development: 2.1 Conceptual background\n3. Data and methodology: 3.1 Data\n4. Empirical results: 4.1 Presentation of results\n5. Concluding implications, caveats and future directions: This study has presented theoretical and empirical arguments for the role of the mobile telephony in promoting good governance in 47 sub-Saharan African countries for the period 2000-2012. The empirical evidence is based on GMM with forward orthogonal deviations. After presenting theoretical frameworks, we have proceeded to discussing the empirical findings by first employing six individual governance indicators and then bundling these into composite indices to make our arguments more robust while at the same time increase the subtlety in our empirical analysis.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present theoretical and empirical arguments for the role of mobile telephony in promoting good governance in 47 sub-Saharan African countries for the period 2000-2012.\nDesign/methodology/approach: The empirical inquiry uses an endogeneity-robust GMM approach with forward orthogonal deviations to analyze the linkage between mobile phone usage and the variation in three broad governance categories - political, economic and institutional.\nFindings: Three key findings are established: first, in terms of individual governance indicators, mobile phones consistently stimulated good governance by the same magnitude, with the exception of the effect on the regulation component of economic governance. Second, when indicators are combined, the effect of mobile phones on general governance is three times higher than that on the institutional governance category. Third, countries with lower levels of governance indicators are catching-up with their counterparts with more advanced dynamics.\nOriginality/value: The study makes both theoretical and empirical contributions by highlighting the importance of various combinations of governance indicators and their responsiveness to mobile phone usage.\n",
"role": "assistant"
}
] |
Intelligence level analysis for crowd networks based on business entropy
|
[
"Differential evolution",
"Business entropy",
"Crowd intelligence",
"Network profession distribution",
"Subject interaction patterns"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Crowd networks where intelligent subjects are collaborated, interacted for a variety kinds of tasks, are widely existed in both human society and nature. In human society, crowd networks are pervasive in the field such as test automation (Bird and Munoz, 1983), customer preference analysis (Xiao et al., 2016.), supply chain, collaborative decision-making (Chen et al., 2017; Decker and Wattenhofer, 2013), currency network and so on. The more sufficiently multiple subjects work together, the easier our mission will be. Therefore, it is of great significance to study the collaborative effect of multiple intelligent subjects for optimizing this crowd network. In this paper, we mainly discuss some factors that affect the level of crowd network. Then we can update the network through these factors to make the network more powerful.\n2. The related work and background: 2.1 Crowd network\n3. Business entropy and its influencing factors: 3.1 The concept of business entropy\n4. Multi-factor business entropy quantitative model: In this section, the two factors including Subject Interaction Pattern and Profession Distribution Deviation are introduced into the business entropy model as parameters to make it more generalized and accurate.\n5. Case studies of analyzing crowd intelligence based on business entropy: According to the types of basic problems solved by crowd intelligence science and the functions of crowd intelligence network system, the basic problems of crowd intelligence science can be divided into crowd decision and crowd cooperation.\n6. Differential evolution model and k-means clustering were used to distinguish intelligent subjects: In the previous sections, the paper discussed the related factors influencing the business entropy. The 3 network examples in the Section 5 can be classified by their division of labor. However, not all networks can find their professional categories only through division of labor. Therefore, in Section 6, k-means clustering algorithm with improved differential evolution is proposed.\n7. Summary: This paper combines qualitative analysis of the factors affecting business entropy and quantitative expression of business entropy to discuss the relationship between professional categories of intelligent subjects in crowd intelligence network and the intelligence level of crowd network. There are many factors that affect the level of crowd intelligence network. Two important factors are listed in this paper, namely, Profession Distribution Deviation and Subject Interaction Pattern. PDD is quantified based on the deviation between the current NPD and optimized NPD. SIP is quantified according to the network topology. By combining these two factors, the MFBEQ model is proposed. By utilizing the model, three typical crowd intelligence networks (including logistics service supply chain, ant division and collaboration network and bit-coin network) are analyzed in detail.\n",
"role": "user"
},
{
"content": "Purpose: Intelligence level of a crowd network is defined as the expected reward of the network when completing the latest tasks (e.g. last N tasks). The purpose of this paper is to improve the intelligence level of a crowd network by optimizing the profession distribution of the crowd network.\nDesign/methodology/approach: Based on the concept of information entropy, this paper introduces the concept of business entropy and puts forward several factors affecting business entropy to analyze the relationship between the intelligence level and the profession distribution of the crowd network. This paper introduced Profession Distribution Deviation and Subject Interaction Pattern as the two factors which affect business entropy. By quantifying and combining the two factors, a Multi-Factor Business Entropy Quantitative (MFBEQ) model is proposed to calculate the business entropy of a crowd network. Finally, the differential evolution model and k-means clustering are applied to crowd intelligence network, and the species distribution of intelligent subjects is found, so as to achieve quantitative analysis of business entropy.\nFindings: By establishing the MFBEQ model, this paper found that when the profession distribution of a crowd network is deviate less to the expected distribution, the intelligence level of a crowd network will be higher. Moreover, when subjects within the crowd network interact with each other more actively, the intelligence level of a crowd network becomes higher.\nOriginality/value: This paper aims to build the MFBEQ model according to factors that are related to business entropy and then uses the model to evaluate the intelligence level of a number of crowd networks.\n",
"role": "assistant"
}
] |
Exploring online and offline shopping motivational values in Malaysia
|
[
"Malaysia",
"Shopping behaviour",
"Online shopping",
"Shopping motivation",
"Offline shopping"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The evolution of a Malaysian shopper's lifestyle has reflected the social and economic changes of the developing nation over the years. Overall income level has risen from RM1,566 in 1992 to RM6,141 in 2014, and average household expenditure from RM1,161 in 1993 to RM4,585 in 2014 (Economics Planning Unit, 2015). Information and communication technology (ICT) usage has increased concurrently, with individual internet users rising from 57.0 percent in 2013 to 71.1 percent in 2015, and households with internet access from 58.6 percent in 2013 to 70.1 percent in 2015 (Department of Statistics, 2016). These domestic drivers, along with an increasingly globalized environment, have resulted in Malaysians recognizing the need for better quality education that improves ICT literacy. The Malaysian Education Blueprint is a prominent initiative by the Ministry of Education that seeks to address this need by integrating relevant learning materials into the national curriculum (Ministry of Education, 2012).\nLiterature review: Shopping motivation theories\nMethodology: Research design\nAnalysis and findings: Overall, the participants were all Malaysians aged 21 years and above, and from major cities such as Kuala Lumpur, Penang, and Sarawak. Demographically, 63 percent were female, and 37 percent were male, with 56 percent of participants below the age of 35 and 44 percent above the age of 35. The majority of the sample was working adults of Chinese descent with undergraduate qualifications. The average time spent on both online and offline stores was less than 7 hours (a week). This demographic information is presented in Table II. The research was designed to be used with two statistical techniques, namely, factor analysis and pairwise t-tests, which will be discussed below.\nDiscussion and conclusion: The purpose of this study was to identify the shopping motivational values of online and offline Malaysian shoppers, and determine if there were differences between the shopping motivational values based on shopping channel, gender, and age group. The results indicated that there is an overall predisposition for respondents of both genders and age groups to place importance on the shopping motivational values when shopping offline.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to combine the elaboration likelihood model and hedonic/utilitarian values into a coherent model of shopping motivational values for online and offline Malaysian shoppers. Once these values were identified, they were compared to one another based on shopping channel, gender, and age group.\nDesign/methodology/approach: The Shopping Motivational Value Questionnaire was developed by the researcher after extensive literature review and measured the respondent's perception toward shopping motivational values in both shopping channels. Data were collected from 306 Malaysians above the age of 21, and analyzed using exploratory factor analysis and pairwise t-test techniques.\nFindings: The factor analysis revealed a three-factor model of shopping motivational values. Gratification Values were the most important to shoppers, followed by Essential and Societal Values. Shoppers perceived all these motivational values as important when shopping offline as compared to online, regardless of gender and age group.\nResearch limitations/implications: The results of the factor analysis revealed a three-factor model of shopping motivational values: Gratification Values which were the most important to shoppers, followed by Essential and Societal Values. These values were then compared to one another based on shopping channel, gender, and age group. Results indicate that shoppers perceived all these motivational values as important when shopping offline rather than online, regardless of gender and age group. Moreover, the perception of shoppers toward specific product categories and other shopping channels were not taken into account.\nOriginality/value: This study attempted to combine the elaboration likelihood model and hedonic and utilitarian values into a coherent model. A questionnaire was developed by the researcher with these two theoretical models as its foundation. Additionally, the shopping motivational values of this study have been formulated to be compatible to both shopping channels.\n",
"role": "assistant"
}
] |
"Sharing is caring": Corporate social responsibility awareness explaining the relationship of information flow with affective commitment
|
[
"Corporate social responsibility",
"Affective commitment",
"Information flow",
"Organizational communication",
"Affective psychology",
"Social responsibility",
"Corporate communications"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: More and more organisations communicate their corporate social responsibility (CSR) activities as a means to enhance stakeholder relations (Korschun et al., 2009), and strengthen stakeholder perceptions of their organisation (Bhattacharya et al., 2009). CSR activities can increase brand value (Luo and Bhattacharya, 2006), stimulate buying (Brown, 1998), and create a reliable and honest reputation (McWilliams and Siegel, 2001; Viswesvaran et al., 1998). While the effects of CSR activities on external stakeholders' perceptions and behaviours have been studied extensively, researchers have largely overlooked the effects on internal stakeholders (for notable exceptions, see Bartel, 2001; Bhattacharya et al., 2008; Dhanesh, 2011). This is unfortunate because employees are arguably the most important contributor to CSR activities, adding credibility to CSR initiatives (Morsing et al., 2008) and corporate brands. It is known that customers are critically important for the survival and growth of business (Turker, 2009). In addition, employees are the most important ambassadors of the organisation because they are the ones communicating organisational values to external stakeholders, and thereby shaping expectations (Collier and Esteban, 2007; Dawkins and Lewis, 2003). We propose that, by using communication to enhance employee awareness of the organisation's CSR activities, organisational communication can increase affective commitment.\nOrganisational communication and commitment: In this study, organisational communication is regarded as the flow, exchange, and creation of information through and by networks and its \"actors\". Information, in this perspective, moves from one point to another, varies in quality, may be embedded in messages that are distinct or equivocal, influences the uncertainties that employees face, and may be distorted by employees, resulting in different interpretations. Once information is obtained, it is processed, interpreted, and integrated in line with available information and schemata, which possibly alters its intended meaning (Conrad and Haynes, 2001). Employees' perceptions of organisational communication can also influence their emotional attachment to the organisation. Being well-informed about job-related and organisational issues (such as job objectives, organisational goals, and new developments) provides employees with empowering job resources, allowing them to contribute to the organisational goals. This could, consequently, lead to a greater attachment to the organisation. The psychological relationship between employees and organisations is often described by referring to the cognitive-motivational construct commitment. Commitment is defined as employees' interest in and connection to an organisation (Steers, 1977). Research has demonstrated that employees and organisations are adversely affected by low levels of commitment, and that both benefit from high levels of commitment (Abelson, 1983; Bakker et al., 2003; Valentine et al., 2002). This is especially true for affective commitment (Meyer et al., 2002). This form of commitment is described as employee's positive emotional attachment to the organisation, whereas normative and continuance commitment are described as feelings of obligation and lack of choice (Meyer and Allen, 1991). Therefore, affective commitment is the preferred form of commitment for organisations and also the form that has been associated with organisational communication most often (e.g. Postmes et al., 2001).Organisational communication considered as flow and exchange of organisational information has been linked to commitment in several ways. In their study, Postmes et al. (2001) show that employees are more strongly committed to the organisation if they obtain adequate information to perform their tasks, through formal channels, than when they receive this information informally. Communication with immediate coworkers (horizontal communication) predicted commitment worse than communication with or from senior management did (vertical communication). Furthermore, Postmes et al. (2001) report a stronger relationship between organisational communication and commitment in large, than in small organisations. Ng et al. (2006) focused on the link between management communication and organisational commitment (see also Carriere and Bourque, 2009; Van Vuuren et al., 2007). Management communication is the extent to which organisations provide organisation-related information to their employees, such as information about policies and procedures, financial results, and organisational changes (Vandenberg et al., 1999). Rousseau (1998) proposes that communication of organisation-related information may enhance employees' psychological attachment to the organisation, because it encourages employees to perceive themselves as members of the organisation and to contribute to the organisation's goals. Therefore, as a starting point, we predict a direct relationship between organisational communication and affective commitment, as formulated in the following hypothesis:H1. Organisational communication is related to affective commitment.Organisational communication and CSR awareness\nMethod: Sample and procedure\nResults: Descriptive statistics and measurement model\nDiscussion: The objective of this study was to expand our knowledge of the influence of organisational (CSR) communication on internal stakeholders. We proposed that by enhancing employee awareness of the different types of CSR activities, organisational communication could increase affective commitment. In line with prior research, we anticipated that affective commitment would not only increase from receiving support, but also from giving support. Grant et al. (2008, p. 902) found that giving initiated \"a process of prosocial sensemaking, in which giving causes employees to judge personal and company actions and identities as caring. As a result, employee giving strengthens their affective commitment to the company\". In the current paper, we study whether this notion can be extended to CSR activities by examining if the organisational communication of a large insurance company can increase affective commitment, by creating awareness about CSR \"giving\" activities.We found support for H1: information flow was positively related to affective commitment. H2 was partly confirmed: we found that organisational communication was related to affective commitment through awareness of CSR activities. However, only awareness about CSR activities aimed at employees and customers mediated this relationship. Awareness of CSR activities aimed at the government and society at large did not influence affective commitment.The impact of the awareness of CSR towards employees and customers can be explained by social exchange theory (Emerson, 1976), which is among the most influential conceptual paradigms for understanding workplace behaviour (Cropanzano and Mitchell, 2005). In a social exchange, Actor A trusts that his or her input will be returned by Actor B at some point in the future, without specifying the future obligation (Blau, 1992, Van Knippenberg et al., 2007). Social exchange often starts in an economic context, in which the value of gifts offered and received is in exact balance. If one receives a gift, one feels the responsibility to give another gift in return. After a long enough period of giving and receiving, the interaction partners no longer count their investments and returns, instead perceiving the relationship as rewarding to both (Organ et al., 2006, p. 55). The continuing exchange of favours results in the development of behavioural patterns and long-term relationships (Cropanzano and Mitchell, 2005). Exchange of positive behaviour between co-workers, such as sharing lunch and helping each other, has been indicated by previous studies (Hodson, 2001; Ten Brummelhuis et al., 2010). The positive relationship between CSR activities aimed at employees and affective commitment can be understood as the outcome of the exchange relation between an employee and the organisation. When employees, through communication, become aware of a CSR program aimed at employees, such as an employee support program, they are more likely to feel that their organisation values their well-being and thus reciprocate by developing affective commitment toward the organisation. The impact of CSR activities aimed at customers on affective commitment can be interpreted in a similar vein. Employees develop an (economic) relationship with customers of giving and receiving. For the maintenance of this mutual relationship, it is in the interest of employees that their organisation takes good care of the customers. In other words, employees will value organisational CSR initiatives aimed at customers, because it is also good for them. Therefore, the awareness of CSR activities aimed at customers will increase affective commitment. After all, by employing CSR activities aimed at customers, the organisation invests in maintaining the work relationships of employees. The latter will reciprocate this investment with increased commitment.Although the positive effect of CSR \"giving\" activities on commitment reaches further than giving support to colleagues, it does not extend beyond giving support to the most direct stakeholders. Therefore, it seems that our findings can be better explained by means of a social exchange perspective, as opposed to the prosocial sensemaking process. Nonetheless, since we did not measure prosocial sensemaking as an underlying mechanism, we cannot exclude that as an explanation either. However, the non-findings regarding CSR aimed at government and the CSR aimed at society at large cannot be explained by Grant et al. (2008)'s prosocial sensemaking perspective. Regarded from this perspective you would expect employees to become more committed as a result of the organisation's efforts to do good, regardless of the proximity of the support receiver. From our study it seems that \"caring\" must benefit stakeholders that are close to the employee, to have an impact on affective commitment. Apparently, when CSR activities are aimed at distant external stakeholders, like the government and society at large, it does not strengthen the emotional commitment of employees. This finding can be explained in two ways:1. employees do not feel (the need for) a strong relationship with these external stakeholders and therefore these CSR activities do not influence employee commitment towards the organisation (social exchange perspective), or;2. the distance felt between the employee and the target of the CSR activity limits the contribution to a prosocial identity (Aquino and Reed, 2002).The results from the present study imply that organisations can use organisational (CSR) communication to enhance affective commitment. Although in general the notion that the quality of organisational communication can enhance affective commitment is well established (Postmes et al., 2001).We show that employees become more committed because they are more aware of the organisation's good deeds. However, to provide more specific practical recommendations, more insight is needed into the underlying mechanism. If it is the social exchange perspective, organisations could stimulate affective commitment by building on a stronger sense of relatedness with the external stakeholders that benefit from the organisation's CSR activities. If employees feel associated with the people that benefit from the CSR activities, they may return the favour the organisation did by enhancing their affective commitment to the organisation.However, if the mechanism at work appears to be prosocial sensemaking, more employee involvement with CSR activities could contribute to a prosocial identity (Grant, 2007, 2008) and subsequently to more affective commitment. Engaging in CSR activities will make the employees feel better about themselves, and about the organisation. Additionally, it is important to note that CSR activities of the company might not have a similar effect as internal employee support programs (Grant et al., 2008), where employees have the actual opportunity to give.Theoretical and practical contributions\n",
"role": "user"
},
{
"content": "Purpose: - The effects of corporate social responsibility (CSR) communication on external stakeholders' perceptions and behaviours have been studied extensively; however, researchers have largely overlooked the effects of CSR communication on internal stakeholders. This study seeks to propose that, by enhancing employee awareness of the organisation's CSR activities (aimed at society, the government, customers, and employees), organisational communication can increase affective commitment.\nDesign/methodology/approach: - Data were collected at a Dutch healthcare insurance company, using a web-based questionnaire (n=301). The proposed model was tested with structural equation modelling (SEM) using Amos.\nFindings: - The bootstrapping results showed a positive association between information flow and affective commitment, with a mediating effect for CSR aimed at employees and customers. CSR projects aimed at the government and society at large did not mediate the relationship between information flow and affective commitment.\nOriginality/value: - The impact of CSR towards employees and customers can be explained by social exchange theory. When employees feel that their organisation values their well-being and that of their customers, they reciprocate what they receive from the organisation by committing themselves to the organisation. As such, this study provides an important argument for the implementation of CSR activities: it increases the emotional attachment of employees towards their organisation.\n",
"role": "assistant"
}
] |
A status quo analysis of disaster risk reduction policy and legislation in Cameroon
|
[
"Resilience",
"Disaster risk reduction",
"Cameroon"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The concept of a \"disaster\" or \"catastrophe\" is not new to the nation of Cameroon. Some of the first recorded disasters (MINATD/DPC, 2002) in Cameroon only date back to three major disasters (the Lake Nyos gas disaster in 1986, Nsam Train Crash in 1998 and Mount Cameroon eruptions). Implicitly, this does not mean that economic and human costs of disasters were not experienced before this time. Two historical accounts of major natural hazards in Cameroon can be traced back to significant recorded rainfall in Debunscha (one of the wettest places in Africa) (Frankham et al., 2002) lying at the foot of Mount Cameroon.\nCameroon's historical hazard and disaster profile: Figures 1 and 2 indicate that geological and hydrometeorological hazards are not the only categories of disasters that occur throughout the area of Cameroon. At the time this article was written, Cameroon's hazard and disaster profile (Aka et al., 2016) indicated an increase in natural hazards (seasonal floods) triggering landslides, mudslides, and to an extent, technological accidents and protracted conflicts (Figures 1 and 2).\nThe concept of civil defence/protection in Cameroon: The concept of civil defence/protection dates back to the general organization of civil defence under Law No67/LF/9 of 12 June 1967, sect. 11 and 18 (MINATD/DPC, 2002). This law mandated the Ministry of Territorial Administration (MINAT) to implement and coordinate civil protection measures. These efforts concerned the risks incurred by the civilian population as a result of calamities such as fire, floods, cyclones, earthquakes and the like. When the law was proclaimed, la Republique du Cameroun had just gained independence from France in 1960, while British Cameroon remained as a trusteeship mandate under the League of Nations. In France, General Charles de Gaulle, the then President of France, instructed the Prime Minster (Georges Pompidou) in 1964 to create an organization to unite all the forces contributing to the protection of civilian populations at the national level. Hence, the first National Federation of Civil Protection (FNPC) was created in France in 1965, while her former colony, Cameroon, mimicked her own Civil Protection laws in 1967 and 1968 respectively. The 1968 Presidential Instruction was more focused on the safeguarding and protection of civil installations of vital importance (No02/CAB/PR of 18 January 1968). Between the period of 1967 and 1972, Cameroon (East and West Cameroon) had a political turnaround from Federalism to a Unitary state (Chem-Langhee,1995) while its first President, Ahmadou Ahidjo, was elected in 1972. This period also saw law No 71/DF/381 of 11 August 1971 supplementing article 6 of law No 71/DF/302 of June 1971 to organize the MINAT. At the international level, 1970 to 1986 saw the creation and evolution of the United Nations Disaster Relief Office (UNDRO) as part of the United Nations Office for Humanitarian Assistance (UN/OCHA), and the International Decade for Natural Disaster Reduction, IDNDR 1990-1999, later extended from 2000 to 2010 as the International Strategy for Disaster Reduction (ISDR)[1].\nCurrent disaster risk reduction and civil protection laws and legislations (1986-present): August 21, 1986, saw one of the deadliest disasters in the Lake Nyos region in Cameroon, with over 1,750 people killed, and thousands of livestock found dead after a sudden limnic eruption, releasing almost 300 000 tons of carbon dioxide into the region. It was as a result of these events that the 1986 structural legislative law No 86/016 of 6th December 1986 - re-organizing the department of civil protection and disaster management in Cameroon - came into force. The National Council on Civil Protection was established (CNPC) by Decree 96/054 of 12 March 1996, two years before the Nsam Train Fire Disaster in 1998 (two trains collided, one carrying tons of oil), triggering Decree No 98/031 of 9th March 1998 to organize emergency management and major risk relief plans. The disaster-driven reactive nature of Cameroonian policy is thus clear. During this time, the UNDP engaged in the National Disaster Prevention and Management Programme (NDPMP). The focus was on capacity building for disaster management with the support of the Cameroon government and the UN/OCHA in Geneva. Other important partner organizations that played a major role in Cameroon's disaster management system are the International Civil Defence Organization (ICDO) of which Cameroon became a member in 2000. Reports show that Cameroon remains a proactive member of the ICDO, focusing on capacity building of staff of the Department of Civil Protection (DPC), the National Fire Brigade, and other government staff like Governors and Senior Divisional Officers. Furthermore, a decree from the Prime Minister's Office (No 037/PM March 19, 2003) saw the creation of the National Risks Observatory Commission (Observatoire National des Risques - ONR) seen in Figure 3.\nResearch design and approach: This research used both qualitative and quantitative research instruments. To answer the research questions:\nRoot causes of mainstreaming disaster risk reduction into sustainable development for Cameroon: Cameroon's Vision 2035 includes factors of climate change and environmental risks (MINEPAT, 2009). However, both Vision 2035 (MINEPAT, 2009) and the National Adaptation Plan (UNFCCC, 2016) do not have feasible action plans and programs to integrate DRR as a development priority. Furthermore, there is weak evidence of collaboration in mainstreaming DRR across sectors within Cameroon. Climate change, sustainable development and DRR/M, are handled by different ministries. Hence, the Ministry of Environment and Forestry (MINEF), now separated from MINEPDED, for example, has identified and elaborated specific aspects of natural resources and natural risks in Cameroon under Project CMR/92/008 (MINEF and UNDP, 2008). Subsequently, three of five aspects proposed by MINEF together with the UNDP focusing on disaster management legislation and sustainable development are the following:\nDiscussion of the empirical research: The results from Table III from the data analysis indicate that majority of the respondents 95 per cent strongly agree that policies and legislation governing DRR in Cameroon are ineffective and need review. Likewise, less than 15 per cent attest to the fact that DRR commitments actually constitute an integral part of Cameroon's development goals, and Vision 2035. More than 80 per cent agreed that there is not an enabling environment for the DRR policy and legislation in Cameroon. According to the GOAL Toolkit, respondents were asked to rank Cameroon's DRR integration and performances. The majority of the respondents (98 per cent) ranked Cameroon under minimal integration with less than a 20 per cent score.\nChallenges that Cameroon face in the implementation of DRR practices: The research highlights principal challenges that Cameroon faces in the implementation of DRR plans, laws and legislation at the national level. These DRR plans either failed to involve multiple stakeholders, were implemented without the consultation of DRM professionals, did not consider or align with the international framework governing DRR, or focused on the reactive and humanitarian approach, not capable enough to be holistic and broad in its functions and operations in risk reduction and prevention. In fact, some of these documents do not embrace the principal UNISDR definition of risk. Basically, Cameroon demonstrates emergency management practices and humanitarian response rather than risk reduction in its institutional framework. The findings of this research also reveal that Cameroon scores poorly with regard to the integration of DRR when measured against the GOAL Scorecard (GOAL, 2015), with a score of 20 per cent, indicating minimal integration. With a critical geopolitical challenge within the framework of the national platform, Cameroon has the task of reviewing its exclusion of DRR English-speaking professionals. The concept of the DRR national platform by definition is multi-sectoral and is not meant for a chosen few political elites. Marginalized groups and professionals should be recommended to get involved to improve DRR results and align with international standards where necessary. Second, over-centralized and bureaucratic DRR coordination and implementation efforts (Bang, 2013; Gaston et al., 2012) should be limited and decentralization by devolution should be an alternative solution.\nConclusion and recommendations: The findings of this article show that DRR strategies, actions and policies in Cameroon are relatively ineffective (Bang, 2014; Ndille and Belle, 2014) within the day-to-day development framework of the Cameroonian territory. This research sampled ten administrative sectors in Cameroon to measure how much of the DRR practices, actions and policy had been integrated into institutional planning and development processes within the past 50 years since disaster risk management laws were re-organized in Cameroon.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze the status quo of disaster risk reduction (DRR) policy and legislation in Cameroon.\nDesign/methodology/approach: Using a qualitative method, this paper examines historical data from sectoral administrative reports, plans, declarations, commitments and speeches, texts and peer-reviewed journals on disaster and risk management in Cameroon for the period 1967-2017. Empirical data from ten selected government sectors were used to analyze the status quo, together with quantitative data collected by using four instruments (i.e. HFA Priority 1 & 4, USAID Toolkit, GOAL Resilience Score and the Checklist on Law and DRR).\nFindings: Findings show that Cameroon largely still practices disaster response through the Department of Civil Protection. Transparency and accountability are the sine qua non of the state, but the lack thereof causes improper implementation of DRR within development institutions. DRR is seen as an ad hoc activity, with the result that there is not effective institutional capacity for implementation. The need to develop a new national DRR framework is evident.\nOriginality/value: Analyzing the status quo of DRR in Cameroon could assist with the review and reevaluation of a new DRR framework within the Cameroonian territory.\n",
"role": "assistant"
}
] |
Targeting the reduction of shipping emissions to air: A global review and taxonomy of policies, incentives and measures
|
[
"Shipping",
"Policies",
"Incentives",
"Abatement measures",
"Maritime air emissions"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Shipping is essential for the growth of international trade, carrying more than 90 per cent of global trade in volume. From this perspective, it is fortunate that shipping is the most energy-efficient mode of transport and, as such, is often held to be the least environmentally damaging mode of long-distance freight transport. This and its other benefits in comparison to other modes are identified by Lopez-Navarro (2013) and Styhre et al. (2014). It is for this reason that a number of EU transport policies, for example, are aimed at incentivising a modal switch for freight from land to water (European Commission, 2011), even though the benefits of such a modal switch are not categorical in every case (Ng, 2009; Baindur and Viegas, 2011; Douet and Cappuccilli, 2011; Tzannatos et al., 2014). This general policy direction is reflected in many similar regional and national policies on a worldwide basis.\nPolicies, incentives and measures targeting the reduction of maritime air emissions: There have been various efforts to review the various technical, operational and market-based measures, policies and incentives implemented worldwide for targeting the abatement of air pollution from ships. However, most of them have only a limited scope, focussing either on specific types of measures (e.g. technical), particular geographical areas (e.g. ports) or individual policies (e.g. vessel speed reduction). None of them has dealt with air emissions from shipping as a whole; most of the studies have been divided into combating local air pollutants (such as SOx, NOx and PM) or GHG emissions that have a global impact.\nMethodology: As referred to above, the scope of this paper is the identification and classification of the policies, incentives and measures that have been implemented across the globe and are related to the abatement of air emissions from shipping. To identify the full range of policies, incentives and measures that have been adopted for the reduction of shipping air emissions, an extensive survey of various sources was conducted during late 2017 and early 2018. Official reports of international and regional institutions (like the IMO and the EU), government policy documents, port authority websites, classification society pages, private firms' sites and the academic literature are included in our research. This exploratory review resulted in the establishment of a comprehensive global database of measures, policy instruments and initiatives proposed and encouraged by the whole range of shipping stakeholders and decision-makers - intergovernmental organizations, public authorities, private firms and the research community - for the reduction of shipping air emissions.\nResults: Organization profile of the institutions that have adopted measures, policies and incentives with maritime emission reduction potential\nDiscussion and conclusions: The abatement of maritime air emissions has become an issue of major concern during the past few decades, resulting in the development and adoption of various measures, policies and incentives targeting the reduction of these emissions by a range of different actors, from inter-governmental organizations to national, regional, public institutions and private associations. Throughout this paper, we established a comprehensive global database of all types of initiatives that target the reduction of shipping air emissions implemented globally in the period from 2008 till 2018 and then classified them on the basis of specific features that they share. Apart from the classification of the various initiatives, we analyzed the organizational profiles of the institutions that have adopted and implemented these policies to give some insight into the organizations that have been most active in combating air pollution and GHG emissions from shipping.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to identify and classify the various initiatives developed and implemented across the globe for the abatement of maritime air emissions.\nDesign/methodology/approach: In this paper, an extensive survey of various sources was conducted, including the official reports of international and regional institutions, government policy documents, port authority websites, classification society pages, private firms' sites and the academic literature. The initiatives were then categorized in accordance with the classification of the Swedish Environmental Protection Agency and analyzed using the SPSS Statistics software to give some insight into their frequencies and the interrelationships between them.\nFindings: This exploratory review resulted in the establishment of a comprehensive global database of initiatives encouraged by the whole range of shipping stakeholders and decision-makers for the reduction of shipping air emissions. According to the findings, economic incentives that provide motivation for the adoption of less environmentally damaging practices are the most commonly used initiative, followed by infrastructure investments and informative policies.\nResearch limitations/implications: The results provide implications for further research that include an in-depth analysis of ports' policies, as well as an evaluation of initiatives applied on a large scale to map their emissions reduction potential for shipping.\nOriginality/value: The main contribution of this paper is the identification and analysis of all the diverse initiatives implemented globally in a comprehensive way and its dealing with air pollution from shipping as a whole.\n",
"role": "assistant"
}
] |
Furthering an understanding of West African plant foods: Mineral, fatty acid and protein content of seven cultivated indigenous leafy vegetables of Ghana
|
[
"Proteins",
"Minerals",
"Food products",
"Ghana",
"Plants",
"Africa"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: We were interested in the mineral, fatty acid and protein content of the leaves of seven indigenous leafy vegetables (ILVs) that grow in Ghana, West Africa as well as in learning about the uses of these plants in the communities where they are consumed. The plants analyzed in this study, with local Twi language names in parenthesis, include: Hibiscus sabdariffa (bitto), Hibiscus cannabinus (berese), Amaranthus cruentus (aleefu), Corchorus olitorius (ayoyo), Solanum macrocarpon (gboma), Xanthosoma sagittifolium (nkontomire), and Vigna unguiculatus (bento). Indigenous leafy vegetables in general are widely regarded as excellent sources of calcium, iron, zinc and many other essential micronutrients. The impetus for investigating the nutritional value of these plant foods is the growing demand for ILVs in Ghana and the significant role they play in the diets of local communities and increasingly in urban communities. As urban centers grow, there is growing demand in cities for these \"traditional\" plant foods by urban populations for preparation in the home as well as for sale in local restaurants. What had often been thought of as \"traditional\" or rural foods are now attaining new status in the growing middle class of urban Ghana.The plants in this study are primarily used as potherbs, meaning they are cooked in a variety of soups that constitute the most common main course of Ghanaian meals. Leaves from the plants are used fresh or may be dried and stored for future use. The leaves are most commonly added to soups that are eaten at meals, which may contain other vegetables or meat. Dokosi (1998) has documented the uses of numerous ILVs in the rural areas of Ghana and when relevant, also highlights medicinal uses of the plants as well.In the present study, we sought to acquire information regarding the content of protein and essential amino acids, fatty acids, and minerals and trace elements in seven cultivated ILVs in Ghana in order to further our understanding of the potential nutritional benefits of these foods and their role in Ghanaian diets.\nMaterials and methods: Collection of plants\nResults: Comments regarding ethnobotanical considerations\nDiscussion: Collectively, the seven plants we describe herein represent a sampling of ILVs in Ghana that are increasing in value in urban centers because of their desirability for consumption by large segments of their population and increasing economic value. As demand for these plants continues to grow, cultivators in rural areas of the country have begun considering cultivating some of the plants on their land to diversify their cultivars.The main finding of this study was that, with the exception of Xanthosoma sagittifolium, which was inferior to the other six plants, the ILVs we investigated appear to contain nutritionally useful amounts of good-quality protein, a-linolenic acid, and many essential minerals.The protein in a typical serving of plant Hibiscus sabdariffa, Hibiscus cannabinus, Amaranthus cruentus, Corchorus olitorius, Solanum macrocarpon or Vigna unguiculatus could contribute significantly to satisfying the daily needs of an adult for all of the essential amino acids, except the sulfur amino acids methionine and cysteine. For example, consumption of about 80 g of fresh Xanthonsoma sagittifolium (20 g dry weight) would provide about 5 g of protein, which compares with a recommended daily intake of 42 g of protein for an adult (World Health Organization, 2007). Noteworthy is the fact that all seven of the plants contained less-than-ideal proportions of cysteine plus methionine relative to the WHO standard protein (World Health Organization, 1985). Plant proteins in general and especially those derived from legumes, tend to be low in methionine and cysteine (McCarty et al., 2009). Exacerbating this problem is the fact that the digestibility of plant proteins is lower than that of animal proteins (Saunders et al., 1973; Pirie, 1978; McCarty et al., 2009). In light of the low scores of all seven plants for the sulfur amino acid pair relative to the WHO standard protein (see Table III), the diets of populations who rely extensively on any of these indigenous leafy vegetables would have to be complemented with some other protein source that could provide these essential amino acids.The seven plants reported in the current study are readily and widely available in Kumasi because they are grown in family gardens and local farms.With regard to the ability of these plants to contribute to satisfying one's mineral requirements, take calcium and iron as examples. The recommended daily intakes of calcium and iron for an adult male are 1,000 mg and 10 mg, respectively. From the data in Table IV, one can estimate that 20 g dry weight of Vigna unguiculatus would provide more than 600 mg of calcium and 4 mg of iron. Clearly, most of the green leafy vegetables in the present study represent impressive sources of protein and minerals.The same can be said for the ability of these indigenous green leafy vegetables to provide the human diet with the essential fatty acid a-linolenic acid. The daily recommended intake of a-linolenic acid for an adult is about 1 g (Spaaij and Pijls, 2004)). 20 g (dry weight) of Hibiscus cannabinus contain 0.13 mg of a-linolenic acid. Although the quantity of a-linolenic acid in the leaves of the seven plants we analyzed in the present study represents only 10-15 percent of the daily requirement for an adult, the fatty acid profiles of all of these leaf species is advantageous vis-a-vis human nutrition because the ratio of linoleic acid/a-linolenic acid in each instance was less than 1/1. To put this low ratio in perspective, consider that the linolenic acid/a-linolenic acid ratio in most western diets is > 5/1 and commonly in the range of 10/1 to 20/1. A high linoleic acid/a-linolenic acid ratio is widely regarded as pro-inflammatory and generally unhealthful, especially with regard to cardiovascular disease and certain cancers (9,10).The amounts of certain nutrients in the leaves of Hibiscus sabdarifa gathered in Ghana are similar to values reported for the corresponding vegetable obtained in the Republic of Niger (Spaaij and Pijls, 2004; Freiberger et al., 1998). For example, the protein content of Hibiscus sabdarifa leaves gathered in the Republic of Niger and Nigeria was in the range 16.9-22.8 percent; however, the methionine plus cysteine content of Hibiscus sabdarifa leaves from Ghana was only about half that of the same species from the Republic of Niger. In all other respects, the amino acid compositions of the proteins from the two countries were similar. Although the total amount of fatty acid in Hibiscus sabdarifa leaf grown in Ghana was only about one-half that of the same green leafy vegetable from the Republic of Niger (Freiberger et al., 1998; Sena et al., 1998), the fatty acid profiles (expressed on a fatty acid-percentage basis) were similar in that a-linolenic acid was the most abundant fatty acid in both sets of specimens and the omega-6/omega-3 ratio was well below 1/1.As for minerals and trace elements, leaves of Hibiscus sabdarifa grown in Ghana and those obtained in the Republic of Niger contained similar amounts of calcium and magnesium (Freiberger et al., 1998; Sena et al., 1998; however, the leaves of Hibiscus sabdarifa gathered in Ghana contained only half or less as much iron, copper, manganese and zinc than those from the Republic of Niger.Relative to nutrient content data reported by Ndlovu and Afolayan (Ndlovu and Afolayan, 2008) for Corchorus olitorius grown in South Africa, the leaves of this same species that we collected in Ghana contained about 30 percent more protein, half as much lipid, one-third as much calcium but similar amounts of zinc and iron.The literature is lacking with respect to reports of the nutrient content of Hibiscus canabinus and Amaranthus cruentus against which we might compare the results we obtained in the present study.In light of the high incidence of iron-deficiency anemia (Agyei-Frempong et al., 2001), calcium-deficiency rickets (Thacher et al., 1999), and deficiencies of zinc and other micronutrients in Ghana and many other regions of sub-Saharan Africa (Lutter and Rivera, 2003; Adu-Afarwuah et al., 2008), the information in this study should serve to underscore the value of certain readily available and economically accessible, but heretofore underappreciated, green leafy vegetables which could contribute to the prevention of these and many other health problems that affect vulnerable populations (e.g. infants and young children, pregnant and lactating women).A limitation of the present study is that we did not investigate the content of anti-nutrients such as protease inhibitors (e.g. trypsin inhibitor), metal chelating substances (e.g. phytate, tannins) and fiber that can reduce protein digestibility (Pirie, 1978). In addition, we did not use in vitro protease-based methods to assess the digestibility of the proteins in the seven Ghanaian plants, nor did we measure the bioavailability of the amino acids in these proteins. Bioavailability studies in general must be performed in order to obtain a more complete picture of the actual value of the nutrient data provided in the current report. The bioavailability of proteins and minerals and trace elements will be the focus of future studies.\n",
"role": "user"
},
{
"content": "Purpose: - The main purpose of this paper is to determine the content of amino acids, fatty acids and minerals in seven indigenous leafy vegetables (ILVs) in Ghana.\nDesign/methodology/approach: - Leaves from plants growing near Kumasi were milled to a fine powder, dried to constant weight in a vacuum desiccator, and analyzed for their content of the afore-mentioned nutrients. The plants were: Hibiscus sabdarifa, Hibiscus cannabinus, Amaranthus cruentus, Corchorus oliforius, Solanum macrocarpon, Xanthomosa sagittifolium and Vigna unguiculatus.\nFindings: - All seven ILVs contained a large amount of protein (15.5-22.8 percent), which compared favorably to the essential amino acid pattern of a WHO standard. They all contained nutritionally useful amounts of a-linolenic acid and had an omega-6/omega-3 ratio of 0.1-0.9. The seven ILVs contained quantities of calcium, copper, iron, magnesium, manganese, molybdenum and zinc that could contribute significantly to satisfying an individual's need for these elements.\nResearch limitations/implications: - The presence of relatively large amounts of various nutritionally essential macro- and micronutrients in these seven ILVs does not necessarily mean these nutrients are bioavailable. Future research is required to determine the amounts of anti-nutrients (e.g. protease inhibitors, chelators) in these vegetables, and the extent to which their protein, lipid and mineral constituents are digested and/or absorbed.\nOriginality/value: - Since malnutrition (e.g. iron-deficiency anemia, rickets, zinc deficiency, protein-calorie malnutrition) is common in sub-Saharan Africa, the information which is provided should increase awareness among agricultural and public health officials of the nutritional value of seven underappreciated and underutilized ILVs that are indigenous to Ghana and many other parts of Africa.\n",
"role": "assistant"
}
] |
The liability of newer than newness: aspiring entrepreneurs and legitimacy
|
[
"Entrepreneurship",
"Crowdfunding"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The negative correlation between organizational age and mortality is one of the most well-established relationships in the social sciences (Thornhill and Amit, 2003). Stinchcombe (1965, p. 148) labeled this condition the \"liability of newness\" in offering the general rule that \"a higher proportion of new organizations fail than old.\" The liability of newness arises as aspiring entrepreneurs experience both internal challenges, such as learning new roles and establishing routines, and external challenges, such as building social networks and establishing legitimacy in the eyes of key stakeholders and resource providers. Early empirical examinations of aging effects on organization mortality rates supported both a monotonic negative effect (Carroll, 1983; Freeman et al., 1983) and a non-monotonic, effect termed the liability of adolescence (Bruderl and Schussler, 1990; Fichman and Levinthal, 1991; Mahmood, 2000). Recent advances in the literature have clearly outlined that very new organizations face more challenges than existing entrepreneurial ventures when attempting to obtain legitimacy and acquire resources for their entrepreneurial efforts (for a review, see Uberbacher, 2014), and that they face a legitimacy threshold that they may have to cross in order to remain viable (Rutherford and Buller, 2007; Rutherford et al., 2016).\nTheory and hypotheses: Liability of newness\nMethods: Research setting\nResults: Descriptive statistics\nDiscussion: Stinchcombe (1965) ushered in a wave on important and rigorous research concerning the relationship between organizational age and mortality when he introduced his ideas about the liability of newness. Subsequent studies validated his initial observations of a negative correlation between organizational age and mortality, a relationship that has become one of the most well established in the social sciences (Thornhill and Amit, 2003). However, over the past decade scholars have increasingly called for more attention to the earliest days of an organization's life because of concerns that the literature on the liability of newness has veered off course from Stinchcombe's original insights by focusing existing rather than emerging businesses (Abatecola and Uli, 2016; Aldrich and Yang, 2012; Zachary and Mishra, 2011). In this study, we follow advice offered by Aldrich and Yang (2012) to help return the emphasis of liability of newness research to Stinchcombe's (1965) original arguments.\nLimitations: Before concluding, we should note a few limitations of this study. We find mixed evidence for the ability of aspiring entrepreneurs to buffer the liability of newness. While the results of the approaches to examine the interaction between loan requests for new businesses and creditworthiness both produce effects in our hypothesized direction, only our analysis through the sub-group models provides statistically significant support. As noted, the differences in statistical significance between the two analyses may be due to inconsistent covariate effects between the two sub-groups, aspiring entrepreneurs and individuals raising money for their existing businesses. Still, this interesting finding requires further study and offers an exciting opportunity for future research to untangle the reasons for the discrepancy.\nImplications: The liability of newness is one of the most pervasive and well-established notions in entrepreneurship and the social sciences (Thornhill and Amit, 2003), although emphasis has been on existing rather than emerging businesses. Our study fills that gap by extending the general theory of the liability of newness to the context in which it was originally intended: the emergence of new organizations. Moreover, we uncover a potential buffer against liabilities associated with generating and managing support from strangers. In particular, the results suggest that creditworthiness may be of strategic value to aspiring entrepreneurs when garnering support for their new business ideas, especially aspiring entrepreneurs who rely heavily on bootstrapping to finance their new business ideas or grow their new businesses.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to return to Stinchcombe's original emphasis on emerging vs existing organizations by examining the cognitive legitimacy challenges aspiring entrepreneurs face vis-a-vis entrepreneurs with existing businesses.\nDesign/methodology/approach: The data collection included content analysis of profiles of an online crowdfunding, peer-to-peer lending market leading to a sample of 507 business loan requests, 123 of which were requests to support new business ideas rather than existing businesses. Negative binomial regression was used to test hypotheses regarding whether aspiring entrepreneurs seeking convenience-based support for their new business ideas would be less successful than their counterpart entrepreneurs seeking support for their existing businesses.\nFindings: The findings show that aspiring entrepreneurs received less convenience-based support for their new business ideas from key resource providers than their peer entrepreneurs asking for support for existing businesses. The findings also suggest that this liability of newer than newness may be able to be mitigated by reputational signals such as the creditworthiness of the entrepreneur making the request.\nOriginality/value: This study focuses on the original insights Stinchcombe introduced when he described the social conditions that produce the liability of newness. Moreover, this study offers explicit theory as to the key mechanisms that cause the liability of newness by focusing on an aspiring entrepreneur's ability to secure convenience-based support and potential ways an aspiring entrepreneur may offset that liability.\n",
"role": "assistant"
}
] |
Analysis of production planning in a global manufacturing company with process mining
|
[
"Process mining",
"SAP",
"Heuristic miner",
"Material management",
"Production planning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Today's highly competitive business environment forces companies to collaborate with trading partners in their supply chain to provide products and services at the right time, of the right quality and at acceptable cost. One of the key business processes towards succeeding in this highly volatile business environment is production planning (PP). PP is the process of ensuring that production occurs in a smooth and timely manner, satisfying a schedule made as a part of the business strategy of a company. The aim of this process is to maximise profits, fulfil demand, attract and retain customers.\n2. Process mining: Process mining is a term describing the tools and techniques for extracting a structured process description from a series of real-life executions (van Der Aalst et al., 2009). The basis of process mining is the analysis of event logs (van Der Aalst et al., 2004), which are records of actual actions and their associated information, such as the time of execution. This analysis gives insight into real control-flow dependencies, data and resource utilisation, and various statistic measures. One of the end products of process mining is a description of the process in the form of a model.\n3. Methodology: The methodology used in this study follows three out of five stages of the L* lifecycle model proposed by van Der Aalst (2011): plan and justify, extract and create control flow model. In the first phase, it is determined that the process mining is driven by the questions related to PP process in the case company. The case company experiences heavy re-scheduling in PP. The company has some ideas that constant adjustment to prognosis from the headquarters may influence other related process such as MM and production. However, they have not performed an in-depth analysis. Based on this, process mining is used to answer the following questions:\n4. Results: 4.1 Case description\n5. Analysis: This section describes results and the analysis of the study structured according to the research questions.\n6. Discussion: Several findings from the research in the case company's PP process need to be discussed further. The discussion will focus on indication of schedule instability, the cause, implication and potential solutions to schedule instability problem found in the case company.\n7. Practical and research implications: This research has several practical and research implications. For the case company, the results of this study demonstrate the implication of accommodating changing forecasts to the company's PP and MM, which may not have been investigated before.\n8. Conclusion: This paper demonstrates the use of process mining in modelling and analysing the event log from PP and MM of ERP systems in a real manufacturing company. The cases in the event log are at the material level in order to be able to tie the PP and MM processes. The results of process mining with Heuristic Miner show that there are many more activities and possible paths in the real PP process compared to the standard guideline determined by the case companies.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present the result of using process mining to model the production planning (PP) process of a manufacturing company that is supported by enterprise resource planning (ERP) systems.\nDesign/methodology/approach: This paper uses event logs obtained from the case company's ERP database. The steps for this research are planning process mining implementation, extraction and construction of event log, discovering process model with Heuristic Miner and analysis.\nFindings: Process model obtained from process mining shows how the PP is actually conducted. It shows the loop in materials requirement planning and create plan order process. Furthermore, the occurrences of changing plan order date and production line indicate the schedule instability in the case company. Further analysis of the material management (MM) event log shows the implication of production plan changes on MM. Continuous change in the plan affects material allocation priority and may result in a mismatch between production needs and the materials available.\nResearch limitations/implications: The study is only conducted in a single and specific case. Therefore, even though the findings provide good insight, the use of solitary case study does not imply a general result applied to other cases. Hence, there is a need to conduct similar studies on various cases so that a more generic conclusion can be drawn.\nPractical implications: The result provides insights into how the current company's policy of adjusting the production plan to accommodate changing demand impacts their operation. It can help the company to consider a better balance between flexibility and efficiency to improve their process.\nOriginality/value: The paper demonstrates the use of process mining to capture the real progression of PP based on the data stored in the company's ERP database, which give an insight into how a real company conducts their PP process, the implication of schedule instability on MM and production. The novelty of this research lies in the use of process mining to attest to the schedule nervousness issue at a process level.\n",
"role": "assistant"
}
] |
Completing the circuit in e-government process automation: A simple cryptographic watermarking scheme for online hand signatures and stamps
|
[
"Electronic commerce",
"Government",
"Digital signatures",
"Automation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The main purpose of automating government services is to render these services available remotely (online) for all types of customers, through a variety of online electronic devices. This saves time and energy, increases efficiency and improves environment conditions. Imagine a government agency providing services for the public with zero visitors and an empty car park. During the past few years, a lot of work has been done to provide automated government services (e-government) in many countries around the world. Despite the efforts to automate, there are some situations where a signed/stamped physical document is required. The customer starts nicely by paying the fees and filling the forms online, but has to present himself personally to get a stamp or a signature. Take for example, the visa process used by specialized ministries and embassies. Theoretically, the customer can submit his application and pay the fees online, but has to present his passport physically to get the visa sticker. Another example is the authentication stamp service required to certify the authenticity of documents (diplomas, for example). Again the customer can apply for the service online, but has to present the document physically to the certifying authority in order to get the stamp. This defeats the main purpose of automation and makes the \"zero visitors, empty park\" goal a far quest.In this paper, a simple scheme for performing remote (online) physical signatures and stamps is suggested. The main motivation is to complete the service circuit and avoid the requirement that customers should present documents physically for signatures or stamps. The suggested scheme is based on digital watermarking and steganography (Petitcolas et al., 1999). Digital watermarking received considerable attention over the past decade and is currently used extensively an image and video authentication.The following sections describe the suggested scheme after giving a brief overview of digital watermarking.\nRelated work: The past few years witnessed considerable progress in cryptographic authentication and authorization techniques. \"Digital signatures\" or \"Electronic signatures\" are well understood and widely used in many applications today. This paper, concentrates on physical signatures and stamps, which are authentic images of the original signatures or stamps. Physical signatures or stamps are still required by many applications.Up to our current knowledge the problem of performing physical signatures or stamps online (remotely) has not been directly addressed. We are aware only of a short web reference to a patent by the University of California, Davis, indicating that researchers at the university have created a new method for remote signature/attestation of death certificates that uses a person's voice as a biometric signature (http://research.ucdavis.edu/tia). The details of the method are not given.However, a lot of work has been done in a related area, namely image and video authentication. A number of methods for authenticating images and videos have been published over the past few years. These methods have been well received in the market and are now widely used by publishers and vendors. Most of the published methods are based on digital watermarking and steganography (Petitcolas et al., 1999). Since, the scheme described in this paper is based on digital watermarking, the following section will give an overview of this technique.\nDigital watermarking: Digital watermarking is an information hiding technique used to hide a mark (image, audio signal, serial number, text, etc.) in a given image (carrier or stego image) (Petitcolas et al., 1999). The hidden object is the watermark while the carrier is the marked object. Fragile watermarks are destroyed if the carrier image is subject to considerable modifications. Robust watermarks cannot be removed from the image unless the image is totally destroyed. Robust watermarks resist all types of known modifications. Invertible watermarks can be recovered from the carrier image using a given key, while non-invertible watermarks can be detected but not recovered. Fingerprints are a special type of hidden features (serial numbers for example) used by intellectual property owners to track the customers who break the license by distributing the property to third parties. The fingerprint is usually a hidden serial number or similar feature that characterizes a particular entity.Figure 1 shows the general embedding process used in most watermarking systems. Given an image I, a watermark M and a key K, the embedding process may be thought of as a transformation on the image I that modifies its digital content without affecting its visual appearance in an obvious way. The key is usually a seed for a random number generator used to distribute the bits of the watermark within the carrier image.The details of the embedding process and the use of the input parameters (I, M, K) depends on the type of watermarking technique.Watermarking techniques are classified into two major classes (Petitcolas et al., 1999):1. Private marking systems. Private marking systems (Barni et al., 1998; Caronni, 1995; Cox et al., 1996a, b, c) require that at least the original image I is available in order to extract or detect the watermark (Figure 2). In class A of these systems the original image is used to find the location of the mark in the marked image. In class B, both the original image and the mark are needed to prove that the mark is embedded in the marked image.2. Public marking systems. In public marking systems (Hartung and Girod, 1998), neither the original image, nor the mark is required to extract the mark from the marked image. Public marks are more useful than other types of marks. The work described in this paper is based on public marking systems and the underlying techniques.Marking algorithms\nA simple scheme for remote signatures and stamps: In this section, a simple scheme for generating remote physical signatures and stamps is presented. The section starts by stating the main conditions for acceptable remote signatures and stamps (RPSS), followed by a high-level view for the software architecture of a working system. After that, a number of suggested solutions are presented.The main conditions for acceptable RPSS:\nDiscussion: There are two major issues relating to the acceptability of remote physical signatures and stamps: authenticity and the possibility of reusing the signature illegally (duplicating the signature). These issues are resolved as follows by the above scheme:1. Authenticity. Because all watermark components are encrypted using the private key of the signer, this proves that the signature is authentic. The hidden watermark also includes detailed information about the signer, including the serial number of the device from which the signature operation has been performed.2. Signature duplication and reuse. A general solution for this problem is a bit difficult, but solutions for specific application domains may be found. The above scheme provides a number of precautions that may be used to formulate solutions for the duplication problem under certain restrictions:* The remote signature/stamp is generated using a secure protocol that maintains information about the signer and the requestor. This information is included within the watermark pack. The destination authority should check the identity of the person who presents the document, and if it differs from the one included in the watermark, the document is rejected. The requestor is accountable for any misuse of the signature.* The main features of the signed document are included in the watermark. If the features of the presented document are different from those in the watermark, the document is rejected.* It is advisable that the signed document is characterized with a unique sequence number which is included in the watermark.* The signature/stamp is associated with a unique sequence number and timestamp information. This guarantees the freshness of signatures/stamps and that only one version is used.* If the application domain allows, active agents may be used to maintain a copy count each time the document is presented. Such agents may be embedded in the cover image.\nA simple example: To illustrate the process, I embed my signature in a given image using a typical watermarking algorithm. Figure 3 shows the signature image, Figure 4 shows the carrier image before embedding the signature and Figure 5 shows the image after embedding the signature. The process has been used to embed a signature image, a text file and an executable. All items have been successfully recovered from the cover image. Figure 6 shows the recovered hand signature.\nConclusions and recommendations: A simple, secure scheme for RPSS has been suggested. The scheme is based on the principles of digital watermarking and public key cryptography. It has been demonstrated that physical signatures produced using the scheme are authentic and are as good as direct signatures and stamps if properly implemented. RPSS saves time and effort and improves the environment. More work is needed to verify the integrity of the approach and investigate the hardware and software requirements to streamline the process. Legal and standardization issues should also be investigated and studied.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to provide a technique that helps in completing the automation cycle in e-government and e-business transactions. Some e-government and e-business transactions require the physical intervention of a user or an employee to place a physical stamp or a hand signature on a document. This paper suggests a digital watermarking approach for online hand signatures and physical stamps. The suggested technique may be used by individuals or organizations to place hand signatures and stamps on remote documents.\nDesign/methodology/approach: - The work started by surveying existing security protocols and authentication techniques. Digital watermarking techniques have then been deeply studied to investigate their security capabilities and effect on the performance of the system. A scenario-based approach has then been used to derive a number of secure digital watermarking protocols that may be used to place remote physical signatures and stamps. The derived protocols were then audited and tested.\nFindings: - It is possible to design and build simple light watermarking protocols for placing authentic hand signatures and stamps on remote documents (online). The process is secure and reliable.\nPractical implications: - The suggested technique will greatly improve and speed e-government and e-business transactions.\nOriginality/value: - According to one's own knowledge and the surveys so far conducted, the suggested system is the first of its kind.\n",
"role": "assistant"
}
] |
Leading learning in Asia - emerging empirical insights from five societies
|
[
"Principals",
"Leadership",
"Asia",
"Instructional leadership"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The field of educational leadership and management (EDLM) is in the midst of a paradigm shift that has only recently begun to be acknowledged widely. This paradigm shift centers on the transformation of the field from a largely Western, Anglo-American literature to a global field of study (Clarke and O'Donoghue, 2016; Hallinger, 2011, in press; Hallinger and Bryant, 2013a, b; Mertkan et al., 2016; Oplatka, 2006; Yalcin et al., 2016). Prior to 2000, virtually all reviews of EDLM research published in our journals were based on studies located in a narrow set of Western Anglo-American societies (e.g. Bossert et al., 1982; Bridges, 1982; Campbell and Faber, 1961; Erickson, 1967; Hallinger, 2014; Hallinger and Heck, 1996; Lipham, 1964).\nConceptual background of the ILEA studies: EDLM scholars in East Asia have responded proactively to the challenges outlined above (Hallinger et al., 2005; Walker and Hallinger, 2015). Initially, this took the form of largely uncoordinated efforts to increase the output of empirical studies (Hallinger and Bryant, 2013a). As the volume of EDLM research conducted in English as well as \"local\" languages began to accumulate, a group of scholars, operating under the auspices of the ILEA project, undertook a series of foundation reviews of research within (e.g. Hallinger and Truong, 2014; Hallinger et al., 2013, 2015; Ng et al., 2015a, b; Pan et al., 2015; Walker et al., 2012; Walker and Qian, 2015) and across Asian societies (Hallinger and Bryant, 2013b; Walker and Hallinger, 2015). These reviews of research were conducted with the goal of revealing what had already been discovered about the EDLM practices \"within these societies\" as well as identifying commonalities and differences across the societies. A central challenge in the minds of these reviewers was linking leadership and management practices to the context in which they were found (Walker and Hallinger, 2015). Thus, the role of \"context\" loomed large in our conceptual lens (see Bajunid, 1996; Clarke and O'Donoghue, 2016; Dimmock and Walker, 1998, 2000; Hallinger, 2016).\nILEA: commonality and distinctiveness: The societies encompassed in this study share contextual similarities as well as differences. \"Contexts for leadership\" have been analyzed in terms of culture, levels of economic development, political systems and organizational structures (Hallinger, 2016). Each of these contextual features contributes toward shaping the school context in which principals work and their approaches to leadership. Moreover, we note that each society evidences a particular combination of features.\nReflections on instructional leadership in eastern and western contexts: This synthesis has sought to identify similarities and differences in instructional leadership practices of primary school principals across five East Asian societies. We have suggested that the particular role set accorded to principals within each society can be traced back, at least in part, to features of the social-political-economic context in which schooling takes place. With the exception of Malaysia these national contexts share a common Confucian cultural heritage which exerts a continuing influence on attitudes toward education as well as broader social relationships. At the same time, these societies evidence different levels of national economic development (e.g. from highly developed Singapore and Taiwan to less economically developed Vietnam and China) as well as political systems.\nConclusion: We close with a note on the process of engaging in this collaborative research venture into studying educational leadership across East Asian societies. The project began in 2010 with the step of developing a regional agenda for research in educational leadership. This led to the first stage in which we developed national reviews of EDLM research. These reviews made findings from research studies written in local languages available to the global community for the first time. Indeed, we expect these reviews of research published in 2015 to lay the foundation for future scholarly work in these societies. This second stage of qualitative research offers new directions on which other scholars can build.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to synthesize findings from studies of principal instructional leadership conducted in five East Asian societies. The authors first identify similarities and then differences in approaches to instructional leadership across the societies. Then the findings of the synthesis are compared with broad findings from the global literature on principal instructional leadership.\nDesign/methodology/approach: The paper employs a thematic approach to synthesizing findings from the five qualitative studies.\nFindings: The authors identified numerous similarities in practices of instructional leadership across the five societies. These included first, a top-down approach to defining the mission and goals of schools whereby principals worked within a fairly narrow zone of discretion. Second, principals devoted relatively little attention to coordinating the curriculum due to working within strict national curriculum frameworks. Third, principals executed their instructional leadership practices with an ever-present sense of the need to honor hierarchical relations and maintain harmony among staff and other stakeholders. Differences across the five societies centered on the extent to which the instructional leadership role of principals was explicitly defined and the extent to which they received training for the role.\nOriginality/value: This synthesis sought to build upon reviews of research published in a special issue of this journal two years ago. The synthesis and this body of research papers have contributed toward moving empirical research on educational leadership broadly, and instructional leadership in particular, forward in East Asia.\n",
"role": "assistant"
}
] |
Resource deployment under consideration of conflicting needs in times of river floods
|
[
"Decision making",
"Flood response",
"System dynamics model",
"Conflicting needs"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Disasters are characterized by severe consequences to society, causing fatalities, loss of property or damage to environment and are exceeding the population's capability to cope with it (Kiss et al., 2014). According to Doll et al. (2014), future changes in precipitation and flood patterns highly influence weather-related delays in transportation. In disaster response, the key logistical issues result from limitations in transport resources and damages to infrastructure (Balcik et al., 2008). Hence, during disasters, the transport sector is critical (Bil et al., 2015). Therefore, it is crucial to better understand the influences of transport disruption on supplying affected areas with commodities, especially as an increasing number of disasters and higher impacts occur in Europe (EEA, 2011). In contrast to commercial supply chains, humanitarian relief is coping with victims suffering from delivery delays and shortages (Chakravarty, 2014). To avoid panic reactions, supplying the affected population is crucial, however, suddenly increasing and unpredictable demand due to uncertainties in post-disaster environments challenges decision makers (Balcik and Beamon, 2008; Safeer et al., 2014). According to Terpstra and Gutteling (2008), the households' perceived responsibility concerning flood risk is low. Hence, sudden shifts toward extreme demands for certain goods or to certain areas may occur due to limited household preparedness. These are main reasons for the uncertainties of decision making (Kumar and Havey, 2013). Baker (2011) show that a hazard threatened population reacts by stockpiling food and water. An increasing demand as a panic reaction is linked to the people's observation of decreasing commodity availability at stores due to supply disruptions and limited reaction time for retailers (Cavallo et al., 2014). During Hurricane Katrina in 2005, people mainly stocked non-perishable food, bottled water, cleaning supplies, flashlights, candles, first-aid supplies, generators, batteries and ice (Katrandjian, 2011). Even though an appropriate forecast was made in the case of Hurricane Katrina, decision-makers reacted ineffectively, which resulted in unsatisfied needs (US House of Representatives, 2006). Holguin-Veras et al. (2012) state that in case of a disaster, supplies have to be delivered for victims, and additionally, for the response process. In the case of the Tohoku disaster in 2011 in Japan, an unexpected demand of supplies occurred because of the combined needs of the affected victims and the needs for conducting relief operations (Holguin-Veras et al., 2014). This challenges decision makers and inadequate ordering patterns may occur due to the complexity of demand and underestimations concerning inventory management (Barlas and Ozevin, 2004). In times of disasters, many retailers order emergency supplies reactively to a sudden increase in demand (Lodree et al., 2012). Regarding inventory control, this wait-and-see approach is currently applied primarily in practice as opposed to pre-positioning inventory before being hit by a disaster (Lodree et al., 2012).\n2. Method: SD, first applied in Forrester (1961), aims to reveal the system's behavior, which is characterized by complexity (Morecroft, 2007). For the SD model, which consists of a qualitative (causal loop diagrams (CLDs)) and a quantitative (stock flow (SF) model) part, various elements have been determined. Table I provides an overview of the most relevant elements, which are used in the model.\n3. Model behavior and numerical investigations: For the computational experiments, we differentiate between excluding (variation 1) and including (variation 2) the additional demand of relief units and evacuees and vary the beginning of the hoarding behavior (scenario A) as well as the initial inventory (scenario B). Moreover, the impact of a low- and high-increase rate of the water gauge is analyzed for each scenario. An overview of the calculated scenarios is given in Table II.\n4. Conclusion: Our results show that a possible stock-out as a result of conflicting needs for commodities lasts longer the later the population and, consequently, the system reacts to a decreasing availability of commodities on site. If the system starts to adjust the number of vehicles at a later point in time, it takes longer to reach a new equilibrium state that is essential to satisfy the demand. Furthermore, in some cases, no equilibrium state is reached due to inadequate ordering patterns. This is a consequence of underestimations in inventory management by wait-and-see strategies of decision makers. Therefore, preventive measures, as awareness raising within the population are of importance to avoid hoarding behavior resulting from insufficient household preparedness. We found that a lack of quantitative data concerning hoarding behavior of the population in case of floods appears. Future research to analyze the behavior of the affected population regarding stockpiling and preparedness efforts is of high importance. Furthermore, investigating the reasons of limited household preparedness is of interest. Moreover, expanding the model to address additional questions such as the needed number of facility locations to better meet the additional needs in the affected region would be a field for future research. The presented SD model contributes to an extended understanding of the impact of various factors and their interdependencies in the context of resource deployment under a sudden demand. Furthermore, supplying the evacuated population and the deployed relief units in the affected region leads to additional needs under limitations concerning transport conditions at the same time. The presented work shows that applying SD to issues of flood response is of value as the system's approach allows one to analyze different scenarios for strengthening understanding of the system's complexity.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present a system dynamics (SD) model that allows one to simulate resource deployment to fulfill increasing needs for commodities such as food and other consumables during disaster situations. The focus is on managing a suddenly increased demand (hoarding behavior) of an affected population under restricted transport conditions. The model aims to support decision makers by fostering comprehension of the systemic behavior and interdependencies of those complex settings.\nDesign/methodology/approach: Through literature review and case study analyses the SD model was established and implemented with STELLA 10.1.1.\nFindings: The needs of relief units for response operations and supply of evacuees in the affected region result in conflicting needs under limited transport conditions during disaster situations. Therefore, uncertainties and dynamic parameters as, e.g., occurring delays, limited information, or delivery constraints and their influence on resource deployment under a sudden demand, have been identified and incorporated in this work. The authors found that an oscillating behavior within the system is possible to occur and is more intensified in case of regarding the additional needs of evacuees and relief units.\nResearch limitations/implications: Due to the high level of abstraction, it is not possible to incorporate all influencing variables in the SD model. Therefore, the authors focused on the most important ones with regard to the model objective.\nPractical implications: To focus on awareness raising is of importance for decision makers in the context of disaster management. Furthermore, the authors found that the oscillating behavior is more irregular in case of assuming a higher increase rate of the water gauge than if a low increase rate is assumed.\nOriginality/value: To the best of the authors' knowledge, none of the work already done refers to providing a flood-prone area with commodities under consideration of a sudden demand, by applying the SD approach. The presented model contributes on the generation of systemic insights of resource deployment under consideration of conflicting needs in times of a river flood to support decision makers in those situations.\n",
"role": "assistant"
}
] |
Listening differently: youth self-actualization through critical Hip Hop literacies
|
[
"Critical literacy",
"Hip Hop pedagogy",
"Adolescent identity",
"Youth epistemologies"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Drew:\nConceptual frameworks: Youth epistemologies and Hip Hop culture\nFoundations of Hip Hop-based education: Over the past three decades, HHBE has increasingly found its way into K-12 and post-secondary classrooms. Morrell and Duncan-Andrade's (2002) action research project was one of the first to explore the use of Hip Hop for the promotion of academic and critical literacies in a high school English classroom. By juxtaposing the study of traditional poetic works with Hip Hop texts in a curricular unit, the authors found that students were able to relate to Hip Hop texts presented as poetry through the resonance of particular themes and recognized the space that a study of Hip Hop texts provides for critical discussion of literary texts.\nMethodology: Site and sample\nFindings: \"I Don't know who I am\": Hip Hop and adolescent identities\nDiscussion and implications: As McLean (2005) argues, part of the process of identity formation for young people occurs through the act of sharing stories and making meaning of the past through reflection. Baszile (2009, p. 16) described Hip Hop culture as a site for counter stories and young people's engagement with this culture as \"both a rearticulation and a resistance to the stories, the identities we tend to impose upon them through the official school curriculum.\" For Drew, the counter stories present in Hip Hop culture offered a framework for hope and resistance. In her autobiography, Sonya engages in what Baker-Bell (2017, p. 531) describes as \"Black feminist-womanist storytelling,\" in that she put her \"racial, gendered, and sometimes classed experiences in conversation with one another through storytelling.\" Mark used his autobiography to reflect on his process of becoming and the influence of his family environment on that process. As literacy development through the lens of Hip Hop \"builds upon youth's lived experiences whilst evoking an imagining of identities\" (Watson and Marciano, 2015, p. 38), CHHL can serve to activate reflections on the past and imaginings of the future.\nConclusion: Previous research on HHBE in the English classroom has documented how Hip Hop and the study of Hip Hop can validate youth identities in school and support teachers' understandings of their students by creating a space for youth to share their stories in class (Hill, 2009; Morrell and Duncan-Andrade, 2002). The results of this study expand on this research by documenting how developing CHHL in the English classroom can also support students' understandings of themselves, not only as Hip Hop-identified youth, but also as agentive, self-actualizing youth with conflicting, intersectional identities. While a CHHL framework may not resolve the inequities that youth face in schools and in society, it can provide tools for their development of theories and strategies for understanding and resisting this inequity.\n",
"role": "user"
},
{
"content": "Purpose: This study aims to refocus the field of Hip Hop based education on youth identities and epistemologies rather than on the tangible artifacts of Hip Hop culture. It argues that centering classroom pedagogy and curriculum on youth self-actualization best supports the critical literacy development of students grappling with social and structural inequities within an ever-evolving youth and media culture.\nDesign/methodology/approach: Building upon previous literature on critical literacy, Hip Hop pedagogy and adolescent identity formation, this paper shares data from a semester-long teacher-researcher case study of a high school Hip Hop literature and culture class to explore how young people develop critical literacies and self-actualizing practices through a critical study of youth culture.\nFindings: For youth engaged in Hip Hop culture, co-constructing spaces to discuss their consumption of popular media and culture in class allows them to openly grapple with questions of identity, provide support for each other in dealing with these questions and reflect more critically upon their self-constructed, performed and perceived identities.\nOriginality/value: This form of English education challenges traditional notions of teaching and learning as it positions students as co-creators of curriculum and as part of the curriculum itself. Building on research that frames Hip Hop pedagogy as a culturally relevant tool for engaging urban youth, this paper argues that educators should approach critical Hip Hop literacy development as a means by which young people across diverse educational and social backgrounds come to know themselves and others as part of the process of self-actualization and critical resistance.\n",
"role": "assistant"
}
] |
Examining the relationship between the economic environment and restaurant merger and acquisition activities
|
[
"Cointegration",
"Economic environment",
"Error correction model",
"Merger and acquisition",
"Restaurant industry"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Merger and acquisition (M&A) has been widely used by the restaurant industry as an important growth strategy (Park and Jang, 2011). According to the Securities Data Corporation (SDC) database, the number of completed M&A deals in the US restaurant industry increased from 534 in the 1980s to 898 in the 2000s and the total value increased from US$34.9bn to US$56.5bn during the same time. Responding to this substantial increase in M&A deals, a growing body of literature has focused on M&A deals in the hospitality industry context (Kim and Zheng, 2014). However, these studies (Chatfield et al., 2012; Oak et al., 2010; Sheel and Nagpal, 2000) have primarily focused on the wealth gains or the difference in gains according to various deal characteristics such as financing method.\nLiterature review: Current economy\nMethodology: Data\nEmpirical results: Descriptive statistics\nDiscussion and conclusions: This study conducted a series of analyses to explore the relationship between the economic environment and M&A activity in the US restaurant industry. First, the long-term equilibrium among variables was examined using cointegration. Then, the short- and long-term effects of individual economic variables were examined using ADL-ECMs. Finally, the three estimated ADL-ECMs were compared with respect to their model fit. The findings of this study demonstrate that the economic environment affects restaurant M&A activity based on the long-term equilibrium relationship between restaurant M&A activity and four economic conditions. Such a finding implies that restaurant acquirers have made M&A decisions rationally by taking economic conditions into consideration. However, this equilibrium relationship is achieved along with short-term irrational decision errors which may result from poor management decisions. According to the herding theory, some managers tend to impetuously follow the successful actions of industry leading firms without careful analysis of the return on the actions (Martynova and Renneboog, 2008). In other words, some restaurant firms fail to effectively use information regarding critical economic conditions in their M&A deals; consequently, the relationship between restaurant M&A activity and economic conditions temporarily deviates from the long-term equilibrium. However, such M&A deals tend to produce unsuccessful outcomes, reducing irrational management decisions, and the deviation is therefore adjusted to achieve long-term equilibrium over time.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to determine whether the economic environment affects the merger and acquisition (M&A) activities in the restaurant industry.\nDesign/methodology/approach: The M&A transactions in the restaurant industry between 1981 and 2013 (n = 1,415) were examined. Data were collected from the Securities Data Corporation (SDC) database. Using an autoregressive distributed lag approach, this study developed three error correction models to explore the short- and long-term relationships between restaurant M&A activities and four macro-economic factors.\nFindings: This study found that there was a long-term equilibrium relationship between the M&A activities and the four economic factors and that economic outlook had a significantly positive impact in the long term, while the effect of cost of debt was significantly negative in both the short and long terms. The findings suggest that restaurant firms are more likely to adopt M&A strategy when they are optimistic about the future economy and can take on debt at a low cost.\nPractical implications: The findings of this study are expected to help practitioners make informative M&A decisions in the restaurant industry taking into consideration the economic environment. They will also help investors effectively manage their portfolios by predicting and ascertaining the proper time to invest in the restaurant industry based on the changes of economic environment.\nOriginality/value: No known study has been identified that examined the relationship between macro-economic factors and M&A activities in the restaurant industry. The findings of the study are expected to fill the gap in the literature by demonstrating the economic environment and the M&A activities in the restaurant industry are in a long-term equilibrium achieved by self-correction of their short-term disequilibrium.\n",
"role": "assistant"
}
] |
Is human capital the sixth factor?
|
[
"Asset pricing",
"Human capital",
"Factor models"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In the recent past, the emerging markets show evidence of significant development and growth, which draw the attention of several investors and researchers globally to access the risk inherent in these markets. Rapid development and growth of the emerging markets give birth to various unidentified factors in the capital markets which are different from the developed markets. That gives a tough challenge to the investors and researchers to identify such unidentified factors prevailing in the emerging capital markets and makes asset pricing difficult. Over the years, several stock market anomalies have emerged and they pose a tremendous challenge to the asset pricing framework. Emergence of new anomalies overturns the CAPM's success and other multifactor models in capturing the risk-return relationship. Even Fama and French (2015) are unable to capture the risk-return relationship for the micro stocks or smallest portfolio. Likewise most of the studies (Hou et al. 2014; Clarke, 2016; Chiah et al., 2016; Fama and French, 2016, etc.) carried out across the world find that Fama-French's (2015) five-factor model is not able to capture all the risk return variations that justify the scope for the new more robust models. Most of the prominent asset pricing models whether it is CAPM or Fama-French factor models have ignored human capital in the model. Human capital is an important factor as it represents the wealth of the company, and Campbell (1996) introduces human factor as an important component of wealth. The study also argues that labour (human capital) captures approximately two-thirds of the GNP and rest is related to the financial wealth. Campbell (1992, 1996) calculates the total wealth by using the following equation:(1)\nLiterature review: The single-factor model CAPM (Sharpe, 1964; Lintner, 1965; Mossin, 1966) is a notable breakthrough in the literature of asset pricing. However, capital markets are more complicated in operation and with single-factor b, it is quite impossible to capture all the risk-return variations. Hence, CAPM was criticised and challenged by Fisher Black (1972), Fama and MacBeth (1973), Ross (1976), Banz (1981), Reinganum (1981), Gibbons (1982), Basu (1983), Shanken (1985) and Bhandari (1988). Further number of sizeable studies by Bhandari (1988), Fama and French (1988), Gibbons et al. (1989), Jeffe et al. (1989), and Lo and MacKinlay (1990) find that size of the company is another important variable along with b in explaining the expected portfolio returns. Similarly studies by Basu (1983), Reinganum (1981), etc., find P/E ratio, and Bhandari (1988), Narayanaswamy and Phillips (1987), etc., find leverage as the other important factors that explain portfolio expected return variations. Then, Fama and French (1993) introduce a more stable three-factor model from the previous studies with b, size and value. Due to the several other unidentified capital market anomalies, Fama-French's (1993) three-factor model was also criticised and challenged by several studies (Haugen and Baker, 1996; Cohen et al., 2002; Fairfield et al., 2003; Titman et al., 2004; Novy-Marx, 2013; Hou et al., 2014), etc. Finally, Fama and French (2015) introduce a five-factor model with b, size, value, profitability and investment factor that is able to capture more risk return variations than Fama-French's (1993) three-factor model. The major problem with Fama-French's (2015) five-factor model is its inability to capture the risk-return variations for the micro stock portfolios. Likewise, most of the studies (Hou et al., 2014; Clarke, 2016; Chiah et al., 2016) conducted across the world found that Fama-French's (2015) five-factor model is not able to capture all the risk return variations and is not a global model; therefore, this justifies the scope for the new and more robust model. Human capital factor is considered to be a key factor as it explains the wealth of the company and all of the studies discussed above have ignored human capital factor in the model. Campbell's (1996) study introduces the human capital factor and argues that the omission of the human capital from the model may lead to problem in the asset pricing as explained in the Introduction section. Jagannathan and Wang (1996) include the human capital return instead of the market portfolio return and find that the unconditional model (CAPM) is able to explain over 50 per cent of variations in the cross-section returns. Jagannathan et al. (1998) find that labour income is an important variable to define aggregate risk which is not captured by the traditional stock return index factor or market returns. An interesting part of the study is that labour b drives out size factor but not the value factor. Rosett (2001) finds that economically and statistically human capital (labour stock) has a direct positive relation with the equity investment risk. The study uses three standard models, both market- and accounting-based risk measures variables to arrive at the said conclusion. Qin (2002) addresses why the asset's market b is unable to capture the risk associated with the human capital. The study derives an intertemporal model that relates market b and human capital b linearly with the asset returns. Pantzalis and Park (2009) and Edmans (2011) find that human capital is related to the stock returns, but the market cannot correctly price the human capital for the small size stock due to small firm effect. In the Indian context, Shijin et al. (2012) find a causal relationship between the asset returns and labour income and his study is mainly based on the NIFTY 50 Index. The study finds that human capital is a significant factor that can affect the asset prices and adds the human capital factor to the model, as it will be helpful in predicting more variations, though the study is conducted for only market index and it does not consider different stock portfolio returns. From the above rigorous literature review, it is concluded that human capital does affect the asset prices and in the Indian context, a comprehensive study on human capital and stock returns is needed and none of the previous studies address whether portfolio returns are affected by the human capital or not in the Indian context. So, the present study will bridge the existing gap by developing a new model with human capital using a different study period and methodology from the previous studies that can capture the risk return variation better than the existing asset pricing models.\nData and methodology: The study uses monthly data of 431 companies listed in the BSE500 index and data are taken from the Bloomberg database for the period July 2004 to November 2016. Market capitalisation (MC) is used as the proxy for the size factor variable and P/B ratio as the proxy for the value factor variable. Total asset (TA) growth is used as the proxy for investment and it is calculated by the formula [(TAt-TAt-1)/TAt-1)] similar to Hou, Xue and Zhang's (2015) study. Return on equity (ROE) is used as the proxy for profitability similar to Haugen and Baker's (1996) study. Studies by Hansson (2004), Draca et al. (2011) and Bell and Machin (2016) use average wages as the proxy for human capital; in similar line, our study uses the growth of salary and wages expenses as the proxy of human capital as a measure of total investment made by company for its human development and it is calculated by the formula [(HCt-HCt-1)/ HCt-1)]. The study uses 91 days T-Bills as the proxy for risk-free rate (Rf) and BSE200 returns as the proxy for market return (Rm).\nExplanatory variables: Table I shows the descriptive statistics for the independent variables. The results show that the monthly average market premium is 0.9 per cent (t=1.461), while monthly size premium is found to be 1.3 per cent (t=5.120). The monthly value premium is of 0.6 per cent (t=1.650), while profitability premium is estimated to be 0.2 per cent (t=0.710) per month. Then, premium for investment is 0.3 per cent (t=1.179) per month. Finally, human capital premium is 0.2 per cent (t=0.910) for the respective mimicking portfolios. The average market premium of 0.9 per cent, size premium of 1.3 per cent and value premium of 0.6 per cent seem large in terms of statistics and practical terms from the investment perspective. The results are echoing the previous studies of Fama and French (1993, 2015). The average profitability premium of 0.2 per cent, investment premium of 0.3 per cent and human capital premium of 0.2 per cent seem much lower than the previous three (market, size and value) and for all the cases standard errors are less.\nEmpirical results: Figure 1 shows that small stocks earn significantly higher returns than big stocks and this phenomenal performance is pervasive in all the portfolios across the combinations (MC_PB, MC_ROE, MC_TA and MC_HC). Hence, this is the manifestation of size, value, investment and human capital effects.\nAsset pricing results: Table IV details about the FFTF regression results for MC_PB, MC_ROE, MC_TA and MC_HC cross-portfolios. Panel A of Table IV represents the FFTF regression result for the MC_PB cross-portfolios. A model is said to be a good if it can capture all a values equal to 0. The a value ranges from 0.001 to 0.014 with an average a value of 0.004 for all 25 portfolios. Four portfolios out of 25 portfolios are found to be statistically significant (a values where t(a) is more than 1.96) though the microcap stocks with low value are not captured by the model. In case of MC_ROE cross-portfolios, the a value ranges from 0.000 to 0.011 with an average a value of 0.005 for all 25 portfolios (see Panel B of Table IV). Six portfolios out of 25 portfolios are found to be statistically significant (a values where t (a) is more than 1.96) though the microcap stocks with weak profitability are not captured by the model. Similar to the Fama-French's (2015) study, our results also find that microcap is the problem in both of the above cases. Three-factor regressions for MC_TA cross-portfolios estimated results are similar to the previous two cases. The average a value calculated is 0.004 for all portfolios, while the a values lie between 0.000 and 0.011 (see Panel C of Table IV). Five portfolios are found statistically significant (a values where t(a) is more than 1.96) though the microcap stocks with conservative investment are not captured by the model. Similar to the Aharoni et al. (2013) and Fama and French's (2015) studies, our results also find that microcap is the problem. Three-factor regressions estimates for MC_HC cross-portfolios show that a three-factor model is unable to capture all a values equal to 0. Three portfolios out of 25 portfolios are found statistically significant (a values where t(a) is more than 1.96). The microcap stocks combination of small size and low compensation is well captured by the model. Panel E represents the R2 values of the FFTF regressions. The average value of R2 for MC_PB, MC_ROE, MC_TA & MC_HC is 82.7, 81.1, 81.7 and 79 per cent, respectively.\nModel performance test (GRS test): The study uses GRS test for model performance test designed by Gibbons et al. (1989). All three-factor models are rejected by the GRS test and it is consistent with the recent global findings of Fama and French (2012, 2015) and do not support Connor and Sehgal's (2003) findings in the Indian context. Similarly Fama-French (2015) five factor models are rejected by the GRS test except for MC_HC cross portfolio, and then except MC_PB cross-portfolios, all other six-factor models are accepted by GRS test. Fama and French's (2015) study argues that lower GRS F-Statistics signifies six-factor model a better model than five-factor model and it is consistent with our results (see Table VIII) and that confirms the superiority of six-factor model in all cases over all both the three- and five-factor models.\nConclusion: There are average returns patterns related to the size, value, profitability, investment and human capital in the Indian capital market. GRS test easily rejects both the three-factor model and five-factor model directed at capturing these average return patterns for size, value, profitability and investment. The study estimated that the both three- and five-factor models have the similar explanatory power and R2 value is nearly similar. In Indian context, study results find LMH factor as an important factor unlike the Fama and French's (2015) findings in the US context but CMA factor as the redundant factor. In case of parsimony issue, our study results suggest that investment (CMA) is a redundant factor and average returns due to CMA is fully captured on exposure to the Rm-Rf, SMB, LMH and RMW factors. The investors whose sole interest is to capture the abnormal returns (regression intercepts/a value), study results suggest that a four-factor model without CMA factor yields similar results as of the five-factor model. Five-factor model will be of interest for the investors who have a choice to tilt towards size, value, profitability and investment premiums; the study suggests using the CMAO factor as a replacement for the CMA factor in the five-factor model. Moreover, the study result suggests the failure of the five-factor model over three-factor model in capturing all portfolios a values equal to 0 in the Indian context as similar to the studies of Hou et al. (2014), Clarke (2016) and Chiah et al. (2016) in the international context. Further study also supports the findings of Griffin (2002) and Fama and French (2012, 2015, 2016) that Fama-French factor models are not global models.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to focus on one of the major emerging Asian economies - India - to examine the role of human capital in asset prices.\nDesign/methodology/approach: The analysis uses various statistical techniques (e.g. multifactor regression model, 3D graphs, GRS test and residual graphs) to test the role of human capital in asset prices.\nFindings: A six-factor model designed for capturing the size, value, profitability, investment and human capital patterns in average portfolio returns performs better than both Fama-French's (1993) three- and Fama-French's (2015) five-factor model. The main problem of six-factor model is its failure in capturing the average returns on \"microcap with low-value stocks that are highly profitable invests aggressively for asset growth but invests much lesser for human growth\" and \"microcap with unprofitable stocks whose returns behave like those of low-value firms with conservative investment\". The study finds the investment factor (CMA) of Fama-French's (2015) five-factor model as the redundant factor for describing the portfolio average returns in the study sample.\nResearch limitations/implications: The paper argues that human capital also plays a role in predicting returns. This has significant public policy content.\nOriginality/value: The present study is novel for several reasons: first, it includes six-factor model descriptions; second, no comprehensive asset pricing study is done with human capital in Asian emerging markets, especially in India. Perhaps, this is the first study to examine whether portfolio returns are affected by the human capital in the Indian context. Third, the study period and methodology used are completely different from the previous studies.\n",
"role": "assistant"
}
] |
Design thinking: a working strategy for the third sector
|
[
"India",
"Innovation",
"Design thinking",
"Desert",
"Marwar",
"Non-governmental organisations (NGOs)",
"Systematic inventive thinking (SIT)",
"Water scarcity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Non-governmental organisations (NGOs) are non-profits that belong neither to the government (public) nor the private (for-profit) sectors. They represent the civil society and constitute the \"third sector\". In emerging economies, NGOs have become the key players providing solutions to many complex developmental challenges and societal problems. The reasons for the emergence and proliferation of NGOs include growing fiscal deficits and lack of viable public-funding models, availability of funding from national and international donors and an expanding neo-liberal ideology among international donors that prefer non-state service providers (Hershey, 2013; Schwenger, Straub and Borzillo, 2014).\n1. Phases of design thinking: The DT process comprises three broad action phases: insight, ideation and impact (King, 2018) (Figure 1).\n2. Assessing innovations: the desirability, feasibility, viability (DFV) framework: An ideal innovation is one that meets three criteria: desirability, feasibility and viability. Desirability refers to providing solutions that customers really need; a feasible solution is one that builds on the strengths of the organisation's current operational capabilities; and a viable solution is one which comes up with a sustainable business model (Figure 2).\n3. The Jal Bhagirathi Foundation: The JBF was incorporated in 2002 to address the water crisis of the Marwar region, primarily the Thar Desert area. It was driven by a vision of water security, sustained by responsive governance and inclusive growth, leading to sustainable development. Marwar, with a population density of 83 persons per km2, is the most densely populated arid zone in the world. The seven districts constituting the Marwar region experience extreme climatic conditions, with temperatures ranging from 0 to 50degC and an average rainfall of 200 mm annually. Estimates place the chances of drought each year at 47 per cent, and the region had already witnessed six droughts from 2010 to 2018. The annual water table has been receding by an average of 1-2 m per year, but up to 2-5 m in some pockets. The total dissolved solids (TDS) of groundwater in the region is approximately 10,000 parts per million, making it unfit for consumption (Table II).\n4. Stage 1: insight: 4.1 Immersing, observing and empathising with the user\n5. Stage 2: ideation: JBF undertook ideation to generate multiple ideas and considered the conventional methods of rainwater harvesting, including building lakes and ponds. However, the problem was not one of merely eliminating water scarcity, but rather doing it in a manner that freed women from drudgery involved in activities associated with drawing and carrying water for the family members and animals, which prevented them from enjoying leisure or going to schools. Hence, DT principles were used during the ideation stage.\n6. Stage 3: impact: prototype and testing: The newer permanent tankas, which were built one per home/hut, went through multiple iterations. The prototypes incorporated multiple features. These were tested and the final tankas had the following features:\n7. Discussion: The use of DT in for-profit organisations has been well recognised. These include the contribution of DT to increased market share, development of new markets, enhanced competitiveness, increased employment, increased turnover, increased profit, development of new products/services and intellectual property (Best, 2011). Non-profit organisations, operating as service providers to meet societal needs and faced by unstructured, human-centric problems, are in a better position to take advantage of a \"design thinking\" approach rather than a conventional \"business\" approach to deliver better organisational outcomes (Table III).\n8. Summary: This study highlights the use of DT as a strategic tool by an NGO, Jal Bhagirathi Foundation, in the world's most water-distressed and densely populated arid zone in the world, the Thar Desert. It describes how, through insight, ideation and impact, the NGO was able to better tackle human-centric, unstructured and complex problems in desert communities. Such problems, though originating as water-scarcity problems, had multiple ramifications in spaces pertaining to sanitation, health, literacy - especially for the girl child - poverty and migration. By involving the users, the NGO was able to achieve significantly superior organisational outcomes. It emerges then that NGOs, especially in emerging markets and developing economies, should use the DT approach to solving the complex, developmental challenges faced by these countries.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to understand how non-governmental organisations (NGOs) can use design thinking (DT) as a strategic tool to improve organisational and societal outcomes.\nDesign/methodology/approach: The paper applies a case study design. Building on in-depth interviews with senior management, beneficiaries (villagers), the village headman and children in Karaliya village (Rajasthan), as also from secondary sources, the paper presents a model of using DT for NGOs.\nFindings: This paper presents evidence of how Jal Bhagirathi Foundation, an Indian NGO working in the most water-distressed and densely populated arid zone of the Thar Desert, used a user-centric, DT approach to solve the water-scarcity problem in villages in the Marwar region. In doing so, several interconnected societal problems were also addressed, including those of sanitation, education - especially of the girl child - poverty and migration.\nPractical implications: DT has been increasingly used by \"for-profit\" business organisations to derive competitive advantage. NGOs have lagged in the use of DT as a strategic tool to drive enhanced organisational and societal outcomes. Such NGOs can strategise to adopt a DT approach.\nOriginality/value: A literature search revealed that while the term \"NGO\" threw up 36,571 results, \"NGOs and strategic management\" had only 363 results, and \"the use of design thinking in NGOs\" had 0 results. This paper contributes to the limited research in the field of strategic management in the NGO space by looking at DT as an important strategic tool for NGOs.\n",
"role": "assistant"
}
] |
Predicting hospitality firm failure: mixed sample modelling
|
[
"Default risk",
"Support vector machines",
"Logistic regression",
"Ensemble of ensembles",
"Bankruptcy prediction of hospitality firms",
"Expert systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: A firm's survival is one of the most important interests to its investors, managers and employees because its failure will cause its investors to lose profits and its managers and employees to lose their jobs. However, firm failure is not occasionally happened. For example, in 2014, Chinese courts filed approximately 2,000 bankruptcy cases and nearly 10,000 firms went bankrupt in Japan. Unlike industries such as manufacturing, mining or energy, the focus of hospitality is not on materials or products but on human beings. Hospitality is the act of being hospitable in the reception and entertainment of guests or visitors. The influence of a hospitality firm's failure is greater than that of firms in other industries because the direct service subject is human beings. Besides the losses caused to a firm's investors, managers and employees, a hospitality firm's failure could cause service suspension for its customers or even a negative social event if many people are unable to return to their homes due to the failure of a tourism firm. For local economies that rely on tourism for jobs, hospitality firm failures will have negative impacts on both hospitality businesses and communities. Thus, helping hospitality firms to avoid failures is of importance to the investors, managers, employees, customers and governments that rely on hospitality firms for jobs.\nLiterature review of failure prediction: Firm failure prediction outside the hospitality industry\nMixed sample modelling for forecasting the failure possibility of hospitality firms: Workflow of mixed sample modelling\nHospitality samples and variables: Hospitality samples for forecasting the failure possibility of hospitality firms\nEmpirical design, results and discussions: Empirical design\nDiscussion: The perspective of the outlier parameter of th2?\nConclusions and limitations: Conclusions\n",
"role": "user"
},
{
"content": "Purpose: Available information for evaluating the possibility of hospitality firm failure in emerging countries is often deficient. Oversampling can compensate for this but can also yield mixed samples, which limit prediction models' effectiveness. This research aims to provide a feasible approach to handle possible mixed information caused by oversampling.\nDesign/methodology/approach: This paper uses mixed sample modelling (MSM) when evaluating the possibility of firm failure on enlarged hospitality firms. The mixed sample is filtered out with a mixed sample index through control of the noisy parameter and outliner parameter and meta-models are used to build MSM models for hospitality firm failure prediction, with performances compared to traditional models.\nFindings: The proposed models are helpful in predicting hospitality firm failure in the mixed information situation caused by oversampling, whereas MSM significantly improves the performance of traditional models. Meanwhile, only partial mixed hospitality samples matter in predicting firm failure in both rich- and poor-information situations.\nPractical implications: This research is helpful for managers, investors, employees and customers to reduce their hospitality-related risk in the emerging Chinese market. The two-dimensional sample collection strategies, three-step prediction process and five MSM modelling principles are helpful for practice of hospitality firm failure prediction.\nOriginality/value: This research provides a means of processing mixed hospitality firm samples through the early definition and proposal of MSM, which addresses the ranking information within samples in deficient information environments and improves forecasting accuracy of traditional models. Moreover, it provides empirical evidence for the validation of sample selection and sample pairing strategy in evaluating the possibility of hospitality firm failure.\n",
"role": "assistant"
}
] |
Putting the "brand" back into store brands: an exploratory examination of store brands and brand personality
|
[
"Generics",
"Brands",
"Shops"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\nIntroduction: A number of researchers have examined the success and failure of store brands (or private labels). However, one area of store brands that has been largely overlooked by researchers is their brand personality (BP). When consumers view a brand as having human characteristics, the brand is said to have a personality. For instance, brands such as Harley Davidson (Ruggedness), Nike (Excitement), Hallmark (Sincerity), Wall Street Journal (Competence) and Tiffany (Sophistication) have all been found to have strong personalities. Yet, what kind of personality traits would a store brand have?According to Nishikawa and Perrin (2005), store brands in North America now account for 16 percent of all items sold in US supermarkets, drug chains, and mass merchandisers. While this market share has been improving for the last several years, US retailers still have a long way to go to reach the healthy numbers gained by some of their European counterparts. For example, Switzerland, Germany, and Great Britain have store brand market shares of 45 percent, 30 percent, and 28 percent respectively. Such disparity in market share, for the most part, probably stems from the fact that US retailers have not spent the time or money to properly develop the \"brand\" of their store brands.Thus, the purpose of this research was to examine the branding of a store brand. Specifically, we looked to see if store brands have a brand personality and how this personality differs from their national brand counterparts. In addition, we explored whether or not the BP of store brand can be changed by having consumers \"experience\" the brand. Lastly, we investigated the relationship between a store brand's perceived personality and its perceived quality ratings.A survey was administered to 139 MBA students at a university in a major city in the Southwest. Respondents were randomly assigned to one of four survey conditions. In each condition, respondents were asked to rate the BP and quality of a brand of cookie (store or national) and brand of cola (store or national). In addition, roughly half of the respondents were only given the brand name (of cookie and cola) on a sheet of paper and then asked to evaluate the BP and quality. The other half of the respondents were able to experience the brand (i.e. see, hold, smell, taste, etc.) before being asked to evaluate the BP and quality. Thus, a 2x2x2 experimental design was employed (two product categoriesxtwo brands in each categoryxtwo experimental conditions).The results revealed that developing a store brand's BP is important. For one, the national brands had higher BP than the store brands in each category. Second, a brand's personality was found to be correlated with a brand's perceived quality. Thus, store brand managers would be wise to further explore how they can create stronger personalities for their brands. In doing so, they might be able to close the gap between their store brand's image and that of the national brands in which they compete.\nLiterature review: Despite the relatively low market shares as noted in the AC Nielsen reports, retailers continue to develop and stock brands under their own label because profit margins are higher than those for national brands (Hoch and Banerji, 1993). In addition, according to Hoch (1996, p. 89), \"Retailers get quick, cheap sales feedback that allows them to mimic what works for national brands while avoiding costly mistakes\". Given this first-hand marketing research and the profit potential of store brands, retailers have worked recently to improve the quality of their store brands.However, as Sethuraman and Cole (1999) found, a national brand and store brand can have similar objective quality ratings (as measured in laboratories with food technicians), but still have different subjective quality ratings (as measured by consumer perceptions). Similarly, Richardson et al. (1994) conducted a study to examine consumer's perceptions of store brands versus national brands. Their procedure included switching national brands with store brands; national brand contents were put into a private label package with a private label price and private label brand contents were put into a national brand package with a national brand price. The results revealed that consumers preferred whatever was packaged as a national brand with a national brand price. And, Burton et al. (1998) developed a scale to measure consumers' attitudes toward private label brands. Using a multi-item measure, the authors reached several interesting findings. They found that consumer attitudes toward private label brands were negatively correlated with brand loyalty, price-quality perceptions and impulsiveness.More recently, a Special Report on Branding and Private Labels (Pollack Associates, 2002) noted that retailers are increasing the number of private label offerings because of the current economic conditions. With store brands being usually cheaper than national brands, the report indicated that consumers are looking to save some money at the grocery store to offset a tepid economy and soaring gas prices.Altogether, more research on how store brands can compete against national brands is needed. Previously, the store brand research that has been conducted has primarily focused on the following aspects:* How national brands should react to the surge of store brands (Hoch, 1996; Quelch and Harding, 1996).* What the optimal price gap should be between national brands and store brands (Heath et al. 2000).* What factors are conducive for introduction of store brands (Raju et al., 1995).* What factors determine the variation in store brand market share among retailers (Dhar and Hoch, 1997).* The appeal factor of store brands (Dunne and Narasimhan, 1999).* What store brands can do to convince consumers of their high quality (De Wulf et al. 2005).However, there has been no prior research that has looked at how the personality of a store brand could be a driving factor in determining the success of that store brand.\nHypotheses: One way to create a strong brand identity is to develop the brand's personality. According to a number of researchers, brands can and do develop personalities or human characteristics (Aaker, 1997; Blackston, 1993; Plummer, 1984; Wysong et al., 2002). Examples of human personality traits that consumers assign to brands include honest, daring, outdoorsy, intelligent, upper-class, etc. (Aaker, 1997). In brand personality research, consumers are asked to think of the brand as a person. With this approach, marketers are able to develop a deeper understanding of the consumer decision-making process with regard to that brand. Given that national brands spend more money to develop their brand's personalities, it is hypothesized:H1. There will be statistically significant (p<0.05) differences between store brands and national brands with regard to their brand personality ratings (as defined by Aaker's (1997) Brand Personality Scale).Aaker (1996) was one of the first researchers to advocate that a brand's product-based attributes could influence consumer perceptions of the brand. According to Aaker (1996), a brand's ingredients, color and packaging can be very important to developing BP. Consequently, if consumers are allowed to try or experience a brand, they should have a better understanding of the brand and its personality. In addition, consumers who experience a brand will be able to judge first-hand the brand's quality. Thus, it is hypothesized:H2. Brand personality ratings (as defined by Aaker's (1997) Brand Personality Scale) of a brand will be significantly higher (p<0.05) when respondents are able to experience the brand (versus just being exposed to the brand name).H3. Quality ratings (as defined by Sethuraman and Cole (1999)) of a brand will be significantly higher (p<0.05) when respondents are able to experience the brand (versus just being exposed to the brand name).Given the Richardson et al. (1994) study cited previously, it would seem that consumers are more likely to rate a brand with a stronger brand identity as having better quality. Moreover, as Sethuraman and Cole (1999) discovered, typically a national brand rates higher in quality perceptions and thus is able to charge premium pricing as compared to store brands. Hence, brand personality and brand quality should be related, as well. Thus, it is hypothesized:H4. Quality ratings (as defined by Sethuraman and Cole (1999)) and brand personality ratings (as defined by Aaker's (1997) Brand Personality Scale) of a brand will be positively correlated.\nMethodology: Products\nResults: The final sample included 139 respondents (Group 1=32, Group 2=31, Group 3=43, and Group 4=33). The sample was roughly equal with regard to gender. Of the respondents, 55 percent were male and 45 percent were female. According to the results, the highest rated brand personality trait for a cookie was for the store brand (Safeway Select) in the experiential condition and the highest rated brand personality trait for a cola was for the national brand (Coca-Cola) in the experiential condition. Table II shows the highest rated brand personality facet by group.H1. Brand personality differences between store brands and national brands\nImplications, limitations and future research: Implications\n",
"role": "user"
},
{
"content": "Purpose: - To explore how store brand personality might play a role in consumer perceptions towards store brands and how such personalities might vary when consumers are allowed to experience the product.\nDesign/methodology/approach: - Two different store brand products from high selling categories were compared with their leading national brands. A total of 139 respondents took the survey. Respondents were divided into four groups, each group representing a different experimental condition. Respondents rated the personality attributes on a seven point Likert scale.\nFindings: - Results show that national brands have stronger brand personality traits when compared to store brands; however, such differences diminished when consumers were allowed to taste and experience the products.\nPractical implications: - Retailers would be better off to let consumers experience and taste the products to drive consumer perceptions of store brands. Retailers should also focus on those categories where the differential in personality traits between store brands and national brands are small and where the profit margins are higher to drive the store brand image.\nOriginality/value: - This paper presents a slightly different approach in understanding store brands and argues that store brands do have a personality and that such personality could be further influenced by allowing consumers to experience the product.\n",
"role": "assistant"
}
] |
The impact of flooding on marriage: evidence from Pakistan
|
[
"Marriage",
"Floods",
"Disasters",
"Difference-in-difference analysis"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Disasters shake human well-beings at individual, family and community level (Cohan and Cole, 2002). While the most immediate impacts of disasters get a lot of attention from media, governments and NGOs, the focus of the aftermath effects of the events mostly remains calm, perhaps due to the difficulty in quantifying long-term impacts (Anttila-Hughes and Hsiang, 2013). The issue of disasters' impacts on mental health has received much attention in research. But only until recently the impacts of hazards on life course transitions such as marriage, divorce and birth had not been studied. Perhaps due to the data constraint, research studies on the impact of floods on marriage have received a very limited attention. This study reduces the gap in knowledge in the literature by using a rich and representative survey-based data set from Pakistan and utilizing the 2010 Pakistani flood as a natural experiment.\nTheoretical background: The theoretical base for this study comes from sociological and psychological (Cohan and Cole, 2002) as well as economic literature, from which the prediction of the disasters' impacts on marriage can be explained. From the known literature, these can be grouped into two categories. The first group includes theories which predict that stressful events may bring favorable impacts on marriage, such as causing an increase of marriage rate. This group includes two evolutionary perspective theories; attachment theory (Bowlby, 1969) and terror management theory (TMT) (Solomon et al., 1991). Second group comprises research on economic circumstances which predict the contrary results. The attachment theory states that \"marital solidarity\" may occur after a disaster. When a disaster happens, then people would like to be near to their closed or loved ones and in case of severe dangers, both adults and children run toward their proximate relationships for security and relaxation. One of the important assumptions of the theory is that the attachment system is active over the entire life span which can also explain the adults' behavior for seeking association and support in time of distress (Hazan and Shaver, 1994). Furthermore, Cohan et al. (2009) also suggest that marital bonds become more strong in case of extreme events which eventually may cause an increase of marriage after a disaster because the role of being a romantic partner rather than stay attached to parents or guardian would change after a disaster and dating couple become life partners and get and maintain more intimacy in partnership (Prati and Pietrantoni, 2014). Based on this theory, marriage rate in flooded and to some extent non-flooded districts in Pakistan due to the flood should increase.\nData sources and descriptive statistics: Marriage data\nEmpirical analysis: Modeling framework\nResults and discussion: Main results\nRobustness checks and further extensions: Placebo experiment\nConclusion: The present study examines the aftermath impact of a natural hazard on one of the life course transitions, marriage, in developing countries with a special reference to the July 2010 Pakistani flood. Based on the terror management and attachment theories, it is predicted that marriages in flooded districts after the flood would increase while research on economic circumstances provides a base for negative association between hazard and marriage.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the effect of the flood on marriages in flooded households compared to marriages in unaffected households by utilizing the 2010 Pakistani flood as a type of natural experiment.\nDesign/methodology/approach: A difference-in-difference approach is used to estimate the effect of the flood on marriages in 62 flooded districts compared with those in 53 non-flooded districts by utilizing the six waves of the household level surveys data from the Pakistan Social and Living Standards Measurement, 2004-2005 to 2014-2015.\nFindings: Results show that the flood decreased marriages; by 17 marriages per 1,000 individuals aged 15-50 years in flooded districts during the flood year and the effect disappeared after the flood year. The negative impact of the flood on rural marriages is significantly higher and robust.\nSocial implications: The flood seemingly discouraged individuals in flooded districts to be engaged in long term relationship mainly due to the flood related economic and financial losses. In order to acquire and maintain individual overall well-being, sexual health in vital to maintain mental and physical health, so policy makers/humanitarian aid-providers should assist the affected adults financially or by arranging their marriages at least during the flood year. The study also suggests that the delay of marriages means the accumulation of human capital in the form of school attainment of male marriages, so younger adult should be discouraged marrying at early age.\nOriginality/value: This study contributes to the literature in the following ways: first, the study empirically investigates the impact of flood - both immediate and long term - on marriage rates by using a natural experiment. Second, it examines the relationship based on geographic location and gender. Third, it investigates the impact of natural hazards on child marriage.\n",
"role": "assistant"
}
] |
How communication climate and organizational identification impact change
|
[
"Change management",
"Organizational identification",
"Internal communications",
"Communication climate"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Successful organizational change requires effective internal communication. As Kotter (1995) pointed out \"employees will not make sacrifices, even if they are unhappy with the status quo, unless they believe that useful change is possible\" (p. 63). Internal communication can build employee support for organizational change. Barrett (2002) identified five primary goals for internal communication in times of organization change: \"(1) ensure clear and consistent messages to educate employees in the company vision, strategic goals, and what the change means to them; (2) motivate employee support for the company's new direction; (3) encourage higher performance and discretionary effort; (4) limit misunderstandings and rumors that may damage productivity; (5) align employees behind the company's strategic and overall performance improvement goals\" (p. 220).\nLiterature review: Organizational change has been defined as a planned process that allows organizations \"to adjust and behave differently to accommodate new contingencies, technologies, values, processes and personnel\" (Seeger et al., 2005, p. 80). Organizations engage in change to pursue new opportunities to strive for competitiveness, productivity, and growth (Choi and Ruona, 2011; Kotter, 1996). The forms of changes vary from leadership adjustment, cultural change, technology change, to downsizing and layoffs (Luo and Jiang, 2014).\nMethod: This study recruited 1,034 employees in June 2018 with the assistance of Dynata (formerly known as Survey Sampling International), a global survey service company with 1.5m-member research panel. Applying the stratified and quota random sampling methods, we obtained a representative sample of employees working in a variety of industries in the USA and with comparable age groups, gender, education, and income, in accordance with the US Census. To qualify for this study, a screening question required that participants indicate that their company is currently undergoing a company-wide change or has gone through such change in the past two years and they have experienced the change while employed at their company. The survey participants were then asked to identify the type of change they had recently experienced using a checklist of eight common types of organizational change derived from the literature as well as the option to enter another type of change as \"other.\" The most common type of organizational change they reported experiencing was reorganization (16 percent), followed by changes in leadership reported by 14 percent, and downsizing, mergers & acquisitions, new product lines and changes in technology, each reported by 12 percent. The average age was 41 years, and 46.7 percent were females. Approximately half had at least a bachelor's degree. Regarding employment history, 31.1 percent of participants have been with their current organizations for 2-5 years, and 22.9 percent for 6-10 years. The largest segment of the sample (38.1 percent) reported working in non-management position, 31.1 percent were in middle-level management, 20.9 percent were in lower-level management, and 9.9 percent were in top management.\nResults: The hypothesized model was tested utilizing two-step structural equation modeling analysis with AMOS 25.0 software. In the structural model, communication climate was identified as a latent variable with two dimensions of openness in communication and participation in decision making. Organizational identification and employee affective commitment to change were specified as unidimensional observed variables. Employee behavioral support for change was identified as a latent variable with two dimensions of cooperation and championship. The test of the confirmatory factor analysis (CFA) model revealed satisfactory fit to the data: kh2(5)=16.21, p < 0.01, kh2/df=3.24, RMSEA=0.04 (90% confidence interval: 0.02-0.07), SRMR=0.01, TLI=0.99, and CFI=0.99. The standardized factor loadings ranged from 0.84 to 0.94, suggesting a good construct validity of the measurement model (Figure 1).\nDiscussion and conclusions: A fundamental task for modern organizations is to manage changes to adapt to the evolving environment and remain competitive in the market. This is especially true in light of the ever-increasing frequency and rate of changes (Venus et al., 2018). Among various factors that impact change management and implementation, communication arguably plays a central role (Elving, 2005; Men and Bowen, 2017). The current study proposed and tested a conceptual model that links communication climate characterized by openness and participation, organizational identification, and employee reaction to change (i.e. commitment to change, behavioral support for change). Two major findings were revealed: First, an open and participative communication climate directly contributes to employee affective commitment to change and behavioral support for change. Second, communication climate featured by openness and participation boosts employee identification with the organization, which, in turn, leads to positive employee reaction to change. In other words, organizational identification partially mediates the effect of communication climate on employee affective commitment to and behavioral support for change. Together, the findings support the role of communication climate in facilitating employee support for change and explained how the process works by showcasing the mediation effect of organizational identification.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine why and how an open and participative communication climate matters for employee organizational identification and their change-specific responses, specifically employees' attitudinal and behavioral reactions.\nDesign/methodology/approach: To test the hypothesized model, the authors conducted an online survey using a stratified and quota random sample of 1,034 US employees working in diverse industry sectors in June of 2018, with the assistance of a premier global provider of survey services, Survey Sampling International. To test the hypothesized model, structural equation modeling analysis was employed using AMOS 24.0 software.\nFindings: An open and participative communication climate directly contributes to employee affective commitment to change and behavioral support. Communication climate featured by openness and participation boosts employee identification with the organization, which leads to positive employee reaction to change. When employees identify with the organization, they tend to believe in the inherent value of the change and are more likely to support the change initiative in action through cooperation and championship.\nOriginality/value: Theoretically, the study contributes to the change management and communication literature by focusing on the role of communication climate in inducing employee reaction to organizational change. Practically, the study offers insights for change managers, internal communication professionals and organizational leaders. Organizational leaders need to be open, create a trusting atmosphere and actively involve employees in the decision-making process. Organizational leaders and communicators should also strive to boost employee identification with the organization, especially during change.\n",
"role": "assistant"
}
] |
Job characteristics affect shared leadership: The moderating effect of psychological safety and perceived self-efficacy
|
[
"India",
"Self-efficacy",
"Psychological safety",
"Job characteristics",
"Shared Leadership",
"Sociograms"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Organizations are increasing their reliance on team-based work due to its high potential to demonstrate quality results in less time. Teams bring diverse perspectives and experience to ensure richness in the workplace decision making and functioning (Torrente et al., 2012). The importance of structuring work around teams stems from emerging trends such as increase in global competition, emphasis on knowledge-work, innovation and flatter organizational structure. These trends inevitably make an external leader incapable to execute the leadership functions for his/her team (Lawler et al., 2002).\nLiterature review and hypotheses development: Job characteristics and shared leadership\nMethodology: Organizational context\nResults: Multivariate techniques were used to test the relationships proposed above with the help of SPSS 20 statistical software. Descriptive statistics were calculated using measures of mean and standard deviation. To test moderating effect of psychological safety and self-efficacy on shared leadership, hierarchical multiple regression analysis was conducted. The means and standard deviations have been reported in Table II. Overall, the means gave indications of a positive perception toward shared leadership in the organization by respondents. The correlation matrix provides evidence that all five job characteristics and psychological safety are correlated with shared leadership indicating that there may be significant main effect on shared leadership.\nDiscussion: This study adds to the extant literature on antecedent conditions of shared leadership by discovering job characteristics as significant determinants of shared leadership and psychological safety and self-efficacy as notable moderators. The results, in particular, indicate that autonomy, task significance, skill variety and feedback impact shared leadership. Autonomy has a significant positive relationship with shared leadership. As experienced in the organizations, employees who have independence to influence and govern their work are more responsible and accountable for the work. Skill variety positively influences shared leadership which suggests that if an employee performs a variety of tasks then he/she is more engaged and motivated. Furthermore, task significance also has a positive relationship with shared leadership, strengthening the fact that jobs with a substantial effect on others elevate the intrinsic motivation of employees and encourage them to participate in challenging tasks. In addition, feedback is also found to be a significant determinant of shared leadership. Employees receiving timely feedback are motivated to pursue their work by bridging the gaps. However, contrary to our expectations, task identity is not found to affect shared leadership among team members. This can be because of the organization chosen for this study, where each employee is held responsible for the task assigned instead of the entire assignment. Hence, employees do not identify themselves with the completion of entire work and the tasks for which they are not responsible. This discovery suggests that designing the jobs with enriched resources (high on core dimensions) positively facilitates shared leadership and may unwind other constructive benefits, supporting Hackman and Oldham's (1974) conceptual framework.\nConclusion: A new facet of leadership has come to light which is important today in modern organizations. Shared leadership in teams or organizations at large is instrumental in achieving quick decision making and performance. This study provides an advanced model of leadership that has contentedly made organizations to quit viewing leadership function through traditional lenses. The authors feel that this study encourages future researchers to explore association of more novel variables with shared leadership and proffers organizations a strong manifestation of positive outcomes.\n",
"role": "user"
},
{
"content": "Purpose: Shared leadership is characterized by exercising lateral leadership influence depending upon the task and the adroitness required for the task. The purpose of this paper is to examine the effects of job characteristics on shared leadership and explore the moderating effects of perceived self-efficacy, and psychological safety on the relationship between job characteristics and shared leadership.\nDesign/methodology/approach: The authors applied hierarchical regression and social network analysis using a sample of 23 teams consisting 219 employees from an Indian-based multinational BPO.\nFindings: Results indicate that skill variety, task significance, autonomy and feedback are significant precursors for shared leadership, and psychological safety and perceived self-efficacy act as moderators.\nResearch limitations/implications: The study extends the literature associated with job characteristics in relation to shared leadership. Managers can encourage shared leadership using job design and by creating a psychologically safe environment to enhance performance of the team. The cross-sectional nature of the study and data from a single organization question the generalizability of the results.\nOriginality/value: The authors provide an initial understanding of the impact of job characteristics in a team for the development of shared leadership, which has not been studied so far. Also, psychological safety and self-efficacy of individuals in teams have not been explored in the past, moderating the job characteristics and shared leadership relationship.\n",
"role": "assistant"
}
] |
Examining customer-oriented positive deviance intentions of retail employees
|
[
"Organizational performance",
"Hope",
"Goal attainment",
"Positive deviance intentions",
"Retail employees"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: What inspires some employees to break the rules to help a customer? For example, a shop assistant who provides a refund without a docket, or a flight attendant who arranges a free meal for a passenger who is unable to pay. Stepping outside defined workplace rules - albeit to benefit a customer - can be a risk for retail and service employees. Studies of employee workplace deviance intentions generally highlight the negative impact of employees' deviant behaviour on organisational outcomes (Coyne and Bartram, 2000; Dunlop and Lee, 2004; Ferris et al., 2012; Mishra and Prasad, 2006; Swimberghe et al., 2014; Jaakson et al., 2017). However, it is also possible that a competent employee may deviate from their company's procedures, norms and policies with the intention to provide support to their customer (Mertens et al., 2016; Bowman, 2013). In line with this view, customer-oriented positive deviance intentions is defined as the actions of an employee who voluntarily departs from norms of a referent group to achieve favourable outcomes for the organisation and themselves (Jung and Yoo, 2019; Kim and Choi, 2018; Leo and Russell-Bennett, 2012). For example, a frontline service employee may deviate from the organisation's policy in an attempt to build relationships with their customers, which may boost sales and subsequently improve the organisation's performance. Alternatively, they may overlook standard procedures in order to attain positive feedback about their performance.\n2. Theory development: 2.1 Employee hope\n3. Method: To test the conceptual model, we measured the constructs of interests via self-completion, an online survey administered to a panel of respondents (n=536) sourced from a national commercial provider. The study adopted a purposive sampling strategy, not only because of time and cost constraints, but also because the purposive selection of employees as respondents can answer the basic research question. This study used Roscoe's (1975) and Cochran's (2007) criteria for determining a sample size. Roscoe (1975) suggests that a sample size of ten or more times the number of variables in the theoretical framework is reasonably good for any research in social sciences. In line with Roscoe (1975), Cochran (2007) states that for a small population (i.e. less than 10,000), a sample set of 10-30 per cent of that population is suitable. His recommendations for a sample size of a larger group (i.e. over 150,000) are as low as 1 per cent. In current study, there are 15 items for four different constructs. Thus, a sample size of 200-250 is appropriate to run a multivariate statistical analysis. However, we were able to collect data from 270 respondents. An online survey methodology was chosen due to its ease and availability of access to a representative sample of participants. Online surveys overall are very valuable when describing characteristics of a large population (Babbie, 2015). Further, with the rise of smartphone technology, participants can be reached at any location at any time, which increases convenience for participant (Evans and Mathur, 2018). It is also possible to remove non-response error in the survey by coding online surveys in a way that does not allow access to following sections without fully completing all previous questions. Additionally, for data analysis purposes, an online survey method allows for the data to be stored and analysed without input of data to a separate programme, thus reducing the possibility of errors when data are transferred from one format to another, such as from paper to statistical programme (Babbie, 2015; Evans and Mathur, 2018).\n4. Analysis and results: 4.1 Confirmatory factor analysis (CFA)\n5. Discussion: This research has contributed to the scholarship on employee-organisation relationship-building strategies by considering hope as a motivational influence on employees' positive deviance intentions, which in turn impacts employees' goal attainment, improving their perceptions of organisational performance. This study differs from previous research (Ferris et al., 2012; Huangand Lin, 2019; Mishra and Prasad, 2006; Swimberghe et al., 2014; Jaakson et al., 2017) in that it considers employees' customer-oriented deviance intentions as positive, as long as an employee's deviance is to benefit the customer and organisation. We found that employees who deviate from the retail organisation's rules with the aim of providing benefits to customers, also expect to bring about benefits to the organisation, presumably as a result of customer satisfaction.\n6. Managerial implications: This study highlights a psychological construct of hope, and develops and tests a conceptual model which is different from general extrinsic or intrinsic motivation or desire-based, goal-directed behavioural models. For theorists, the employee hope model may provide a better understanding of a psychological mechanism, namely, hope, that can explain why employees' customer-oriented positive deviance intentions are productive (or devastating) for customers, employees themselves and their organisation. For practitioners the employee hope model offers a better understanding of positive outcomes of employee deviance, suggesting that retail managers should invest resources to build strong employee-organisation relationships.\n7. Future research and limitations: As with any study, this research contains limitations that serve as potential avenues for further inquiry. The cross-sectional nature and single-level data collection may limit the generalisability of this research and create issues other than common source variance. Response bias is a widely discussed issue in behavioural research where self-reported data are used, occurring when respondents offer self-assessed measures of some phenomenon (Rosenman et al., 2011). In light of our reliance on self-reported front line employee data, to maximise measurement validity, future research may consider combining the survey approach with other techniques, such as assurances of confidentiality, indirect questioning, face-saving questioning or the bogus pipeline (BPL) approach (McDonald, 2008; Nancarrow et al., 2001). A future extension of this work may be to focus on both \"trait\" and \"state\" hope, or multi-level studies on frontline employees, managers and organisations to ascertain differences. This study has not considered temporal effects of hope on employee productivity and performance and, consequently, employee choices and preferences. Another limitation of this study is that it excludes variables that might be perceived as close to hope, such as desire, expectation and optimism. A research design that establishes a control for these variables may bring some different results as compared to the current study's results. Further, an experimental approach could manipulate the levels of hope (high hope vs low hope) to clarify its impact on deviance intentions and employee goal attainment processes.\n",
"role": "user"
},
{
"content": "Purpose: Occasionally, retail employees \"break the rules\" in order to help customers. Currently, there is little research on the mechanisms by which a sales assistants' positive deviance intentions help them attain specific personal and organisational goals. The purpose of this paper is to examine one mechanism, hope, which develops employees' deviance intentions to provide benefits to the customer, themselves and the organisation.\nDesign/methodology/approach: A survey captured responses from 270 frontline employees from the retail and services sector. AMOS 23 was used to conduct measurement, path and mediation analyses.\nFindings: This study highlights the role of employee hope in developing employees' positive deviance intentions, and improving perceptions of organisational performance. Results demonstrate that the direct positive impact of hope on positive deviance intention was significant. Furthermore, positive deviance intention was found to positively impact employee goal attainment and perceived organisational performance. The authors' employee hope model offers a better understanding of positive outcomes of employee deviance, suggesting that retail managers should invest resources to build strong employee-organisation relationships.\nOriginality/value: This is the first study to empirically demonstrate that employee hope can explain how customer-oriented positive deviance intentions help employee goal attainment and improve their perceptions of organisational performance.\n",
"role": "assistant"
}
] |
Parental risk perception of mycotoxins and risk reduction behaviour
|
[
"Child health",
"Germany",
"Consumer protection",
"Optimistic bias",
"Willingness to pay"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In the past, many surveys have been conducted on the public perception of food-related hazards (Sparks and Sheperd, 1994; Fife-Schaw and Rowe, 1996, 2000; McCarthy et al., 2006; Ohtsubo and Yamada, 2007; Cunha et al., 2010). A common methodology to measure public perception of risks is the psychometric paradigm developed by Slovic and his co-workers (Slovic et al., 1979, 1980; Fischhoff et al., 1978; Slovic, 1987). This psychometric paradigm was originally designed to analyse the public perception of technology hazards. It indicates that every hazard has a unique pattern of social and psychological characteristics which are related to the perception of risks. This risk-specific pattern of characteristics helps the understanding of lay people's responses to a hazard and provides a basis for effective risk communication (Slovic, 1987). Fife-Schaw and Rowe (1996, 2000) modified the psychometric paradigm and extended it to the field of food-related hazards. They created the Perceived Food Risk Index (PFRI), a scale to analyse the public perception of food risks (Fife-Schaw and Rowe 2000).\n2. Mycotoxins - occurrence, health risk and consumer protection: Mycotoxins are naturally occurring mould toxins produced by fungi that often infect the crop already in the field. The most frequently detected mycotoxin in cereals is deoxynivalenol (DON) (Gareis et al., 2003). In the European Union, mainly wheat and maize are contaminated by DON (Gareis et al., 2003). DON is very heat stable, tasteless and invisible; thus, a DON contamination cannot be identified by the consumers. Mycotoxins are noxious in small quantities; however, their health effects depend on the mycotoxin type. In the case of DON, immunosuppressive effects, resulting in increased susceptibility to infectious disease, have been demonstrated (Scientific Committee on Food, 1999).\n3. Objectives: The health risks formed by mycotoxins are still largely unknown among the general public. Also, deeper knowledge about the public perception of health risks due to mycotoxins is rare. In the past, only two surveys on the public risk perception of mycotoxins have been carried out in Germany (Freese, 2010; Raupach and Marggraf, 2011). Both studies were related to health risks for adults only.\n4. Material and methods: 4.1 Questionnaire\n5. Results: 5.1 PFRI related to mycotoxins and child health in Germany\n6. Discussion: The aim of this study was to analyse the parental risk perception of mycotoxins and to elicit parental aWTP for cereals containing only an amount of mycotoxins that health risks for children can be excluded. It was found that parental risk perception of mycotoxins related to child health is characterised by two dimensions, Dread and Control, which are similar to those found in other food-related studies (Fife-Schaw and Rowe, 1996, 2000; Cunha et al., 2010). This result might indicate that risk perception of food-related hazards does not depend on the target subject. However, in the present study only the risk due to mycotoxins has been considered. Hence, further research is needed to clarify if this finding is transferable to other food hazards particularly affecting child health.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyse the parental risk perception of mycotoxins (mould toxins) related to child health in Germany. It ascertains the parental risk reduction behaviour operationalised as parental additional willingness to pay (aWTP) for special child products. It investigates the interrelationships between parental risk perceptions and risk reduction behaviour.\nDesign/methodology/approach: Altogether, 771 questionnaires were sent to 17 day-care centres in Lower Saxony, Germany. A total of 238 questionnaires were completed. The Perceived Food Risk Index was used to identify the main dimensions of parental risk perception of mycotoxins. Open and closed questions measured aWTP for risk reduction and parental risk estimates.\nFindings: The respondents believed that mycotoxins posed a moderate health risk for children in Germany. However, parental risk assessment was found to be subjected to an optimistic bias. Nevertheless, the parents were willing to pay a premium to protect their children's health against mycotoxins. Parental risk perception of mycotoxins could be described by two dimensions named \"Dread\" and \"Control\". \"Dread\" and \"Control\" predicted aWTP for risk reduction as well as being involved in the genesis of optimistic bias.\nResearch limitations/implications: Information regarding mycotoxins provided in the questionnaire probably influenced parental responses. This limits the generalisation of the findings.\nOriginality/value: This research provides initial findings regarding parental risk perceptions of mycotoxins and its impact on risk reduction behaviour. The results are useful for the development of effective risk management and thus for the protection of child health.\n",
"role": "assistant"
}
] |
Landslide problems in Uttaranchal, India: issues and challenges
|
[
"Landslides",
"Natural disasters",
"India"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Landslide disaster in the state of Uttaranchal in India has a very long and old history. Bill Tillman, during his ascent of Nanda Devi in 1935 writes:A first report from the local headman that a mountain had fallen was ignored. But when Lt. Col. Pulford, Superintending Engineer visited the site, it was found that successive landslides had formed a 300 m high dam. He opined that nothing would happen until water topped the dam, which according to their estimate would take at least a year. Accordingly, the danger limit down the valley in the Birehi and Alaknanda rivers was marked, people were warned, suspension bridges were removed and the pilgrim traffic was diverted. Finally, on 25 August 1894 water began to trickle over the dam and at midnight it collapsed. The flood lasted until the morning of 26 August destroying part of the blocked landslide. In the flood only one person was killed but the town of Srinagar was damaged.Reports of landslide and sinking, at Tawaghat and its surrounding areas, have been reported in the past with noticeable amount of loss of lives and also with properties. A sizeable landslide occurred in this region of Uttaranchal in December 1962. During August 1977, a major landslide came roaring down the main stream of Khaila village and destroyed sheds and houses, killing 44 persons and 76 heads of cattle. About 150 acres of standing crops were washed away. In August, 1979 again a major gravitational slide of the steep mountain slopes between Rauntigad and Tawaghat region had occurred, seriously affecting Sinsa, Gachhila and Syankuri villages situated on the above slopes.There were two major landslide events in Okhimath area on 11-12 and 18-19 August 1998. These slide events were one of the most devastating days in the history of the region. The continued and heavy rains from 2 to 3 days before slide events had confined most of the people to their houses. When the first slide came on 11-12 most people moved to safer areas and some of them lost their property and lives. The heavy rainfall of 18-19 created fresh slide in the region and one of the most devastating slides occurred along the left bank of Madhmaheshwar river. This slide destroyed two villages (Bhenti and Paundar) completely and blocked the flow of Madhmaheshwar river for 12 hours between 2 am and 2 pm on 19 August 1998. It created havoc down stream. Fortunately lake was breached and water started flowing. A total of Rs 41 million worth of property, 101 human lives and 422 cattle loss were estimated due to this disaster. From the geomorphology of these northern regions, it is clear that most of the villages here are located on the debris of old landslides or receding glaciers. In Chamoli district, nearly one-third of the settled villages face the threat of landslides. Similarly, in Uttarkashi and Pithoragarh districts of Uttaranchal, more than one-third of the villages are listed as unsafe (Figures 1 and 2).\nCauses: Slope convert gravitational energy into kinetic/mechanical energy (k=mv2/2 g) and simple diastrophic structures into complex denudational features and all landslides, rock falls, slumping, etc. depend on this process of mass movement. It is clear that the slope instability is the result of disturbances in equilibrium between shear stress and shear strength governed and regulated by various exo-indogenetic forces. However, it should be noted that for slope instability only one factor cannot be held responsible, though the role of any single factor may be vital in specific place and time. In the case of fragile Himalayan environment the slope instability seems to be governed by various set of following causative factors.Earthquake\nSome major landslides in Uttaranchal: Landslide and mass movements are recurring phenomena in Himalayan region. The consequences in recent times have become more severe in terms of casualties and extensive damage to the roads, buildings, forests, plantation, and agriculture fields. In recent years, the intensive construction activity and destabilising forces of nature have combined to generate a class of problem, huge and complex, never encountered before. Implementation of number of hydro-electric schemes, large-scale construction of dams, roads, tunnels, buildings, towers, ropeways, tanks and other public utility works as well as indiscriminate mining and quarrying have brought most of instability problems such as never witnessed before.The Karmi landslide\nLandslide in Uttaranchal: due to construction activities: With the increasing infrastructure development in the Central Himalaya, an anthropogenic change has brought a threat to the environment. Recently it has been experienced that constructional activity is promoting the instability. This process continued till date. It was further devastating by growing demand for housing and by increasing population particularly in the urban areas. Ill conceived, mismanaged, unrestrained and unplanned construction activity, both legal and illegal activity has given rise to large-scale cutting of hill surface for the construction of roads and buildings. Road networking in the Himalayan region is highly developed. It has been calculated that 600 m3 of debris per km of road is produced annually by landslides and rockfalls, causing about 24 x 106 m3 of sediment to slide down the slope.Along the Dehradun-Tehri road, there are ten landslides in every kilometer. Each of which generates 500 m3 of debris. Most of the roads in Himalayan region have been unscientifically constructed. For example, Panar-Ghat road sections (Distt. Pithoragarh) is constructed parallel to North Almora Thrust. Similarly, Ratighat-Garampani road runs parallel to fault line. At 147 km on Haridwar-Badrinath road (Garhwal Himalaya) a multitier slide, having combination of surficial and deep-seated movement of fragmented rock, had occurred.Unplanned urbanization in hill areas has also given rise to a large-scale cutting of the hill slopes for construction. This process disturbs the critical equilibrium point and results into slope instability. For example, Nainital town, after the historical landslide of 1880, in the Sher-ka-Danda ridges overlooking Nainital that created the flats. Parts of the town, including a hotel, a shop, assembly rooms used by British civil servants and army officers and the rooms of their orderlies, were buried under the rubble. The main reason for the catastrophe, which left 151 persons dead or missing, as cited in official documents, was that the:... digging of fresh building sites on the Alma hill - which the Sher-ka-Danda ridge was a part - together with the cutting of the hillside for gardens, roads and tennis courts permitted the crust to become overcharged with water.It is also said that the \"destruction of the grassy covering of the slopes had greatly facilitated the introduction of rain water into the hill\". In April 1885, British government drafted a set of norms for the civil construction. Up to 1886 all the constructions were implemented and social attentions were paid towards the sensitive areas with the remark that tennis courts and platforms should be covered with six inch layer of well rammed clay or other material impervious to water and they should be well drained. However, after independence, increasing population pressure, expansion of construction activity for administrative purposes, tourist industry, etc. quashed all the norms of town planning, specially designed for this highly sensitive town of Central Himalaya. At present, the hill slopes around the lake have developed significant impression of strain caused by slope instability. China Peak area of the Nainital town is demarcated as landslide prone areas but the construction activities are still going on.\nRelief and rehabilitation (R&R) issues in landslide prone areas: The main issue behind R&R is related to the socio-economic background of the people in the area concerned.Rehabilitation or resettlement programme must aim at providing the living atmosphere for the people, in which they used to live before the calamity. What is emphasized in such programmes is provision of dwellings, land pieces and some monetary packages in the name of compensation, which often lacks the societal set-up at par with which the resettled and masses were living.Most of the resettlement programmes suffer a set back because, it is next to impossible to resettle a huge mass. Moreover, land, livestock and other economic assets, which are provided as compensation are, in most of the cases, far beyond the expectation limits.In many cases (in the cases cited) the affected population does not encourage the resettlement programme because they enjoy a regular tenure. It has become a habit of people that they claim for compensation after every event, and it may be quested, as a fallacy of administration system, they used to get it without much difficulty.People very often refuse to resettle, because they usually have deep bandages towards their roots of parental belonging, socio-cultural relations and affectionate feelings towards the immediate surroundings (land, forest, water, people).The public policy on landslide hazard management encompasses decision making on the issue: when at what cost of various mix of preventive and ameliorative of palliative measures be applied to mitigate the hazards of mass movement. Two aspects, regulatory and legislative, are involved. Since, damages consequent to occurrence of involved effect one and all in one and all in one way or another, there should be mandatory regulation governing the procedure of hazard management and for providing guidelines in the matter requiring fixing responsibilities for causing/escalating landslides, providing deterrents to activities detrimental to environmental balance and awarding costs of damage by the law courts.The landslide zones and areas identified in the landslide zonation maps (copies of which should be available in offices of all planning and executive bodies and in the district, block and Panchayat offices) be prohibited or avoided for developmental programmes that involve reshaping or modification of the topography, utilisation of natural resources and charge in the balance loads on the ground medical aid and rehabilitation of the uprooted/threatened population. Every district of Uttrakhand should evolve respective district relief programme, involving co-operation of all segment of the administration and service organizations.\nIssues emerging: * Rehabilitation is how long feasible on landslide hit/prone areas.* Construction strategies in landslide prone areas for rehabilitation.* Coverage of infrastructure facilities while developing the Resettlement plan.* Meeting people's aspiration in the Resettlement areas.* Financial allocation for R&R form the concerned authorities.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to describe how landslide and mass movements are recurring phenomena in the Himalayan region. The consequences in recent times have become more severe in terms of casualties and extensive damage to the roads, buildings, forests, plantation and agriculture fields.\nDesign/methodology/approach: - The study is based on field work and secondary source on information about various landslides in the Uttaranchal, India.\nFindings: - In recent years, the intensive construction activity and destabilizing forces of nature have combined to generate huge and complex problems, never encountered before. Implementation of number of hydro-electric schemes, large-scale construction of dams, roads, tunnels, buildings, towers, ropeways, tanks and other public utility works as well as indiscriminate mining and quarrying have brought most of instability problems such as never witnessed before.\nResearch limitations/implications: - Mitigation of an active landslide is a difficult option. It has been observed that despite of best technological options available the mitigation of an active landslide costs a lot which is, some times not economically viable.\nPractical implications: - Prevention is the best practice for landslide management which can be introduced through proper landuse planning.\nOriginality/value: - The proper and scientific management of landslide will certainly reduce the miseries of the community living in this region. However, for that we must be addressed certain issues.\n",
"role": "assistant"
}
] |
Study shows how cognitive style and personality of leaders determine what they are good at
|
[
"Identity work",
"Cognitive style",
"Big Five",
"Institutional work",
"Integrative work"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An Indian study has established links between a manager's cognitive style and personality, and the types of work they do best. The data from 101 managers in diverse industries found that an \"intuitive feeling\" cognitive style is suitable for identity work, whereas an \"intuitive thinking\" style works well for integrative work. Meanwhile, \"openness to experience\" and \"conscientiousness\" are correlated positively with identity work. \"Conscientiousness\" was also the most important trait for managers doing institutional work, but \"agreeableness\" had a negative effect on both institutional and identity work.\n",
"role": "user"
},
{
"content": "Purpose: The authors wanted to find out if they could establish a connection between the cognitive styles and also the personality traits of managers and the roles they were best at.\nDesign/methodology/approach: They tested a series of six hypotheses on 101 managers and senior managers in diverse leadership roles in India. The four major types of cognitive style were intuitive thinking (NT), sensing thinking (ST), intuitive feeling (NF) and sensing feeling (SF). Meanwhile, the Big Five personality traits were openness to experience (O), extraversion (E), consciousness (C), agreeableness (A), and neuroticism (N). Finally, the three categories of leadership roles were identity work, institutional work, and integrative work.\nFindings: The data from 101 managers in diverse industries found that an \"intuitive feeling\" cognitive style is suitable for identity work, whereas an \"intuitive thinking\" style works well for integrative work. Meanwhile, \"openness to experience\" and \"conscientiousness\" are correlated positively with identity work. \"Conscientiousness\" was also the most important trait for managers doing institutional work, but \"agreeableness\" had a negative effect on both institutional and identity work.\nOriginality/value: The authors said that no previous researchers had measured the impact of personality and cognitive style on work types. They said it was important for organizations to make sure they found the right roles for their managers depending on their individual characteristics.\n",
"role": "assistant"
}
] |
New economics
|
[
"Facilities",
"Best practice"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Urban FM: Urban FM - the integrated provision of public service accommodation and community support services - can make a significant contribution to the principle of putting people first, and meeting the objectives of best value. The private sector has already shown that it has the skills and capabilities of managing the capital and service challenges that are required. However, if the FM industry is to step up to this challenge, it will have to move away from traditional contracting models and towards models based on the public interest company. It will have to become skilful in brokering agreements between community service organisations. It will have to develop new models of governance that allow the citizen to play a full role in co-producing the services that are provided.Urban FM - the idea that community management can be wholly externalised to professional service providers, responsible for investment and management of the public infrastructure and its associated services. Urban FM is simply a logical extension of the need to reinvest in community facilities and systems, and provide a flexible \"platform\" in which agencies and the private sector can come together in new and innovative settings for the benefit of the community.Indeed, support for much of the underlying philosophy of Urban FM was provided in a recent study by the Public Management Foundation (PMF) and sponsored by the Gulbenkian Foundation. The study sought the views of a wide range of senior public service managers about some of the challenges of the best value agenda.A number of areas of concern were identified:*difficulties in creating partnerships between community service organisations;*access to capital;*the need for greater organizational autonomy;*difficulties in balancing accountability to both service users and the public through the democratic process; and*lack of capacity to attract high calibre managers.Potentially, Urban FM provides solutions to a number of these problems through the creation of \"arms length\" organisations with greater autonomy and access to capital. Amongst the advantages would be:*the introduction of the techniques of business management, in particular efficiency improvement to public services;*the introduction of market mechanisms and competition into public life; and*a greater service and customer orientation within public services.However, the PMF report also highlights concerns about the quality of service delivery by the private sector. Respondents felt that experience had shown that conflicts of interest between the commercial interest of private service providers and the public service interest could not be resolved easily. The report highlights the uneven nature of the risk transfer when public services are outsourced. The closure of private residential elder care homes following disputes over cost and quality of services or the withdrawal of contractors carrying out rail maintenance places immediate and urgent burdens on public agencies to maintain the continuity of service.Social enterprise.\n",
"role": "user"
},
{
"content": "Purpose: - To provide a concise viewpoint on Urban FM.\nDesign/methodology/approach: - An opinion piece based on current and recent research on service delivery.\nFindings: - Urban FM - the integrated provision of public service accommodation and community support services - can make a significant contribution to the principle of putting people first, and meeting the objectives of best value.\nOriginality/value: - The paper provides a valuable summary of the public management foundation report.\n",
"role": "assistant"
}
] |
Sustainability coursework: student perspectives and reflections on design thinking
|
[
"Design thinking",
"Creativity and innovation",
"Reflection",
"Sustainable design",
"Environmental engineering education"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Background and relevant literature: Solving today's complex environmental and societal engineering challenges will require the current generation of students to possess a creative skillset and mindset. Design thinking, a human-centered design approach, has recently gained traction as a creative approach for solving societal and environmental sustainability problems. In Design for the Environment, an undergraduate engineering course in sustainability, students have begun to apply the design-thinking process to sustainability-related engineering challenges over the past several semesters. Specifically, a design-thinking approach was introduced to the course in the fall of 2017 and was used again in the fall of 2018. Enhancing student creativity was a primary driver and reason for incorporating the design-thinking process as an underpinning of the course. This paper describes a case study of the associated changes in students' creativity and anticipated future sustainability practices with this course modification. The modification was motivated and enabled by the instructor's experience at Stanford's d. school during the summer of 2017. In addition, feedback from the fall 2016 semester suggested the use of more hands-on innovation and design activities, including in-class use of the school's makerspace. The resulting course re-designs involved activities during class that enabled students to gain hands-on experience with the five stages of the design-thinking process. They ultimately applied this process to a semester-long sustainability design challenge, which was assessed for creativity. By applying a case study approach with multiple forms of assessment, the following research questions of interest were investigated:\n2. Methods: 2.1 Course context\n3. Results: 3.1 Survey and focus group results\n4. Discussion: Design for the Environment, a civil and environmental engineering course that is part of the sustainability course offerings in the department, was significantly modified so students could apply the design-thinking process to sustainability design challenges. Recent literature has called for the use of design thinking in addressing today's complex societal and environmental sustainability issues; thus, this research is relevant and timely.\n5. Summary and conclusions: A case study research approach was applied to investigate several questions related to changes in students' creativity and anticipated sustainability practices associated with design thinking. Based on a qualitative content analysis of survey and focus group responses, students' perceptions of design thinking, in particular its impact on their creativity and future sustainable engineering practices, were favorable. Survey respondents most frequently identified the design-thinking process as the aspect of the course that will positively impact their sustainable engineering practices in the future. The specific sustainability topic areas and activities mentioned most frequently by survey respondents as being influential to their future career practices were the sustainable design frameworks, life-cycle-assessment and biomimicry, each mentioned by approximately one-fourth to one-third of survey respondents. The focus group participants freely mentioned design thinking in approximately two-thirds of responses about perceived changes in their creativity as a result of the course. A statistically significant increase in the project novelty score was found with design thinking, along with a medium-to-large effect size, which indicated practical significance of the result. Based on a qualitative content analysis of student reflections on their use of design thinking, the majority of reflections were (at a minimum) moderately comprehensive in their discussion of the prescribed design stages. Thus, students' application of the design-thinking process generally aligned with the positive direct and indirect assessment outcomes observed. The two-week, introduction-to-design-thinking course module and the notable use of active learning during class were the two most-frequent positive impacts on students' application of design thinking, each associated with over 40 per cent of survey responses. Thus, it appears that these two approaches and activities may be important when teaching sustainability courses in design thinking.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study was to assess particular student outcomes when design thinking was integrated into an environmental engineering course. The literature is increasingly promoting design thinking for addressing societal and environmental sustainability engineering challenges. Design thinking is a human-centered approach that identifies needs upfront.\nDesign/methodology/approach: In an undergraduate engineering course, Design for the Environment, students have begun to obtain hands-on experience in applying design thinking to sustainability challenges. This case study investigates the association between the use of design thinking and student creativity with sustainability design solutions. Student perspectives on their own creativity and future sustainable design practices as a result of the course were also investigated.\nFindings: The findings were favorable for design thinking, being associated with a significant difference and medium-to-large effect with regards to solution novelty. A qualitative analysis showed a positive association between design thinking and students' perceptions of their creativity and future anticipated sustainability practices. Using a content analysis of reflective writings, students' application of design thinking was assessed for comprehensiveness and correctness. A two-week introductory design-thinking module and significant use of in-class active learning were the course elements that most notably impacted students' use of design thinking.\nPractical implications: This case study preliminarily demonstrates that application of design thinking within an environmental engineering course may be associated with beneficial outcomes related to creativity and sustainability.\nOriginality/value: A review of the literature did not uncover studies of the use of design thinking for undergraduate socio-environmental challenges to promote creativity and sustainable-practices outcomes, although the literature has been calling for the marrying of these two areas.\n",
"role": "assistant"
}
] |
A review of machine efficiency in mass customization
|
[
"Metrics",
"Efficiency",
"Mass customization",
"Parameters",
"Capacity",
"Machine"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Mass customization (MC) emerged as a popular strategy in the late twentieth century with the manufacturing paradigm shifting from mass production through lean to agile manufacturing. Davis (1987) framed the word \"mass customization\" while outlaying the future roadmap for companies. Pine (1993) envisioned MC-based supply chains to be able to provide the variety that fulfills the specific choices and needs of customers with affordability. Kidd (1999) proposed that next generation enterprise has to be able to change to high-variety production in its niche as a part of transition to agility refers to embracing change through high variety from being lean which is generally not that variety intensive. Radder and Louw (1999) described MC to be able to cater any volume at any time for any variety for any number of customers. All industries today offer significant varieties in their products and services with an objective to fulfill choices of their customers and in order to delight them. In general, offering variety has become a zero-sum game wherein variety needs to be offered to be able to survive in a given business domain. Variety spins across features, combination of features, functions, market segmentation based on price and also service levels. This paper tries to explore about whether the MC poses a challenge to manufacturing in the realm of efficiency management and about the tools to achieve the same by identifying the influencing factors.\n2. Methodology: There is a shortage of direct literature connecting MC and its impact on operational efficiency though there are some initial definitions and fragmented research papers on the same. This research tries to close the gap by searching the connections scattered across the literature and making a summary that creates a link between the two.\n3. Issues and challenges impacting efficiency: In this section, a detailed explanation of identified issues is provided through a chronological review that was conducted using the method explained in Section 2 that is followed by a detailed explanation about each identified issue and the specific challenges posed by each of them toward machine efficiency.\n4. Chain diagram: A chain diagram is shown in Figure 1 connecting the 29 parameters to emulate a hypothetical dependent sequence of events starting from customer inputs and product variety which later flows into a sequence of events that impacts machine and manufacturing performance. Review shows that the various parameters are valid across all the different industries and applications with varying levels of intensity.\n5. Review of performance measures: A simultaneous review on efficiency measurement methods is considered important here and briefly explained to complement the MC efficiency research. The OEE metric was coined in 1971 in Japan.\n6. Metrics on degree of customization: For efficiency to be correlated with MC, we need to be able to measure the level or degree of MC. McCarthy (2004) explained that MC varies for different firms based on product volume/variety ratio, complexity and value of product complexity, point of customer involvement, the degree of involvement, the type of product modularity offered and the nature of the customized offering and the perceived value.\n7. Results of the literature analysis: Figure 2 depicts the summarized parameters according to the descending order of their rankings to highlight the most influential factors followed by the lesser influential ones. In total, 207 papers by different authors emphasize different parameters and issues on which they individually conducted their research. All these parameters were identified and ranked based on their frequency found across the reviewed papers. For the ranking purpose, each parameter is given unit weightage for occurring in an individual research paper, more occurrences in the same paper do not increase weightage but an occurrence in each of the 207 papers adds unit weightage for the studied parameter. To find the ranking, the sum of unit weightages is done for all the parameters after which the parameters are sorted from highest to lowest weightage of occurrence. All the parameters are explained in detail with the support of literature in Section 4 which follows this section. Interestingly, all papers (including newer papers published till early 2018) also mentioned about issues which are fitting exactly into these 29 parameters. This affirms that all the factors related to efficiency can be rightly summarized into the same.\n8. Toolkit of best practices: The 29 parameters and challenges are proposed to be addressed through actions in seven broad areas with their respective solution strategies and tools synthesized from this literature review. This framework can shape the success of MC along with meeting the efficiency objective. The seven areas identified through the review are maintenance practices, knowledge culture, process principles, design principles, manage uncertainty, OEE management and technology. Table II explains a summary of solutions to overcome the problem of achieving efficiency while trying to survive the MC environment related to the various issues and parameters mentioned in earlier section.\n9. Conclusion and implications: This review focused on closing the gap between efficiency and MC by exploring beyond the boundaries of available definitions through three decades of literature. It finds that MC can impact efficiency and has a potential for creating several new issues which may go hidden. The study identifies and classifies the challenges into 29 issue-based parameters that challenge machine efficiency in a MC environment. Ranking of the twenty nine issue-based parameters (as shown in Figure 2) explains the impact of MC on efficiency. Problems found from papers of current decade also fit into the 29 parameters. Figure 3 provides yearly trend of research paper count in the areas related to machine efficiency in MC from 1987 to early 2018. The trend shows that research in this area has gone up multifold in the current decade compared to the previous decades when the MC paradigm shift had just began.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to address the gap between definition and practical aspects of production efficiency in mass customization (MC). The paper summarizes all major issues impacting efficiency in MC. Also, the paper reviews metrics, relationship between various parameters and provides a best practices benchmark toolkit to achieve higher machine efficiencies.\nDesign/methodology/approach: The paper identified and categorized multiple challenges impacting machine efficiency in MC through a literature review spanning over three decades, and also ranked the identified issue-based parameters. Top issues were found varying across different types of industries identified through the review. Metrics pertaining to efficiency and degree of MC are reviewed in the paper. A chronological review of issues is presented, and a chain diagram is built in the paper. Toolkit of best practices created with solution strategies and tools are summarized through the review.\nFindings: The paper found that MC reasonably impacts machine efficiency which needs to be addressed. Major issues through literature review-based ranking are uncovered, and worldwide research trend and comparison are presented. Active research in this area is observed to be at its peak since 2010. The extensive use of strategies and benchmark toolkit for improving efficiency are summarized.\nResearch limitations/implications: Ranking of issues has been done through a literature review; hence, there can be skewness depending on the frequency of issues researched by various authors in various areas of industries.\nPractical implications: This paper is useful for manufacturing managers and companies willing to increase the size of their product portfolio and choices within their available resources without compromising machine efficiencies and, thereby, the cost. The identified issues help in providing a comprehensive issue list to the academia.\nOriginality/value: This paper describes what is believed to be the first study that explicitly examines the issues faced in achieving machine efficiency while manufacturing in an MC environment.\n",
"role": "assistant"
}
] |
Entrepreneurship students distilled their learning experience through reflective learning log
|
[
"Entrepreneurship",
"Thematic analysis",
"Critical reflection",
"Action research",
"Critical thinking",
"Reflective learning log"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Entrepreneurship has always been conceptualized as a process of discovering an opportunity, sourcing and manipulate resources, planning, and execution intelligence. An entrepreneur knows all the parts and knows how to match and fit all those resources together. The above narration signifies that a process is quite predictable, but in reality, the nature of entrepreneurial activities (to undertake) is not predictable at all. Greene (2014) proposed a portfolio of four complementary techniques for teaching entrepreneurship as a method, not as a process. They are: businesses start-up, reality-grounded simulations and gamification, design-oriented learning, and reflective practice.\nCritical reflective learning: It is not easy to write our thinking process in a form of reflection but it should be practiced in learning, it helps to encapsulate learning in much more meaningful manner. The use of reflection journals can assist to document thinking process which in turns helps teacher to draw conclusion on the learning progress of the students. The entire process of writing per se may have somehow encouraged students to reflect on what they have been taught in class and thus facilitate learning further. According to Cowan (1998, p. 16), a student is said to be doing reflection when \"[...] she notes that there is something different about the case that she is considering, in comparison with the examples she has encountered in class; and when she also identifies what the difference is, and what she should do about it.\"\nMethodology: The research methodology used in this study was classroom-based educational action research (Angelo, 1991; Elliott, 1991). Part of effective teaching is the ability to reflect on what is happening in the classroom, and to identify any differences in what was planned and what actually occurred. By conducting \"systematic, intentional inquiry\" within his/her own classroom, the instructor builds a better understanding of his/her own practice (Cochran-Smith and Lytle, 1993, p. 7).\nAnalyses and discussions: Based on the mark sheets for reflective learning log (see Table AII), Table I tabulates the frequency of score percentage for graded reflective learning log.\nDiscussions: The students' reflections on their entrepreneurship's experience could be grouped into four different reflection organizing themes comprised of:\nDescriptive writing (non-existence of reflection): In this SoTL, there were 75 samples of reflective learning logs that scored 5 percent and below 5 percent point (see Table AII). All of them achieved Level 1 (non-existence of reflection) in the reflective framework proposed by Hatton and Smith (1995). Descriptive writing is a description of events or literature reports. There is no discussion beyond description. The writing is considered not to show evidence of reflection. The followings are some of the accounts:Asmida: Franchise is a semi-independent business that individual pays the fees and royalty for the trademark that will be used to sell the product and their service. This business is easier compare to the individual that want to open their own business. For example like Mc Donald which has franchise for all over the whole that many franchisees pay the fees to run their business(sic).\nImplications and limitations: This study highlighted that reflective writing was an unfamiliar genre to many business students and needed to be explicitly taught in the respective subject. The results of this study have strongly suggested an urgent need for the development of reflective writing skills among the students during their higher education studies so as to assist in the professional and entrepreneurial behavior development as they entered the job market. In addition, there is a necessity to equip the undergraduate with basic building blocks needed to mold them as reflective practitioners. The reflective framework provided by Hatton and Smith (1995) proved to be useful in categorizing the written reflections and making the grading of the reflective learning log as less subjective. The weightage given to this graded assessment was 10 percent and it is considered substantial in grade. While the result of the study may not provide strong evidence on the ability of the students in reflective writing, it will be strange should inability exist as the outcome of attitudinal problem of willingly forgo a set of marks by not writing reflectively in which may warrant further investigation.\nConclusions: By maintaining a learning log for the past 14 weeks in this entrepreneurship module, students can record and comment their every week entrepreneurship lessons in a profound manner. They can self-track their perspectives in their first class in this module and monitor their evolved entrepreneurship knowledge and perspectives as the lecture progress until the end of the semester. In the process, they immerse in self-discovery in the world of entrepreneurship and self-tracking personal growth and cognitive development. Reflective learning log enables students to clearly identify their own thinking in entrepreneurship, the risks involved, the myths of becoming an entrepreneur, opportunity and execution intelligence - a method that promotes deep learning as depicted in the student's comments in the findings section in this study. Research has identified that reflection can help people to change. Although most of the students registering for this module are young undergraduate, and have limited experience to draw from to reflect upon certain issues (this is the findings from this study). But despite these barriers, reflective learning log makes entrepreneurship learning congruent with the suggestion to teach entrepreneurship as a method rather than as a process. Reflective learning log provides students with a profound insight into their own world of academic and; hopefully in future, their professional growth as reflection become a habit for my student.\n",
"role": "user"
},
{
"content": "Purpose: This Scholarship of Teaching and Learning research is a part of the larger study grant to analyze written reflections through learning log among the third and final year students undertaking BPME 3073 Entrepreneurship module in University Utara Malaysia (UUM). The paper aims to discuss these issues.\nDesign/methodology/approach: The data collection techniques are researcher-directed textual data through reflective learning log, taken from 140 students from three classes. A thematic approach was utilized to present the reflections of the students and all data were recorded in a verbatim format.\nFindings: The findings show that most students have never written a reflective log or essay in the formative assessment. As a consequence, they had difficulty in writing the reflection when being requested to do so. A total 75 (approximately 55 percent) of the reflective logs were identified as level 1 (from 1 to 5 percent) in which reflections were simply written in a descriptive manner, resulted in a balance of 61 learning logs being utilized for further analysis. The students' reflections on their entrepreneurship's experience systematically categorize into four different themes comprised of: the nature of entrepreneurship module, entrepreneurial characteristics, opportunity recognition, and creativity and innovation.\nResearch limitations/implications: As for the limitation of the study, it is important to not to underestimate the challenges of introducing a grade assessment that most of them are not familiar with in their university academic journey. Students need guidance, assurance and confidence writing something that require personal opinion, own thinking, sensitive and personal nature of narration. For most students as found out in this study, self-confessional writing is hard to come by (they dare not attempt it in the first place), only a handful appreciating the writing start with \"I,\" \"me\" as first person. More research in this study should be conducted across the university to gauge the response from the students to see if the result of this study is only applicable to this group of students or to this discipline of studies. The researchers would also like to recommend for future studies which take the form of a longitudinal study of similar kind to examine the problems and challenges with regards to promoting learning reflection at the undergraduate level.\nPractical implications: Based on the result of the 61 students who had demonstrated an ability in reflective writing, it is suggested that perhaps the university should consider offering coursework that contains a component of reflective writing as part of the assessment. As such, if this is implemented, students of such ability like the one in this sample group would have been benefitted from such assessment which look at reflective ability (Greene, 2014) and which they were allowed to form a broader perspective in relation to the module undertaken. This in turns will foster the growth of reflective ability which is recognized as a learned behavior (Gustafson and Bennett, 1999). In addition, for the future exercise of this reflective learning log, the researcher opined that we should encourage our students to engage with another student (e.g. close friend) in a way that encourages talking with, questioning, or confronting, helped the reflective process by placing the learner in a safe environment in which self-revelation can take place. In addition, students were able to distance themselves from their actions, ideas and beliefs, by holding them up for scrutiny in the company of a peer with whom they are willing to take such risks (Hatton and Smith, 1995).\nOriginality/value: The results of this research have strongly suggested the need to urgently develop among the students the skills in writing reflectively as they go through the process of higher education which is useful in molding their future professional and entrepreneurial behavior as when they entered the job market which requires a critical reasoning ability.\n",
"role": "assistant"
}
] |
Domestic structure and sub-national foreign policy: an explanatory framework
|
[
"Intergovernmental relations",
"Institutionalization",
"Domestic structure",
"Legal grounds",
"Level of regional autonomy",
"Sub-national entities",
"Sub-national foreign policy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The link between domestic structure and foreign policy has been studied by several scholars. In 1969, Kissinger advised that the USA should examine the interplay of domestic forces in the Socialist bloc to better understand the Soviet actions. He argued that domestic structure is crucial to the understanding a state's foreign policy, as it determines allocation of resources, interpretation of other countries' actions and the goals of a countries' foreign policy. Kissinger particularly focused on the nature of the administrative structure and the leadership as essential factors to the shaping of a state's relationships with other nations.\nMethodology: The study will be carried out on the basis of a theoretical framework proposed by the author to explain the impact of domestic structure on SFP (Figure 1). The proposed framework includes the following variables and their indicators:\nSub-national foreign policy: The practice of foreign affairs by sub-national entities or regions has been coined in a multitude of different ways[2]. SFP refers to direct international activities by sub-national actors. These activities can either be in concert with and complementary to, parallel to, or in conflict with parent state diplomacy (Duchacek, 1990; Kincaid, 1990; Criekemans, 2010). The patterns of the relation between SFP and national foreign policy has to be examined and evaluated on a policy by policy basis (Mccallion, 2011). SFP can be understood as \"a broadening of the universe of international affairs, in which states are no longer the sole actors\" (Keating, 1999: 9).\nDomestic structure and sub-national foreign policy: an explanatory framework: To understand the external activities of regional governments, one has to have a greater understanding of not only the domestic frameworks regions are working within, but also their ability to act within these frameworks. Domestic structure plays an important role in shaping and explaining SFP as it provides opportunities for action while imposing constraints. In essence, SFP does represent further evidence of the breakdown of the distinction between domestic and international affairs (Keating, 1999). Put differently, a clear understanding of the motivations for specific modes of SNE international activity needs to relate domestic to international forces and the interactions between the two (Hocking, 1999).\nFindings: The main finding of this study is that the domestic structural context is important in shaping and explaining the level and form of SFP. That is consistent with Lecours' (2002) conclusion that domestic opportunity structures may enable or constrain the international activity of regions.\nConclusion: This paper attempts to explain the impact of domestic structure on SFP. In doing so, domestic structure was broken down into four variables, including legal grounds, the level of autonomy, the pattern of intergovernmental relations, and institutionalization. Together, they can be utilized to measure the power of a region. We expect that more powerful regions will be more likely to try to influence, not just respond to, international phenomena. In their study on regional offices in the European Union, Marks et al. (2002) find out that regional offices representing more powerful regions will be more oriented towards gaining influence than those representing less powerful regions.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to examine the effects of domestic structure forces on \"sub-national\" foreign policy (SFP); an analytical concept provides a suitable operational framework for research on international activities of sub-national entities or regions.\nDesign/methodology/approach: The paper is carried out on the basis of a theoretical framework proposed by the author. First, the dependent variable is defined. Then, domestic structure is broken down into four variables, including legal grounds, the level of autonomy, the type of intergovernmental relations and institutionalization. A comparative method is used to examine the validity of the theoretical framework.\nFindings: The paper finds out that domestic structural forces influence level and form of SFP with some regularity. The influence of these forces on SFP can be explained, as they recur and have such consistent effects that they create patterns and regularities in SFP. Such regularities can be detected through systematic analysis.\nOriginality/value: The topic of SFP is relatively controversial because of academic debate over international agency of substate actors. However, it is a worthwhile subject of research, as it has the potential to revolutionize research in foreign policy analysis. Moreover, the phenomenon of SFP is in need of theorizing and comparison as the literature on SFP is still in its infancy.\n",
"role": "assistant"
}
] |
Employee service innovative behavior: The roles of leader-member exchange (LMX), work engagement, and job autonomy
|
[
"Employee behaviour",
"Innovation",
"Leaders",
"Service industries"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The world's emerging and advanced economies are showing an inclination toward services rather than products (Thakur and Hale, 2013), generating over 70 percent of the gross domestic product (GDP) through service sector only (Ostrom et al., 2010). However, to survive the intense competition in the business world today, organizations need to constantly make innovations in their services (Kindstrom et al., 2013). Service innovation brings value to the organization and acts as a medium for enhancing market efficiency and customer loyalty (Chen et al., 2011). The significance of service innovation has been acknowledged by both academicians and practitioners considering the fact that the service sector in both developed and developing economies largely contributes to wealth generation (Thakur and Hale, 2013). Despite the acceptance of the significance of service innovation, it remains an underdeveloped and less studied concept (Grawe et al., 2009; Dhar, 2015b). Hitherto, innovation studies have paid more attention to technological innovation by manufacturing organizations (Toivonen and Tuominen, 2009), and their focus has been more on product (i.e. goods) and process (i.e. production systems) innovation (see e.g. Pan and Li, 2016). This limited focus on service innovation may be the result of the belief that services are activities that do not require technological development or innovation (e.g. Pavitt, 1984; Kohler et al., 2012). Literature on product-process innovation emphasizes manufacturing as the main driving force in an economy. However, in emerging and advanced economies, the service sector is now contributing more toward the GDP (Ostrom et al., 2010). Thus, both services and service innovation have come to be considered as key elements for substantial economic growth (Thakur and Hale, 2013).\n2. Literature review and hypothesis development: 2.1 Employee service innovative behavior\n3. Research methodology: 3.1 Sample and procedure for data collection\n4. Findings: 4.1 Descriptive analysis\n5. Discussion: The present study examines employee service innovative behavior in the Indian banking industry, thereby contributing to the limited extant literature on service innovation (Grawe et al., 2009; Fujii et al., 2014). The study presents a model which comprises LMX, work engagement, job autonomy and employee service innovative behavior. Findings show that LMX is positively linked to employee service innovative behavior. Furthermore, findings lend support to the hypothesized effect of the mediating role of work engagement and the moderating influence of job autonomy. A closer investigation shows that work engagement fully mediates the positive relationship between LMX and employee service innovative behavior. In addition, examination of the interaction effect reveals that high levels of job autonomy strengthen the relationship between LMX and employee service innovative behavior mediated by work engagement by reinforcing the influence of work engagement on employee service innovative behavior.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to extend the under-researched work on service innovation by examining employee service innovative behavior in the Indian banking industry. In doing so, this study addresses the call for carrying out context-based research to advance service innovation literature.\nDesign/methodology/approach: Using convenience sampling and self-administered survey methods, data were obtained from professionals employed in Indian public sector banks (n=294).\nFindings: Findings reveal that leader-member exchange (LMX) shares a positive relationship with employee service innovative behavior via work engagement. Results also indicate that job autonomy as moderator strengthened the relationship between LMX and employee service innovative behavior mediated by work engagement.\nPractical implications: This study recommends that higher levels of LMX quality should be combined with enhanced levels of job autonomy for significantly influencing employee service innovative behavior.\nOriginality/value: Even though previous studies on service innovation have produced a significant piece of work, this study is among the first to propose and analyze a comprehensive and theoretically grounded structure of LMX, work engagement, job autonomy and employee service innovative behavior by incorporating social exchange and job characteristics theories.\n",
"role": "assistant"
}
] |
No more Mr. Nice Guy: social value orientation and abusive supervision
|
[
"SVO",
"Abusive supervision",
"Victimization",
"Justice"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Abusive supervision refers to subordinates' perceptions of the extent to which their supervisors engage in the sustained display of hostile verbal and non-verbal behaviors, excluding physical contact (Tepper, 2000, p. 178). Although it has detrimental effects on employee productivity and well-being, relatively less is known concerning its antecedents (Tepper et al., 2017). A growing stream of research suggests that subordinate and supervisor characteristics, such as low conscientiousness, high neuroticism, and low agreeableness (Breevaart and de Vries, 2017; Henle and Gross, 2014; Wang et al., 2015), may be important predictors of abusive supervision.Importantly, Aquino and Thau (2009) suggest that employees with a selfish motivation may be targets of aggression motivated by punitive sentiments. Organizational justice research (e.g. Jawahar, 2002) also suggests that actions such as violations of norms, rules, and expectations may engender feelings of injustice and result in aggression toward violators. Based on Aquino and Lamertz (2004)'s relational model of workplace victimization, this fits into the workplace aggression by a reactive perpetrator directed at a provocative victim. A \"reactive\" perpetrator only aggresses with the intention to retaliate against a \"provocative\" victim who violates norms, rules, or principles that the reactive perpetrator values. Thus, supervisors with a concern for collective welfare may show abusive supervision to subordinates who are mainly concerned with their self-interests.Under this lens, we argue that a particularly relevant predictor of such abusive supervision is social value orientation (or SVO), which refers to individuals' relatively stable preferences concerning the distribution of outcomes between the self and others in social dilemma situations (Messick and McClintock, 1968; Van Lange, 1999). In other words, SVO corresponds to the weights individuals assign to outcomes for the self and outcomes for others in interdependent situations. Most people are classified as either prosocials or proselfs (Van Lange, 1999). Prosocials (or cooperators) assign positive weight to both their own and others' outcomes, value fairness, and thus attempt to maximize joint outcomes, or collective welfare. Proselfs assign positive weight to their own outcomes but assign zero or negative weight to the outcomes of others. Thus, while proselfs are concerned with narrow self-interests, prosocials are concerned with collective welfare. Supporting this conceptualization, research shows that prosocials are more cooperative than proselfs in social dilemmas (Balliet et al., 2009), and this effect appears to exist above and beyond the effects of the big five traits and honesty-humility (Hilbig and Zettler, 2009; Hilbig et al., 2014).In the organizational context, Nauta et al. (2002) found that Prosocials show higher concern for organizational goals and engage in higher interdepartmental problem-solving behaviors than Proselfs. Prosocials also view uncooperative others as immoral (Van Lange and Kuhlman, 1994), show negative emotional reactions to others' noncooperation (Stouten et al., 2005), and tend to punish those who, they believe, have selfish intentions (Steinel and De Dreu, 2004). Thus, drawing on the relational model of workplace victimization and the literature on SVO we argue that supervisors who endorse a prosocial orientation would be reactive, in this case by showing more abusive behaviors, to subordinates who have a proself orientation (i.e. who lack a concern for collective welfare). We also examine the consequence of this interactive effect for subordinates' in-role performance (Figure 1).The present research contributes to the growing literature on the role of subordinate and supervisor characteristics in the occurrence of abusive supervision. Scholars (e.g. Martinko et al., 2013; Tepper et al., 2017) often call for integrating subordinate and supervisor related factors to better understand the development of abusive supervision. Yet, research mostly focuses on subordinate and supervisor characteristics in isolation (e.g. Henle and Gross, 2014; Wang et al., 2015; Breevaart and de Vries, 2017). Further, abusive supervision by a reactive perpetrator toward a provocative subordinate is likely to be severe and systematic (Aquino and Lamertz, 2004). To address this call, we examine how supervisor and subordinate SVOs may potentially produce a reactive supervisor - provocative subordinate relationship characterized by increased abusive supervision. Moreover, SVO is an important determinant of cooperative behavior in interdependent social interactions (Balliet et al., 2009) and is a stronger predictor of cooperation than the big five dimensions and honesty-humility trait (Hilbig and Zettler, 2009; Hilbig et al., 2014). However, its effects in social interactions in organizations remain largely unknown. Therefore, the present research also contributes to the literature on SVO by extending our knowledge of its implications to the abusive supervision literature.Integrating subordinate and supervisor characteristics\nStudy 1 method: Sample and procedure\nResults: Data analytic strategy\nStudy 2: We conducted a second study in order to address three limitations of study 1. First, it tested the role of employees' and supervisors' general cooperative motives, as measured by their SVOs, in the development of abusive supervision. Although previous research shows that SVO predicts employees' concern for organizational goals (Nauta et al., 2002), study 1 did not explicitly test the role of subordinate and supervisor prosocial/proself motives toward the organization in the development of abusive supervision. In study 2, we manipulated prosocial and proself SVOs in relation to supervisors' and subordinates' high/low concerns for the organization, respectively.Second, in study 1, abusive supervision represented subordinates' subjective perceptions. Thus, proself subordinates may have interpreted \"ambiguous supervisory behaviors' from prosocial supervisors as abusive. To address this issue, in study 2 we adapted scenarios that represent actual abusive behaviors. Finally, in study 2 we randomly assigned participants to experimental conditions to strengthen our causal claims. Participants assumed a supervisor role who had either high concern or low concern for the organization. They read a scenario describing abusive supervisory behaviors towards one of their subordinates who was also described as either having high concern (prosocial) or low concern (proself) for the organization. Participants then responded to several questions assessing their views about the abusive behavior. In line with hypothesis 1, we predicted that participants in the prosocial motivation condition (i.e. supervisor with high concern for the organization) would be more likely to assess the abusive behaviors in the scenario as \"more positive\" when the subordinate had low concern for the organization (proself).Sample and procedure\nResults: Manipulation checks. Participants rated the employee as significantly less prosocial in the low organizational concern condition (M = 2.81 SD = 0.80) than in the high condition (M = 5.02 SD = 1.03), F(1,76) = 111.98, p < 0.001. Likewise, participants rated themselves (supervisors) as significantly less prosocial in the low organizational concern condition (M = 2.93 SD = 1.13) than in the high condition (M = 4.14 SD = 1.19), F(1,76) = 21.03, p < 0.001. These results support the effectiveness of our manipulations.Hypothesis test. We performed a full factorial ANOVA with supervisor's and subordinate's concern and gender as fixed predictors and participants' views of the abusive supervisory treatment as the dependent variable. There was a main effect for gender, F(1,67) = 9.53, p < 0.01, e2 = 0.12, CI = 0.03, .25, as males (M = 2.08 SD = 0.84) viewed the abusive treatment as more positively than females (M = 1.56 SD = 0.64). There was not a main effect for supervisor's, F(1,67) = 0.01, ns, e2 = 0, CI = -0.00, 0.01, or subordinate's organizational concern, F(1,67) = 1.73, ns, e2 = 0.03, CI = -0.00, 0.11. However, their interaction was significant, F(1,67) = 4.72, p = 0.033, e2 = 0.07, CI = 0.003, 0.18. When the supervisor had low organizational concern, there was no difference in the assessment of abusive behavior across low (M = 1.65 SD = 0.82) and high (M = 1.78 SD = 0.76) employee concern, F(1,67) = 0.35, ns, e2 = 0.01, CI = -0.00, 0.07. However, participants in the supervisor high organizational concern condition viewed abusive treatment more positively when the employee had a low concern for the organization (M = 2.1 SD = 0.74) than when the employee had a high concern for the organization, (M = 1.49 SD = 0.71), F(1,67) = 6.49, p = 0.013, e2 = 0.09, CI = 0.01, 0.21. These findings are consistent with the SVO interaction effect in Study 1.\nGeneral discussion: The present research examined the joint roles of subordinate and supervisor SVOs in predicting abusive supervision and its consequences for in-role performance. Supporting our hypothesized model, in study 1 we found that when supervisors were higher in prosocial motivation, subordinates with a lower prosocial motivation (i.e. proself) were more likely to perceive abusive supervision, which, in return, was negatively related to in-role performance. Study 2 revealed a similar interaction effect while manipulating supervisor and subordinate orientations toward organizational goals.Theoretical implications\nConclusion: Our research suggests that \"nice,\" cooperative supervisors may show abusive supervision if paired with uncooperative subordinates. Thus, the right pairing of individuals in the organization, based on their SVOs, might be as important as hiring the \"right\" people.\n",
"role": "user"
},
{
"content": "Purpose: The present research examined the interactive effect of subordinates' and their supervisors' social value orientations (SVO) on abusive supervision and its consequence for in-role performance.\nDesign/methodology/approach: In study 1, we provided a survey to 420 subordinates and 115 supervisors from 42 organizations. HLM was used to test the hypothesized cross-level moderated mediation model. In study 2, 78 participants were asked to imagine they were a supervisor and responded to a potential scenario where supervisor and subordinate prosocial and proself orientations toward the organization were manipulated (2 x 2 design).\nFindings: Study 1 showed that when supervisors have a higher prosocial motivation, subordinates who are more self-interested (proself) report more abuse than those with a higher prosocial motivation, with negative consequences for in-role performance. Study 2 replicated the pattern: participants (in the role as supervisor) with induced prosocial goals rated abusive supervision behaviors as more justified and acceptable toward a proself employee than they did toward a prosocial employee.\nOriginality/value: This research is innovative by bridging SVO and organizational literatures and demonstrating that a dyadic interaction between a proself subordinate and a prosocial supervisor may produce a reactive perpetrator - provocative victim relationship characterized by higher abusive supervision.\n",
"role": "assistant"
}
] |
Role of gender and corporate risk taking
|
[
"Board of directors",
"Risk taking",
"Glass cliff",
"Gender",
"Chief financial officer"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: There has been a growing interest in analyzing gender differences in management styles and decision-making. This has followed a rise in the number of females going to business schools, entering the workforce and their steady climb up the management ladder. McGregor (2017) reports that the number of female chief executive officers (CEOs) reached a record high of 32 (6.4 per cent) at Fortune 500 companies in 2016 compared with 24 companies in 2014. Similarly, Zwirn (2016) points out that females make up 13.8 per cent of the Fortune 500 companies' chief financial officers (CFOs) in 2015, compared with 6.8 per cent in 2006. Additionally, in 2015, females occupied 19.9 per cent of the board seats of the Fortune 500 companies compared with 9.6 per cent in 1996 (Catalyst, 2015; Brown, 2017). With this slow but gradual rise in gender diversity in senior management ranks, there has been continued interest in determining whether there are differences in risk-taking behavior between male and female leaders.\n2. Analysis of risk taking: role of gender: Until recently, much of the research on corporate decision-making in finance had focused primarily on the role played by firm characteristics rather than characteristics of managers leading these organizations. Researchers such as Adhikari et al. (2019, p. 2) have attributed this to the fact that preference for risk is \"an innate personality trait which is largely unobservable.\" Yet, in practice, executives who lead the firms display heterogeneity in preference and behavior and differences in their leadership styles that affect the financial performance of firms they manage and lead. For example, as pointed out by Huang and Kisgen (2013), Apple would not be the same company without Steve Jobs at its helm. Adhikari et al. (2019), Faccio et al. (2016), Midavaine et al. (2016) and Van Knippenberg and Schippers (2007) have also argued that characteristics of teams, such as gender and diversity, can influence firms' financial performance and outcomes. In corporate investment decisions, it is equally important to study both the investment and the investors' behaviors because the success or failure of an investment is undoubtedly tied to the investors' decisions and characteristics. In recognition of the role played by individual characteristics in decision-making, gender behavioral differences that were traditionally studied in cognitive sciences such as psychology have recently garnered attention among researchers in the field of behavioral economics and finance. Research by Baker et al. (2007), Bertrand and Schoar (2003) and Malmendier and Tate (2005), among others, focuse increasingly on the role of managerial characteristics in business decision-making.\n3. Data and methodology: The CFOs and their board members selected for this study are chosen from the 2014 ranking of Top 250 CFOs published by ExecRank (2020). ExecRank used proprietary methods for evaluation and ranking based on statistical and algorithmic analysis of 24 key areas such as experience in the role, business results during tenure, board member appointments, significant professional achievements, published thought leadership (books, speeches, articles and media mentions), technology adaption, education, industry professional reputation and peer recommendations and nominations. The exacting methodology used by ExecRank provides high confidence in the quality of CFOs for this study. We focus on the top 38 male and 20 female CFOs who served at least five consecutive years as CFOs between 2012 and 2016. Although 25 female CFOs make up the exhaustive list of top female executives in ExecRank's top 250 CFOs for 2014, only 20 of these females are included in this study based on the availability of data for the five years of study. Using a tenure of five years as a cut-off is appropriate because, as pointed out by Huang and Kisgen (2013), it takes at least one to two years before an executive can make a significant impact on corporate policy. In instances where executives have served consecutively as CFOs for longer than five years, they are also included but only the 2012-2016 data are considered in this paper. After the selection, the work history and tenure of these CFOs are verified using Bloomberg and LinkedIn websites. In addition, data on the PPE investment, debt/equity ratios, stock returns, net revenue (NR), net income (NI), number of employees and gender composition of the Board of Directors are collected primarily from the company annual reports.\n4. Findings: The descriptive statistics that are presented in Appendix 2 show the 58 firms' PPE, debt/equity ratio, stock return volatility, number of employees, NI, NR and share of female board members. The data are grouped by the gender of CFOs leading these businesses. Although female CFOs lead smaller firms (measured by the number of employees) on average, they are also found in businesses with relatively smaller NI and NR with the average values of $2.05bn and $26.6bn, respectively, compared with the average of $3.2bn in NI and $44.4bn in NR for male CFO-led firms. Tests for mean differences in size, NI and NR between the two gender groups are found to be significant at 5 per cent.\n5. Conclusion: Findings in this paper show that having female CFOs and a larger share of females on the board of directors influence the risk-taking behavior of firms analyzed. In examining the impact of CFOs' gender on risk taking, the results are mixed and depend on the selection of risk measures used. Female CFOs show risk aversion when the debt/equity ratio is used as a risk-taking measure. However, the effect of having female CFO is found to be positive, although insignificant, for stock volatility compared with the significant positive influence when the PPE measure is used. Meanwhile, we find that a firm's risk taking generally decreases with having a greater share of females on its board.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to examine the role of chief financial officers' (CFOs') gender in financial risk taking of 58 US companies along with the impact of having women board members.\nDesign/methodology/approach: Using a panel data of 58 selected S&P 500 companies during the period 2012-2016, this paper determines whether the gender of CFOs and having women board members play a role in risk-taking behavior of firms.\nFindings: Firms led by female CFOs are smaller in size with lower net income and net revenue. The panel data analysis shows that the impact of female CFOs on firms' financial risk is mixed, depending on risk measures used, whereas increasing female board members reduces that risk.\nResearch limitations/implications: The data used is limited to 58 S&P 500 companies, and two of the three risk-taking measures used in the study, specifically investment in property, plant and equipment (PPE) and debt/equity ratio, may not be applicable to some industries.\nPractical implications: The findings provide mixed evidence of risk aversion by females in executive and leadership positions, depending on the measures used and the management responsibilities they undertake (CFO versus board member) with support for the glass cliff phenomenon in which females may be leading financially precarious organizations.\nSocial implications: Female CFOs are found to be leading relatively smaller and financially poor-performing firms compared with the male CFO-led firms, thereby giving support to the glass cliff arguments.\nOriginality/value: The paper examines the role of CFOs' gender and board diversity in risk taking as measured by the investment in PPE, debt/equity ratio and stock return volatility.\n",
"role": "assistant"
}
] |
School teachers' job satisfaction and personal characteristics: A quantitative research study in Greece
|
[
"Job satisfaction",
"Personal characteristics",
"Primary education",
"School teachers"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper addresses issues related to the job satisfaction of primary school teachers in Greece. As a country which is experiencing a major economic crisis, Greece provides an appropriate context within which to examine the relationship between job satisfaction and education as a means of boosting productivity for the future development of the country.\nReview of relevant literature: Based on relevant literature, job satisfaction can be defined in many ways. According to some researchers (e.g. Arches, 1991; Butler, 1990; Dressel, 1982) job satisfaction can mean the positive emotional state of a person arising from an assessment of the work situation and is related to the job's characteristics and requirements. Locke (1976, p. 1304) defines job satisfaction as \"an enjoyable or pleasurable emotional state, which is the result of the valuation of work or employment experiences of a person.\" Hulin and Judge (2003) indicate that job satisfaction involves multidimensional psychological responses to one's work and that these responses have cognitive, emotional and behavioral components. Dawis and Lofquist (1984) determine job satisfaction to be the result of an evaluation by the employee on the extent to which the working environment meets their personal needs.\nAims of the study: As outlined above, research on the effects of individual characteristics on teachers' job satisfaction in Greece has only taken place over the last decade. As far as the authors are aware, there has been no empirical research on this matter in the last three years - a period of time when the country (and public sector schooling in particular) has been hit by a heavy economic recession, with adverse consequences for school teachers and their levels of job satisfaction.\nMethodology: In order to collect primary source data for this survey, 360 questionnaires were administered to primary school teachers in the metropolitan area of Athens (region of Attiki). The reasons for choosing the region of Attiki (the largest metropolitan region of the country) to conduct the study was fourfold: Attiki (Athens) dominates every facet of social and economic activity; the region is characterized by a high degree of urbanization; according to the population census of 2011, about 35 percent of the total Greek population live in the prefecture of Attiki whereas the demographic characteristics of the population vary considerably; and it covers a large number of public primary schools. The latter implies that a large number of primary school teachers are located in the prefecture of Attiki (Athens).\nResults: Personal characteristics of the primary school teachers\nDiscussion: The findings of this study imply that Greek school teachers are generally satisfied with their profession. This result, which is in agreement with the conclusions of Koustelios et al. (2004), may be attributed to the fact that school teachers in Greece are public servants and that the public sector offers job stability and security. Indeed, perceptual and attitudinal variables such as the intention of quitting and commitment to the organization were strongly correlated with the job satisfaction in those jobs related to human services (Spector, 1985). In the Greek context, the public sector secures jobs and offers employees stability. Hence Greek public servants (and educators working in public education) do not entertain the idea of quitting since their job in the sector is protected constitutionally. The feeling of job stability, even when the jobs are low-paid, enhances job satisfaction. And perhaps Greek school teachers' general satisfaction from their profession mainly stems from the actual attitude and the socio-economic structure of Greek citizens' receptiveness toward employment in the public sector.\nConclusions: It is evident from this study that, although personal characteristics of the Greek school teachers do not correlate with their overall satisfaction, they appear to be generally satisfied with their profession. However, an interesting finding of this study is that Greek school teachers are more satisfied with three aspects of job satisfaction, namely, \"administration,\" \"colleagues\" and \"nature of work\" and less satisfied with \"salary,\" \"benefits\" and \"potential rewards.\" Considering that over the last two years Greece has been suffering from a deep recession and severe economic problems, the latter result is of crucial importance. Given that the way out of recession and toward both financial and social development depends mainly on the quality of the education provided, quality in education is heavily associated with the workforce (the teaching staff), and job satisfaction is a strong determinant in retaining qualified teachers, then the results of this study may be considered as an extension of the knowledge base as regards job satisfaction within the Greek educational context and may help Greek policy makers to recognize it as an acute issue for boosting productivity and for the future development of the country.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper (based on the relevant literature) is to: investigate, through empirical analysis, primary school teachers' perceptions regarding their job satisfaction, and examine whether or not the personal characteristics of primary school educators (such as gender, age, family status, educational level, and the total years of service in public primary education) have any impact on their job satisfaction.\nDesign/methodology/approach: In total, 360 questionnaires were administered to primary school teachers in the metropolitan area of Athens (region of Attiki). The sample was randomly selected. The questionnaire was based on 41 closed and was divided into two sections. The Job Satisfaction Survey developed by Spector (1985) was implemented.\nFindings: Greek school teachers are generally satisfied with their profession. There is no statistical correlation between personal characteristics and the overall satisfaction while indicated that teachers are more satisfied with three aspects (subscales) of job satisfaction, namely, \"administration,\" \"colleagues\" and \"nature of work\" and less satisfied with \"salary,\" \"benefits\" and \"potential rewards.\" Age correlates with the levels of satisfaction with reference to administration, potential rewards, colleagues and the nature of work. The overall satisfaction positively correlates with all nine aspects of job satisfaction (subscales) and gender affects the aspects of \"promotion\" and \"colleagues.\"\nResearch limitations/implications: This study only analyzes a small sample from the Athens region and hence the results cannot be used to generalize about the whole of Greece. Since other Greek regions operate in different socio-economic environments, an analysis of additional data from other regions (rural and urban areas) would be necessary to compare and confirm the results.\nOriginality/value: The findings of this study a valuable extension of other relevant research as it provides the first empirical study of the Greek school system, investigating the relationship between certain aspects of job satisfaction and the personal characteristics of school educators as well as the relationship between these aspects of job satisfaction and total satisfaction. In the context of efficient educational policy, a greater understanding of educators' job satisfaction could facilitate the development of more effective policy practice that would increase not only the level of educators' satisfaction, commitment and morale but also improve the performance of the school system.\n",
"role": "assistant"
}
] |
Study shows positive impact of numeric diversity and diversity climate on psychological outcomes for faculty of color
|
[
"Stress",
"Discrimination",
"Faculty of color",
"Invisible labor",
"Numeric diversity",
"Racial climate"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: A study of academics at US colleges and universities has shown that both numeric diversity and racial climate have a big influence on psychological outcomes for faculty of color (FOC). The researchers went further than previous studies of FOC in highlighting the \"additive\" effects of having both numerically diverse staff and a positive racial climate. The main driver, they suggested, in mitigating the disparities in outcomes for FOC and white faculty was diversity climate, although numeric diversity should not be neglected.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of the study was to determine, first, whether both numeric diversity and racial climate impacted the psychological well-being and workplace experiences of faculty of color (FOC). But the authors also considered whether there was an \"additive\" effect when both diversity factors existed at the same time.\nDesign/methodology/approach: The authors used surveys of academics from various backgrounds in multiple US regions to test the effects of numeric diversity and racial climate on three dependent variables - invisible labor, stress from discrimination and dissatisfaction with co-workers. For the purposes of the study, FOC were defined as black, Latinx and Asian faculty.\nFindings: The results showed significant support for the authors' prediction that there was more stress from discrimination, invisible labor, and co-worker dissatisfaction against FOC in institutions with low numeric diversity and poor racial climate compared with institutions with high numeric diversity and a positive racial climate. They also found that negative impacts were smaller in institutions with both high numeric diversity and a positive racial climate, compared to institutions with high numeric diversity, but a poor racial climate. However, similar results were not found for institutions with low numeric diversity and positive racial climate.\nOriginality/value: The authors concluded that the findings showed that \"diversity climate may be the primary driver of mitigating psychological disparities between FOC and white faculty\". They said that education officials should take action to construct a positive racial climate, but neither should they ignore numeric representation.\n",
"role": "assistant"
}
] |
Stages of diversification in high performing Asian economies
|
[
"Economic development",
"High performing Asian economies",
"Industrialization",
"Manufacturing concentration"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Several studies on the evolution of industrial structure over the economic development path discovered a U-shaped relationship between industrial concentration and per capita income (Imbs and Wacziarg, 2003; Imbs et al., 2012; Samaniego and Sun, 2012; Clark and Sawyer, 2014; Clark et al., 2016). Poor countries have high levels of manufacturing concentration. Early in the development process, countries begin to diversify their manufacturing base in the sense that economic activity is spread across more sectors. This process takes place over most of the development path. A point is reached late in the development process when countries start to specialize again. Specialization is encouraged by policies that channel scarce resources to comparative advantage activities and stimulate exports.\n2. Industrial concentration and development: The relationship between manufacturing industry concentration and economic development is a classic question in the development literature, going back at least to Chenery et al. (1986). Imbs and Wacziarg (2003) discovered a U-shaped relationship between manufacturing concentration and per capita income in a large cross-section of countries. Poor countries were found to have high levels of manufacturing concentration. These countries initially exploit natural resource endowments. Early in the development process, countries begin to diversify their manufacturing base in the sense that economic activity is spread across more industries. Facing high trade costs, growing diversity in consumer tastes is satisfied by opening up new activities domestically rather than by relying on expensive imported goods. A relatively closed economy will develop new industries as productive factors are accumulated. Expanding the number of industries serves to dampen aggregate effects of sector-specific shocks (Koren and Tenreyro, 2007). Diversification continues until relatively late in the development process. A point is eventually reached beyond which countries start to specialize again. This sequence of diversification and specialization over the development path is encouraged by the interaction of productivity increases and trade costs. Initially, an increase in a country's aggregate productivity level relative to other countries expands the range of goods produced. Later in the development process, falling trade costs provide incentives to specialize in accordance with comparative advantage as countries grew and became more open to trade.\n3. Industrial policy: Industrial policy refers to government assistance using policies that involve more than simply altering industrial structure (Lall, 1994). Policies are needed to encourage diversification into new areas of production and technological advance beyond that permitted by market forces. The practice of industrial policy requires governments to favor some manufacturing industries over others. According to Rodrik (2004, 2009), government interventions must assist countries in diversifying their economy early in the development process and generating new areas of comparative advantage. The government must help entrepreneurs identify new products that can be profitably produced[2]. This will require strategic coordination between government policy makers and private firms to identify significant obstacles to restructuring and to implement policies to remove them. Industrial restructuring will require significant government assistance.\n4. Policies in HPAEs: There is no single HPAE model of economic development. HPAEs used very different combinations of interventionist and market-oriented policies to foster industrial development and diversification by encouraging technical upgrading of products and processes, developing new products, raising local content, deepening the base of technological capabilities, and by promoting exports. Policy interventions typically included targeted and subsidized credit to specific industries, keeping borrowing rates low, establishing and supporting government banks, investment guarantees conditioned on performance, protection of domestic industries, subsidies to offset the resulting anti-export bias, upgrading labor skills, acquiring technology, public investments in applied research, firm and industry-specific export targets, and sharing information between firms and the public sector.\n5. Empirical strategy: Our panel data set consists of eight HPAEs and 18 two-digit ISIC manufacturing industries over the 1963-2009 period. Yearly coverage varies across countries because manufacturing value-added data are limited in certain cases[3]. The Gini concentration ratio is based on industry shares in manufacturing value added, reported by United Nations Industrial Development Organization (2012)[4]. Per capita income, measured in year 2005 constant US dollars, is from the World Bank (2012) and the Council for Economic Planning and Development (2013).\n6. Results: Figure 2 shows the relationship between manufacturing concentration and real per capita income from the lowess procedure. The figure reveals what a typical country can expect to experience as it moves along the Asian economic development path. The non-parametric curve has three distinct stages. Diversification takes place quickly over the first stage that ends at $250 per capita GDP measured in 2005 US dollars. Considerable differences in degrees of diversification are found at the low end of the income scale. Per capita income does not rise rapidly over this stage. This suggests diversification alone cannot be expected to encourage economic development in poorer countries until obstacles such as weak institutions, poor governance, corruption, and a lack of infrastructure are addressed. Diversification and development both rise over the second stage. As countries attain higher levels of development, the diversification process slows down. Limits to diversification may be due to economies of scale considerations, the need to form linkages among manufacturing activities, and the desire to not enter too many activities. The turning point is estimated to be $20,000 real per capita GDP, beyond which countries start to specialize in accordance with comparative advantage. Countries further along the development path such as Japan, Korea, and Taiwan made extensive use of targeted intervention policies. A U-shaped relationship between manufacturing concentration and per capita widely held to encourage economic development is found for HPAEs.\n7. Conclusions: This paper studied the evolution of manufacturing industry structure over the economic development path for eight HPAEs. Increasingly prosperous HPAEs have long been hailed as models for success by other emerging economies. Focusing on a regional group of high performing economies enables us to relate policies used directly to observed patterns of manufacturing diversification. The evolution of industrial structure in HPAEs was found to exhibit the traditional U-shaped pattern reported in earlier studies that were based on a large cross-section of countries. Diversification of manufacturing activity was encouraged early in the development process by using interventionist and market-oriented policies to transfer resources from traditional to non-traditional activities and to develop new areas of comparative advantage. Specialization, much later in the process, involved steering resources to fewer export industries. Japan, South Korea, and Taiwan are held to have used the most extensive set of government interventions. These countries are further along the development path than are Indonesia, Malaysia, and Thailand. Given the success of HPAEs, and a heavy reliance on government interventions, we conclude that industrial policy did not disrupt and may have encouraged the pattern of industrial diversification followed by specialization held to foster economic development.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to determine whether the evolution of industry structure in the World Bank's eight high performing Asian economies (HPAEs) displays the U-shaped relationship between manufacturing concentration and per capita income widely held to foster economic development. Increasingly prosperous HPAEs have long been hailed as models for success by other emerging economies. Focusing on a regional group of high performing economies enables us to relate policies used by successful HPAEs directly to observed patterns of manufacturing diversification and provide policy guidance to emerging economies.\nDesign/methodology/approach: A robust locally weighted scatterplot smoothing procedure is employed to generate the U-shaped relationship between manufacturing concentration and level of economic development. Policies used by the most successful HPAEs are discussed.\nFindings: The relationship between manufacturing concentration and level of economic development is found to be U-shaped. Diversification of manufacturing is a prerequisite for successful economic development. Countries further along the economic development path such as Japan, South Korea, and Taiwan made extensive use of active and selective interventionist policies to diversify manufacturing before eventually specializing in a narrower range of export activities.\nPractical implications: Emerging economies should follow examples set by the most successful HPAEs that demonstrated significant government assistance is required to foster economic development.\nOriginality/value: The paper is the first to investigate the evolution of manufacturing concentration over the economic development path HPAEs. Success enjoyed by HPAEs holds important lessons for developing and emerging economies.\n",
"role": "assistant"
}
] |
Understanding Finnish student craft teachers' conceptions of sustainability
|
[
"Higher education",
"Sustainability",
"Education for sustainable development",
"Craft education",
"Student craft teachers",
"Sustainable craft"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: An old Finnish saying translates Let's model it with iron wire, when trying to understand and/or explain something difficult. It means that at times abstract concepts are hard to grasp, yet made concrete through \"iron wire\", they become comprehensive. Similarly, learning by doing in the context of education for sustainable development (ESD) is regarded effective (Tarrant and Thiele, 2016) and crafts has been discussed to be effective in concretising abstract concepts (Bennett, 2015). The role of craft in the framework of sustainability has been recognised and more emphasis in craft education is needed not only for the fading traditions but also for the importance of maintaining the craft-related abilities (e.g. to produce, repair and maintain products) and develop abilities to reflect and evaluate products and their properties (Cox and Bebbington, 2014).\nEducation for sustainability through craft: Initially, ESD was launched as part of environmental education and has prompted intense ongoing discussion and research ever since (Leal Filho et al., 2015). During this discourse, concern has been expressed that ESD in the professional development of educators is more focussed on administrative certification of environmental behaviour than on shifting the curriculum, teaching and learning aimed at sustainability (Wals, 2009). However, contemporary research on ESD suggests that a holistic understanding of interdisciplinary systems and a reflective pedagogical approach improves students' conceptions of SD at a practical level (Biedenweg et al., 2013; Coops et al., 2015; Howlett et al., 2016). The concept of ESD encompasses critical thinking and systems thinking skills as well as communication and collaboration (Tarrant and Thiele, 2016). Howlett et al. (2016) argue that curricula and pedagogical practices in higher education require interdisciplinary and creative ways of thinking to develop the appropriate mindset for a sustainable future. The purpose of ESD in higher education is to acquire and generate relevant knowledge, to reflect on decisions related to global and future-oriented responsibility and to understand the effects and complexity of behaviour (Barth et al., 2007).\nMethodology: Data collection and participants\nResults: The data and analysis revealed rich and detailed information about sustainable craft from the perspective of the student craft teacher. These conceptions of sustainable craft were first displayed as a system formed by product, practice and immaterial craft. Second, qualitative properties were quantified and statistically described, and dependencies were explored and analysed in relation to the structuredness of conceptions. The observed differences between these student teachers suggested conceptual differences in their understandings of sustainability through craft regarding practice, product, immaterial and holistic orientation.\nDiscussion: This study investigated the conceptions of craft student teachers in the form of written essays to understand their ideas and to develop craft education to achieve a higher level of sustainability (cf. Cox and Bebbington, 2014; Wals, 2009). The findings suggest that the concept of sustainable craft is an intertwined system of three fundamental elements, each with its own character and properties, as argued in a recent industry study of professional craft practitioners (Vaananen et al., 2017). The present study views sustainable craft from an educational perspective. Combining these findings with the existing literature on sustainability and craft, it is possible to draw some conclusions about approaches to sustainability through craft (see Figure 3).\nConclusions: This study articulates a holistic view of sustainability through craft; the system of sustainable craft as practice, product and immaterial craft and the orientations and approaches to SD through craft. The emergent framework provides a holistic understanding of sustainability and contributes novel information to the discussion of ESD. This study explains the intertwined nature of these different aspects in detail. On theoretical level this study confirmed the connections between variables, for instance, how skills, techniques and design connect the sustainable practice in making the crafted products. On practical level, these results reinforce practicing crafts demonstrate understanding of sustainability, cultural change and value the products affiliated with, in a concrete manner. The results signify the ability to commence critical thinking, attached to the maker's values, develops together with one's skills and knowledge through craft. The diverging emphases in these student teachers' descriptions and their varied knowledge of sustainable craft suggests a need to rethink craft education pedagogy at all levels, including higher education and, to diversify these conceptions through the curriculum. It is important to gradually turn the focus from one aspect to holistic and systemic thinking.\n",
"role": "user"
},
{
"content": "Purpose: This study aims to explore student craft teachers' conceptions of sustainable craft. This is an important issue because the Finnish National Curriculum of Basic Education emphasises sustainability, especially in craft education, and teachers play a vital role in preparing pupils to meet the future challenges. Because the concept of sustainable craft is open-ended, there is a need to rethink pedagogy in craft education and higher education.\nDesign/methodology/approach: The data were collected in the form of essays from future craft teachers (N41) studying craft science in the Finnish higher education system. The essays concerned both general conceptions of sustainable craft and reflections on the student teachers' craft practices. The data were analysed using grounded theory to gain a deep understanding of how student craft teachers conceptualise sustainable craft. The data were quantified and statistically assessed for dependencies between variables and transferability of results.\nFindings: The study revealed that sustainable craft is conceptualised as a system and that student teachers approach sustainability from different orientations: practice, product, immaterial and holistic.\nOriginality/value: The emerging theory offers a new concrete tool for understanding the abstract concept of sustainability in higher education and suggests that sustainability can be addressed through tangible methods of craft. This theory proposes craft as a tool to conceptualise of sustainability for broader use in education for sustainability (ESD) and as a concrete tool for developing pedagogy for ESD.\n",
"role": "assistant"
}
] |
Information search for home, medical, and financial services by mature consumers
|
[
"Older consumers",
"Services",
"Decision making",
"Information retrieval",
"Consumer behaviour"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\nThe \"mature\" consumer: For purposes of our research, we consider a mature consumer to be one who is 55 years of age or older. Today, individuals over the age of 55 make up almost one-fifth of the US population (Moschis et al., 2004; US Census Bureau, 2002). The aging consumer population is expected to double in size by 2030, with one in three citizens falling into this consumer age segment by 2050 (Kennett et al., 1995; US Census Bureau, 2002). In addition to the shear size of this segment, mature consumers also control much of the wealth in the USA. Estimates over the past two decades report that over 70 percent of all US assets are controlled by those aged 55 and older; and these individuals control over 50 percent of the discretionary income spent by US consumers (Age Wave, 2006; Linden, 1986). In fact, consumers over the age of 55 tend to have twice the discretionary income of their children and have higher per capita income than members of any other age group (Leventhal, 1990; Polyak, 2000). The first \"baby-boomers\" (US citizens born between 1946 and 1964) turned 55 in 2001 and some companies estimate that the current annual spending power of members of the baby-boom generation approaches $2.1 trillion (Senior Journal, 2005). As the remaining 77 million baby-boomers enter this age demographic, the total spending power of the mature consumer segment will continue to grow.Brock (1998), citing data from the Center for Mature Consumer Studies, reports that mature consumers spend three times the national average on health care products and services and spend more than younger counterparts on other services, such as travel and leisure. In addition, this segment is likely to be very loyal to their chosen service providers (Brock, 1998). Thus, companies, marketers, psychologists, sociologists, financial analysts, medical professionals, and public policy professionals should try to better serve this segment by improving their understanding of how this group makes decisions for service-related products and by examining the mechanisms that increase the quality of such decisions.\nInformation search: During the decision-making process, following the recognition of a consumer need, consumers gather information about alternative solutions. This information is later used to make comparisons across options and to make the final decision (Bettman, 1979; Howard and Sheth, 1969). This search process may be either internal or external: internal search refers to the consumer's mental reflection on their own knowledge and experience of past events (Bettman, 1979; Biehal, 1983), whereas external search refers to the consumer's ability to gather information about the consumer problem from the external environment (Bettman, 1979; Heaney and Goldsmith, 1999). When consumers rely on internal search to solve a consumer need, they may recall their own past experience(s) with a similar consumer need, call to mind relevant advertising information and/or remember another person's experiences. External search requires consumers to use other people for information or to find information from non-personal sources, such as advertisements or internet information.Information search for service-related purchases\nInformation sources used in decision making: Sources of external information search can be divided into personal sources (e.g., family, friends, co-workers, and experts), non-personal sources (e.g., advertisements, media, Yellow Pages, and disinterested third parties), and physical search (e.g., information about product choices noticed at random, at drive-by locations, or through direct contact with a service provider) (Beatty and Smith, 1987; Freiden and Goldsmith, 1989).Information sources used for service-related decisions\nRelationship between information search and other decision variables: Two additional constructs are important to explore within the context of older consumers' service decisions: responsibility for the decision and satisfaction with the outcome of a decision. It is possible that a consumer's level of responsibility for making decisions may vary with age and also may vary with the type of product being purchased (e.g., tangible good versus service products). Very little research in marketing has explored this relationship, but there is evidence from the organizational buying literature that finds as responsibility (i.e. accountability to superiors or peers) for the decision increases, the organizational buyer searches more extensively for information (Doney and Armstrong, 1996).Another question of interest is how satisfaction with the outcome of a service decision is affected by the information search process? Punj and Staelin (1983) found that satisfaction with an automobile purchase was not related to external search behavior. However, Crossley and Highhouse (2005) found that the more a person searched for information relevant to job opportunities, the higher his/her later job satisfaction. The relationship between information search and service satisfaction for mature consumers has not been reported.\nCurrent study: In this study, we examine the mature consumer's service choices in three service categories: medical, financial, and home. Selecting a doctor or pharmacy, choosing or changing medical insurance providers, or choosing home nursing providers were classified as medical service decisions. Choosing non-medical insurance providers, hiring estate planners, choosing investment advisors, changing banks, and obtaining new credit cards or loans were classified as financial service decisions. Hiring a home repair service provider, hiring a plumber, choosing a pest exterminator or roofer, finding a cleaning service, and choosing a landscaper or lawn care specialist were identified as home service decisions. Primary research questions of interest were:1. How many information sources were consulted and how many service providers were considered by mature consumers prior to making a service decision?2. How is age related to information search for service products by mature consumers?3. Is the search process different for mature consumers across different service contexts?4. Does personal responsibility for the mature consumer's service decision affect search?5. Is the mature consumer's satisfaction with a service decision contingent on information search and consideration of a variety of providers?\nMethodology: Participants\nResults: Demographics\nDiscussion: Limitations of the study\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to examine the information search behaviors of mature consumers (age 55 and older) for new service purchases across several contexts.\nDesign/methodology/approach: - Almost 200 mature consumers in the Midwestern USA were surveyed regarding new home, medical, or financial service purchases made within the past 12 months. Questions regarding information search, importance of different information sources, satisfaction with the service decision, and demographics were collected.\nFindings: - Mature consumers consulted few information sources for service decisions, irrespective of context. Service context did affect the number of service providers considered; significantly fewer providers were considered for medical and financial services than for home services. Information search was negatively related to age, but unrelated to service outcome satisfaction and responsibility for the decision.\nResearch limitations/implications: - No comparison was made between the search behavior of older and younger consumers. Although the breadth of information search was examined, the depth of search within each type of information source was not. As a descriptive study, it was not possible to address why such little external search was conducted. A convenience sample was used.\nPractical implications: - Mature consumers conducted more external search for less-involved service contexts (e.g. home services) than for those more-involved (e.g. medical and financial services). Companies are beginning to realize the value of maintaining relationships with mature consumers; the results may suggest ways to improve those relationships with mature consumers by detailing the service decision process. Additionally, public policy interventions or education programs based on the work might help mature consumers to learn better service decision strategies.\nOriginality/value: - A large range of ages was sampled in the study (respondents ranged from 55 to 93) and multiple responses were obtained from some participants across several service contexts, allowing for a within-subjects design of the survey. It was then possible to explore the effect of context on the number of service providers considered and the number of external information sources consulted prior to purchase.\n",
"role": "assistant"
}
] |
Climate change and food security of the Malaysian east coast poor: a path modeling approach
|
[
"Malaysia",
"Bootstrap method",
"Climate changes",
"East coast economic region",
"Household food security",
"Path modelling"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Food security is a burning issue and of great concern for many nations on the face of the earth. At national level, the adequacy of domestic food production or capability to mobilize food on time from international market may be sufficient. However, it does not automatically guarantee sustained food security at household level. The factors determining food security at household level are more crucial than those at national level.\nData, model, and methodology: Data collection\nEstimation of the household food security: a path modeling approach: To determine the influence of climatic and non-climatic general factors on the level of household food security in Malaysia, the study conducted bootstrap procedures for second stage formative measurements path modeling. In the initial model, only six variables have been found insignificant, whereas, both the weight and loading of these indicators are smaller than the critical t-values (1.65) for a two-tailed test, even at 10 percent significance level. These insignificant indicators are locality (X7), spouse doing job (X4), head of household having supplementary job (X5), earning ratio (X6), and household poverty/economic status (X3) under the resilience factors (P\nConclusions and recommendations: The study finds that the constructs of direct and indirect impacts of climate change, resilience factors, and other non-climatic general factors have statistically significant positive impacts on all the four dimensions of household food security. Among the four dimensions of household food security, it is found that the impact of food accessibility (D\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the direct and indirect impacts of climatic changes as well as the effects of non-climatic factors on household food security in Malaysia.\nDesign/methodology/approach: The study is based on primary data collected in the months of July - October, 2012 through a questionnaire survey on 460 low-income households from East Coast Economic Region (ECER) of the country. The samples were selected from E-Kasih poor household database, based on cluster random sampling technique. The statistical estimation was done through two-stage formative path measurement model by using non-parametric bootstrap procedure under Smart PLS.\nFindings: This study finds that household food insecurity ECER - Malaysia is not only related to social and economic factors, but also statistically significantly linked to the direct and indirect impacts of climatic factors. The policy implications of the study suggest that design of food security programs must be integrated with climatic change adaptation programs.\nOriginality/value: The government and other relevant agencies should jointly develop appropriate policies for poverty alleviation, household-level food security, and adaptation with climatic changes in Malaysia.\n",
"role": "assistant"
}
] |
Supply chain value stream mapping: a new tool of operation management
|
[
"Supply chain quality management",
"Supply chain"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The issue of the supply chain has been given much attention in recent decades, both in the academic arena as well as in the business world. It has become so important that organizations have begun to directly manage the supply chain, appointing specific managers to perform this function. New challenges now include a focus on supply to determine the right time and place for product delivery (Chin et al., 2004; Robinson and Malhotra, 2005). International business competition is no longer limited to organizations but now includes supply chains (Li et al., 2006; Kuei et al., 2001).\n2. Lean manufacturing and value stream mapping: The term Lean Manufacturing was developed by Krafcik (1988); Womack et al. (1990) as a management principle different from Western production methods. This management approach has been recognized for its effectiveness in productivity, continuous improvement, product quality and timely delivery to customers (Gjeldum et al., 2011), as well as for its wide potentiality, through a variety of tools, to identify waste within value chains (Cookson et al., 2011). Value Stream Mapping is one of these tools and it is used to visualize a full production system, which helps identify activities that do not add value (Shararah, 2013).\n3. The new concept of supply chain quality management - a possible link: The new concept of supply chain quality management is becoming important because companies rely on supply chains which are dependent on the partners involved in the supply chain to ensure the success of quality initiatives, volume and delivery of products and services (Ramos et al., 2007; Rashid and Aslam, 2012). Foster and Ogden (2008) suggest that the efforts of supply chain and quality management in turn improve each other's performance and that the integration between the two functions may be beneficial for an organization in many ways. The transition to supply chain quality management requires knowledge in two key areas of management: processes (quality, volume and delivery), and the supply chain (Fish, 2011).\n4. Research methodology: The research design of this paper was based on the case study approach (Yin, 2004). This approach is very useful in understanding complex social phenomena. The purpose is thus to include a thorough understanding of a situation and its context, in order to propound possible theories arising from the understanding of the phenomenon (Eisenhardt and Graebner, 2007). For this type of methodology, sample selection is crucial. The continuous aim was to find cases where a wealth of theoretical data could be obtained (Yin, 2004). For our study, two company cases were selected with a consolidated supply chain and stable operating performance (see Table I). Thus, the main criterion for selection was that both supply chains would allow the identification of the variables of quality, volume and delivery for the components that make up the products (upstream) and for the finished products (downstream). Both companies are large with over 300 employees and engage in manufacturing production processes. The research project was conducted from January to June 2014.\n5. Applying SCVSM: As indicated in Section 2, the supply chain is the network of services, material and information flow that link a firm's customer relationship, order fulfillment and supplier relationship processes to those of its suppliers and customers (Krajewski et al., 2008). Considering this definition, managing a supply chain consists of at least three flows o three indicators among the players involved. These flows are the material flow, i.e., the supply of raw materials or inputs for the source company who requests them for processing; the information flow, including the traditional purchasing function to generate the purchase order with information regarding both volume and delivery times of raw materials; and finally, the cash flow, credit or cash movement within the supply chain.\n6. Case studies which apply SCVSM: 6.1. A company that manufactures glass door display coolers\n7. Main findings for applying SCVSM: By observing the SCVSM it was possible to find several disruptions in material and information flow, i.e., in the variables volume and OTD. Thus, by considering the material flow map for the cooler manufacturing company the following was identified: for tier 1 suppliers, specifically regarding the components: labels, thermostats and specification plates, there was a divergence between the volume ordered and volume received, of 50, 99,500 units, respectively (see Figure 2). Only the glass was available in the amount initially ordered. The most critical disruption in the material flow in the cooler company's supply chain was the thermostat, a key component in the production process. It resulted in 2 percent less than requested being delivered.\n8. Final conclusions and managerial implications: The purpose of this research was to describe the implementation of SCVSM in order to visualize the operation of competitive priorities, volume and OTD, of a supply chain. The evidence shows that it is possible to apply the tool to visualize the variables volume and OTD in supply chains. In consequence, this research has made a small empirical contribution to the theoretical field of supply chain quality management.\n",
"role": "user"
},
{
"content": "Purpose: The issue of the supply chain has been given much attention in recent decades, both in the academic arena as well as in the business world. It has become so important that organizations have begun to directly manage the supply chain, appointing specific managers to perform this function. New challenges now include a focus on supply to determine the right time and place for product delivery. Some authors indicate a close relationship between Total Quality Management and the supply chain regarding its philosophical and operational goals and perspectives. Lee (2004) goes even further by indicating that only organizations that devise agile, adaptable and aligned supply chains will be able to continue competing in today's environment. Therefore, observing and measuring the supply chain at all times is critical for today's organizations. The purpose of this paper is to describe the implementation of a tool called Supply Chain Value Stream Mapping (SCVSM) in order to thoroughly understand competitive priorities of volume and delivery (On-time Delivery (OTD)) for any supply chain in organizations.\nDesign/methodology/approach: An exploratory case study was conducted. Two manufacturing company which have been established in Mexico with a robust supply chain for at least 20 years, were selected. The paper chronicles the implementation of SCVSM in both case studies. Three methods were used to gather data: direct observation; documentary analysis and semi-structured interviews.\nFindings: This paper proposes a new tool for operation management called SCVSM. Its development, refinement and applied in both case studies. The evidence shows that it is possible to apply the tool (SCVSM) to visualize the variables volume and OTD in supply chains. In consequence, this research has made a small empirical contribution to the theoretical field of supply chain quality management.\nResearch limitations/implications: This paper has the same limitations as all other case-study-based research, including analysis subjectivity and questionable generalization of findings. There may well also be a theoretical gap in the geographic region because the paper only analyzes two companies established in Mexico.\nPractical implications: Derived to describe both case studies on how to apply the SCVSM tool it may prove to be of value to practitioners and managers involved in the field. Similarly, a section on managerial implications has also been included.\nOriginality/value: The study contributes to the limited existing literature on the Supply Chain Quality Management and subsequently disseminates this information in order to provide impetus, guidance and support toward increasing the development companies in an attempt to move the Mexico manufacturing sector toward world class manufacturing performance.\n",
"role": "assistant"
}
] |
Assessing the field: Students and teachers of writing in high-stakes literacy testing in Australia
|
[
"Writing",
"Assessment",
"Identity",
"English teaching",
"Curriculum English",
"Assessing writing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nOpening reflections setting the scene: Teaching writing Emily\nBackground and context - NAPLAN tests in Australia: Before moving to an analysis of the NAPLAN writing test documents, it is useful to provide some further background about the tests themselves, their role in remediating what has, for four decades, been perceived by governments as a deficit in literacy performance and the associated apparatus of the more recently implemented My School Web site that records and publishes school results - thus ensuring an accountability agenda is foregrounded in Australian education (Corbett, 2008; Dall, 2012; OECD, 2015).\nThe NAPLAN writing test: changing the rules of the game: A survey of media articles relating to the NAPLAN reveals that it is the writing component that is the most contested and controversial. Although the argument we now put forward has not been picked up in the media, we argue that a further controversial aspect to this test is that it effectively champions one approach to writing pedagogy, the genre approach, and does not test (and, therefore, legitimize) approaches to writing that exist outside this paradigm. The focus on genre adopted by the NAPLAN writing test articulates with theories of writing that developed in Australia in the 1980s. At this time, amidst rising concerns about teacher accountability, the profession was engaged with divisive debates regarding how writing should be taught (Doecke and Parr, 2005b). This debate was the result of different allegiances to two distinct approaches to writing in English: process and genre writing. A personal growth view of writing in Australia had been influenced by international theorists such as Dixon (1967) and Moffett (1968), who advocated a student-centered approach to learning, and a heightened valuing of creativity and writing. While process writing emphasized student-centered learning and exploration, the genre theory experts advocated a structural focus on the conventions and textual skills that students could learn. The rise of an Australian genre theory was advocated by linguist Michael Halliday and continued by Christie, Martin, Rothery and Derewianka (Coe and Freedman, 1998). Genre theorists conceptualized six genres that each school student needed to master in their writing: report, explanation, procedure, discussion, recount and narrative (Cope and Kalantzis, 1993, p. 9). Internationally, Scott (2008) has noted that the surge in outcomes-based education and standardized testing that came to typify the 1980s and 1990s did not correlate with process writing models of teaching or assessment, and remains discordant with high-stakes testing to this day. Nonetheless, the genre approach to writing continues to hold political currency in Australia in the form of the NAPLAN, where students are examined for their ability to adhere to structural features and conventions in their writing.\nLimiting the pedagogical field: Turning now to an analysis of the writing test and rubrics, we examine the persuasive and narrative genres, making use of Bourdieu's terminology to shape our analysis. In terms of what has been prescribed for the NAPLAN writing test, from 2008-2010, a narrative writing stimulus was set for students; however, since 2011, students have responded to a persuasive stimulus. This change was justified on the grounds that persuasive writing is more prominent in schools (ACARA, 2012). On one level, one can see how this argument can be justified. O'Neill (2012) argues that persuasive writing has always been important due to its centrality to passing high-stakes tests that offer entry to tertiary study, and the way in which it may be explicitly taught for language features and metalanguage. Further to this, a range of research and scholarship has considered the value of persuasive writing in terms of students developing \"real-world\" literacy skills that enable them to have success in the workplace (Moss et al., 2006; National Commission on Writing, 2003; Norman and Spencer, 2005; Zemelman and Daniels, 2004). In contrast, research concerned with narrative or expressive writing has characterized it as a strongly democratic activity linked to the articulation of personal voice (Andrews, 2008; Norman and Spencer, 2005; Smith and Wrigley, 2012); in a more process-driven sense, personal and creative writing has also been cited as a means for affording students \"an opportunity to clarify their thinking\" (Daisey, 2009, p. 157) and teaching them:\nConclusion: In this article, we have explored the ways in which high-stakes testing impacts on teachers' pedagogy and students' literacy and English experiences. Taking the NAPLAN writing text as an example of a high-stakes testing regime, we conclude that this test is significantly impacting on what it means to be both a student and a teacher of writing. By drawing on the work of Bourdieu, we have argued that the prevalence and authority of the NAPLAN, and the way in which the writing test privileges a particular genre approach to structuring written texts, serves to legitimize a reduced and limited view of writing and marginalize other models of English and writing, and this impacts on teachers' autonomy and professionalism. Under this changing field of writing and English, the ideal teacher is not only one who knows the rules of the NAPLAN game and is able to provide students with the necessary skills to successfully adhere to the rules, but that one who is able to provide students with the capital needed to break the rules in legitimate ways - a somewhat paradoxical need, but evidently one that is required and privileged in an ideal composition.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore the interface between high-stakes testing, disciplinary knowledge and teachers' pedagogy in English. The most prevalent standardized assessment form in the current Australian context is the National Assessment Program for Literacy and Numeracy (NAPLAN) undertaken each year by students in Years 3, 5, 7 and 9 in all Australian States and Territories. Understood in the context of the Global Educational Reform Movement (GERM) (Sahlberg, 2011, pp. 100-101) - the NAPLAN tests serve as a bi-partisan governmental response to a perceived need to improve the quality of teachers and schools in Australia.\nDesign/methodology/approach: The authors draw on the key sociological constructs of Pierre Bourdieu (1995) to analyze the ways in which the writing component of the suite of NAPLAN tests serves to legitimize and idealize particular kinds of writing, writers and teachers of writing.\nFindings: The authors suggest that in the absence of current literacy policy and curriculum instability, this national test shapes the literacy field, influencing the direction of writing practices and pedagogy, and, therefore, subject English itself, in Australian classrooms.\nOriginality/value: This assessment intervention is considered in the context of the history of writing, and addresses accordingly fundamental questions concerning the changing nature of the writing/writerly field, the impact of assessment on teachers' conceptions of disciplinarity and pedagogical content knowledge and students' experiences of writing and thinking in subject English.\n",
"role": "assistant"
}
] |
The effect of framing on willingness to buy private brands
|
[
"Consumers",
"Brands",
"Individual behaviour",
"Retailing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\nIntroduction: Framing effect\nMethod: Participants\nResults: The research hypothesis was that the willingness to buy private brands would be higher in the negative framing condition relative to the positive framing condition. That is, the mean of the participants' answers on the above nine-point scale is expected to be higher in the negative framing condition relative to the positive framing condition for all four products.Participants who did not mark their preferences for any of the four products were removed from the analysis. A two-way ANOVA was performed for framing (positive versus negative) X sample (students versus non-students) on each of the preferences of the four products: cream shower soap, floor shampoo, olive oil and pasta. The effects were statistically significant for the framing variable for all four products, F(1,478)=5.60, 4.93, 5.41 and 11.61 (respectively), all p's<0.05; the effect was statistically significance for the sample variable only for the olive oil product, F(1,478)=4.56, p<0.05, and not for the other three products, F(1,478)=1.41, 1.35, and 1.53 (respectively), all p's>0.05; the interactions between the framing and sample were not statistically significance for all four products, F(1,478)=0.82, 1.29, 0.39 and 0.98 (respectively), all p's>0.05.The lack of an interaction effect indicates that the effect of framing was similar for both samples. This finding, along with the lack of significant difference between the two samples regarding three out of the four products, led to a decision to examine the framing effect for the two samples collapsed. Table I presents the descriptive statistics of the participants' preferences for all four products according to the framing conditions (positive and negative). Table I shows that all four private brands were preferred more in the negative (lose) condition relative to the positive (gain/save) condition. The effect sizes (Cohen's d; Cohen, 1988) of framing were about one-fifth for the first three products and about one-third for the fourth one. All effects were statistically significant, t(480)=2.36, 2.22, 2.29, and 3.37, all p's<0.05. Additional examination found no consistent gender differences regarding the four products.\nDiscussion: As predicted, participants displayed stronger preferences for the private brand products when the situation was framed in a negative manner (i.e. lose money if they are not purchased), relative to the very same situation framed in a positive manner (gain or save money if they are purchased). This effect was found for all four products used in this study - two food products and two non-food products. These results are consistent with our prediction derived from implementing goal framing into the context of willingness to buy private brands. The theoretical explanation can be formulated in terms of the prospect theory (Kahneman and Tversky, 1979): When given a choice to buy either a private brand product or a national brand, people contemplate the price difference differently when they are presented with the same prices in reversed approaches. When presented with the gain associated with the purchase of a private brand, the subjective value associated with it is lower than the one associated with the loss of the very same amount that corresponds to not buying it.The more persuasive effect of the negative message is consistent with previous findings of goal framing effect in other contexts, such as breast self-examination (Meyerowitz and Chaiken, 1987), mammography screening (Banks et al., 1995) and credit card use (Ganzach and Karsahi, 1995). In all these examples (as well as many others, see detailed review in Levin et al., 1998), the persuasive effect of the negative framing describing the negative consequences of not taking an action was higher than the positive framing presenting the very same objective situation in positive terms of the gain associated with taking the action.Managerial implications\n",
"role": "user"
},
{
"content": "Purpose: - This research seeks to demonstrate the effect of framing on consumers' willingness to buy private brands.\nDesign/methodology/approach: - Using an experimental design, 500 participants answered a questionnaire addressed to examine their willingness to buy four private brand products framed either in positive (save) or negative (lose) terms.\nFindings: - Consumers' willingness to buy private brands was higher when the information was presented in negative (loss) framing relative to presenting the very same information in positive (save) framing.\nResearch limitations/implications: - This research examined the effect of framing on willingness to buy private brands. Future research is needed to examine the effect of framing on actual purchasing behavior and regarding other products retailers wish to promote, other than private brands.\nPractical implications: - Retailers typically try and promote sales of private brands using messages suggesting that consumers will save money by buying their brand. The results of this study suggest that retailers could improve their marketing of private brands by framing their products negatively in terms of the loss consumers incur when not purchasing private brands.\nOriginality/value: - The paper suggests adopting Levin et al.'s typology that distinguished between various types of framing. The typology explains apparently contradictory results found in the literature. While goal framing more accurately describes the more effective messages discussed in this study, attribute framing would better describe the more effective messages mentioned in other contexts.\n",
"role": "assistant"
}
] |
The Orkney based craft entrepreneur: remote yet global?
|
[
"Entrepreneurialism",
"Globalization",
"Craft production",
"United Kingdom"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Since 1971 and the publication of the Committee of Inquiry Report on Small Firms, commonly known as the Bolton Report (1971), small and medium sized enterprises (SMEs) have attracted a great deal of attention from those interested in the development of the UK economy. The Bolton Report helped to establish the important contribution, actual and potential, that smaller enterprises could make to the economy. The OECD estimates that 95 per cent of enterprises within the OECD are SMEs, which account for 60-70 per cent of jobs in most member states. They are the source of most new jobs and make significant contributions to innovation and employment, often based on high-technology (OECD, 2000). On the negative side there is tremendous \"churn\" in SME activity with less than 50 per cent of start-ups surviving for more than 5 years and only a very small percentage develop into high-performance enterprises.Within the European Union it is estimated that 19 million SMEs exist (using a definition of enterprises employing 250 employees or less) and these represent 99.8 per cent of all enterprises. These enterprises account for 70 million jobs, which is approximately two-thirds of all employment in the EU. SMEs account for at least half of OECD manufacturing employment and on average firms with fewer than 50 employees account for 20 per cent of OECD manufacturing output.In the UK there are approximately 3.75 million enterprises of which 99.1 per cent employ fewer than 50 people (Small Business Survey, 2003). Most of the growth in the business population has been in the number of micro enterprises employing fewer than ten people and in the number of one-person businesses.In Scotland the statistics reflect the trends at UK and European levels. In 1995 micro firms (1-9) and small firms (10-49) each accounted for 18 per cent of employment. From 1995 to 1999, 45,000 jobs were created in micro enterprises but some 93,000 were lost so overall employment in micro firms shrank, with only large firms showing a net gain in employment. In comparison with the UK the fertility rate (birth rate of new enterprises) was slightly above average for micro firms and the proportion of micro firms achieving growth was the same (19 per cent) but Scottish firms grew on average by a smaller extent, perhaps limited by market size (Scottish Economic Report, 2001).Within this context the crafts sector largely sits within the smaller end of SME and often, more correctly, the many one or two person enterprises should be thought of as micro enterprises. However, within this sector there is the same possibility for business formation, growth, development, decline and closure as seen in their larger cousins in the broader SMEs sector. To a large extent enterprises in the crafts sector face the same problems and opportunities as other SMEs but to date have been essentially unseen as a \"group\" although they may have received support on an individual basis within enterprise networks.The crafts as part of the creative and cultural industries\nMethodology: No single study is going to achieve overnight a complete picture of the sector, but it can provide a starting point and establish the most comprehensive picture yet of the sector. Thus, the essential aim of this study was to evaluate the Scottish craft sector. To this end four key questions were formulated:1. What definition of the crafts sector should be adopted as a basis for strategic development?2. What is the current scale and nature of the sector in Scotland?3. What sources of finance and business support are being used by the sector and are there any identifiable gaps?4. Are there initiatives specific to crafts enterprises which can be regarded as exemplifying best practice?The methodology of the project combined both quantitative and qualitative approaches using a postal questionnaire survey, personal interviews and case studies. There were four main parts to the project: establishment of a database; a postal questionnaire survey; follow-up interviews with a sample of craftspeople and case studies of selected areas which were Orkney, Republic of Ireland and the Scottish Borders.The case study areas were selected because it was believed that they would offer examples of good practice and could inform the future development of the sector more generally in Scotland. Orkney had been identified in previous work by the authors and from discussions with Made in Scotland as an area with a thriving craft industry, and therefore was deemed worthy of inclusion. The Republic of Ireland was perceived to be an economy where the craft sector was recognised as a major contributor to the cultural economy. The borders, as an area in mainland Scotland, was seen as a suitable balance to the study of Orkney.In each of the case study areas key individuals involved in economic development in general and with the craft sector specifically were interviewed. The interviews consisted of open-ended discussions focused on the activities of the individual and their organisation in relation to the sector. In addition, craftspeople were interviewed to provide another perspective on the local situation. All interviews lasted between one and two hours and were conducted at the interviewees' office or craft studio. Two interviewers were present each time and notes were taken which were then written-up as soon as possible after the interview. It was agreed with each interviewee that they could be approached after the interview to clarify any points of detail and to agree the sense of the reported discussion.This paper focuses on the case of Orkney and the initial lessons derived from the study (the results from the main study are summarised in Scottish Arts Council, 2002). It is deliberately presented as a case study of the Orkney story ignoring for now small business, economic development and other potentially relevant literature. The narrative is derived from interviews with informants involved in economic development, craft associations and business groupings including Orkney Islands Council, Orkney Enterprise, the Craft Industries Association and a sample of five respondents from the questionnaire survey. In addition a number of workshops and retail outlets around the island were visited to assist in building up a picture of the sector. It is presented as a narrative with the activities of the main players in the Orkney story being highlighted. Following the narrative an attempt is made to make sense or synthesise the story into a template, which reflects the story in relation to the potential for economic development within the craft sector. This template was presented to the key players and their reflections were incorporated into the final version of Figure 1.\nOrkney case study: Background\nAnalysis and discussion: The crafts in Orkney have benefited from being a unique island community where the agencies are working together to support the sector. In understanding Orkney it is important not to underestimate the importance of good design and a reputation for quality which is now associated with products from Orkney. For some, this design is based on traditional roots drawn from the history of the island, while for others it is a process of evolving and re-interpreting historical influences in a contemporary way. In-migration has also enriched and diversified the local economy. Initiatives such as the Orkney Marketing Scheme have provided simple tools including leaflets while also assisting with product development.However, there is some concern about the dependency of the sector on public funding. In addition there is a perception that it is difficult to identify the potential growers, for example, those businesses on the cusp of a break through to the next level of development, be it the shift from part time to full time, expansion to employ staff or achieving the first export order.In terms of this study what emerges from the analysis of the craft sector in Orkney is summarised in Figure 1. At the base of the craft sector are the kinds of enterprises typified by the Craftsman Guild. These are the basic entry-level enterprises, which will mostly be part time or even hobby ventures. The general quality of the work will be very variable from the innovative to mundane and poor. Within the sector there will exist some enterprises, which are capable of making it to the next stage of development. In order to facilitate this transition these individuals need basic business skills training, for example, simple book keeping, and the opportunity to participate in local events in order to build confidence and their experience.At the next level are the members of the Crafts Industries Association which currently has 24 members. In this group the quality and design component of the work is much more consistent and of higher quality. The Craft Trail allows the group to showcase the best producers and to enhance the general reputation of Orkney as a centre for the crafts. Within this group some are still part-timers and some have developed to full-time workers, for example, Sheila Fleet and Aurora. Between these two groups is a third group, currently estimated to be six of the 24 members who are in a \"Zone of Transition\" between part-time and a full-time enterprise. For that zone to be crossed more sophisticated input is needed of business skills including marketing, training and employing staff.Beyond this group are the established \"big players\" which in the case of Orkney is dominated by two jewellery companies, namely, Ortak and Ola Gorie. These enterprises operate in international markets and are well-established in their markets. They fly the flag and are recognised as representatives of Orkney and Scotland. In part they base their success on their Scottish environment but they are evolving into companies whose reputation is based more on the strength of their contemporary design skills. This is where these particular companies may develop in the future in that they will have Scottish roots but will achieve the greater part of their success in the future on the fact that they are British and modern, embracing all that is good in contemporary European and international design. It is in these companies that the remote yet global idea is most evident. Their existence and future development has the potential to offer inspiration to other entrepreneurs and, if managed properly, can provide sustainable growth for the sector.In this model of the Orkney sector it is easy to see stages of development and the need for the input of different types of support at various stages. The agencies which already exist can help to provide this input either on a face-to-face basis or via material to support self-learning either through traditional hard-copy distance learning or via CD-ROM or web-based material. This is certainly true of the generic help which some enterprises need at the level of the Craftsman's Guild. At the second level the Craft Industries Association could help to identify the enterprises at the three different sub-stages within this level whose needs could then be addressed by the Small Business Gateway scheme. At this level more face-to-face contact is perhaps appropriate for delivering specific support to these growing enterprises. At the top end of the sector the larger players are possibly a source of information and guidance to younger firms as well as perhaps needing more specific advice, for example, on exporting via Scottish Development International.Clearly, there are many issues highlighted in the discussions with the interviewees in Orkney that have a direct bearing on marketing activities which need to be explored further in other areas of the research project. In particular the development of business skills, including marketing, through education courses during the start-up period and beyond; the development of the domestic market in Scotland and the rest of the UK by marketing aimed at the individual consumer and the corporate buyer; the development of the international market for traditional and contemporary Scottish crafts by building on Scotland's image and growing reputation for quality in this sector; the development of a communications strategy to promote the sector internally to craftspeople themselves; the enhancement of quality exhibition/trade fair activities to promote the sector to the business market.\nConclusion: In conclusion, it seems there are initiatives and structures in Orkney which work and could provide a template for the support of the craft sector elsewhere in Scotland. The situation in Orkney helps to show what can be achieved when public, private and independent agencies work together. The structure in Figure 1 is the result of evolution over time but the principles behind the structure could be applied to Scotland as a whole.There are interesting parallels with the approach as represented in the template and the frameworks that Scottish Enterprise already operates under its Small Business Gateway scheme. While this needs further exploration, the template in Figure 1 does have an intuitive appeal to the policy makers within economic development organisations. One of the key attractions of the levels approach to the policy makers is that it would allow the business owners to fit themselves to an appropriate level and plug into suitable assistance. Better targeting means better utilisation of scarce resources and, hopefully, a greater economic impact on the development of the craft sector. However, the complexities of the relationships within the template need further reflection and articulation especially if the \"remote yet global\" hypothesis is to be supported.Further insights could be achieved from an analysis of the development of the crafts sector in the other case study areas - Scottish Borders and the Republic of Ireland. Of even more interest would be an update to the Orkney story 5 years after the original visit to see how initiatives such as the Craft Trail have evolved as well as how the public-private sector co-operation has developed. At the same time the influence of Orkney College and the training of young designers is another potential research avenue extending further our understanding of the vertical integration of the sector. With so many research possibilities, Orkney provides a microcosm for the study of many expressions of entrepreneurial and small business phenomena.\n",
"role": "user"
},
{
"content": "Purpose: - This paper reports a study of the craft sector in a remote area of Scotland. The objective is to understand what has made the sector in Orkney relatively successful, and whether elements of best practice can be transferred to other parts of Scotland and the UK.\nDesign/methodology/approach: - The analysis and discussion is based upon face-to-face interviews with public and private payers in the sector and is supplemented by discussions with craftspeople conducted as part of the wider study of which the Orkney story is but one part.\nFindings: - A vibrant and innovative sector, parts of which are global, is revealed as having existed in Orkney for some time. There are, inevitably, some stresses and strains between the main players and different directions in which policy could develop. However, a framework based on businesses at different levels of experience is outlined as the basis for targeted economic intervention in the future.\nResearch limitations/implications: - Further understanding of the craft sector in Orkney could be achieved by conducting a longitudinal study or by exploring specific aspects of the sector, e.g. by looking at the contribution of Orkney College to the emergence of new designers.\nPractical implications: - If the template, outlined in the paper, were to be implemented then business support for the sector would potentially be improved.\nOriginality/value: - The paper draws together elements of the Orkney story in relation to crafts and provides insights to economic development which potentially has lessons to teach the rest of the sector in Scotland and the UK.\n",
"role": "assistant"
}
] |
Multichannel integration quality, online perceived value and online purchase intention: A perspective of land-based retailers
|
[
"Land-based retailer",
"Multichannel integration quality",
"Online perceived value",
"Online purchase intention"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In recent years, e-commerce has been growing steadily on the strength of rapidly evolving consumer shopping habits and game-changing retail technology. Impacts on the traditional bricks-and-mortar retailers are profound and far-reaching. In fact, the dominance of big box stores are declining, replaced by online alternatives offering superior selection, lower prices and even more convenience (Nelson and Leon, 2012). In the severe situation, an increasing number of traditional land-based retailers[1] have migrated to the multichannel format by establishing their own online stores, thus providing customers with multichannel shopping experience. However, many online stores of land-based retailers achieve much smaller sales than pure e-tailers since a lot of shoppers are more likely to purchase through well-known pure e-tailers. The report of Nielsen (2008) also shows that online stores of multichannel retailers continue to underperform on shoppers' purchase intentions compared to online stores of pure internet players. Therefore, traditional land-based retailers need to better understand online consumer behavior based on multichannel environment to compete more effectively against pure internet retailers (Konus et al., 2008; Rigby, 2007).\n2. Literature review and hypotheses: 2.1 Multichannel integration quality\n3. Research methodology: 3.1 Data collection\n4. Data analysis and results: Structural equations modeling (SEM) appears to be the appropriate method for addressing the research question about multiple relationships between the dependent, mediating and independent variables (Zweig and Webster, 2003). Therefore, this study conducted SEM analysis by employing the statistical software Amos 16.0 and maximum likelihood estimation method.\n5. Conclusion: The purpose of this research is to understand the role of multichannel integration quality in enhancing online perceived value and online purchase intention via the online store operated by a land-based retailer. Based on the \"quality-value-purchase\" chain proposed by Zeithaml (1988), we proposed and empirically tested a model of online behavioral intention in the integrated multichannel context, and, by doing so, underscored the importance of the multidimensional concept of multichannel integration quality and the influence of each dimension on online purchase decisions. Overall, our results have shown that the monetary savings, convenience and hedonic values attached to high quality of multichannel integration are important antecedents of online purchase intention.\n6. Implication, limitations and future research: 6.1 Theoretical contributions\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to understand the role of multichannel integration quality in enhancing online perceived value and online purchase intention via the online store operated by a land-based retailer.\nDesign/methodology/approach: This paper develops a research model based on the \"quality-value-purchase\" chain, with four dimensions of multichannel integration quality as antecedents and three dimensions of online perceived value as mediators of online purchase intention. Empirical data were collected from 390 multichannel shoppers and structured equation modeling was used to test the research hypotheses.\nFindings: Among the four multichannel integration quality dimensions, transparency of service configuration, process consistency and business ties positively affect online purchase intention through online perceived value, whereas the effects of information consistency are not significant; process consistency exerts a stronger influence on online perceived value than business ties; the effect of online convenience on online purchase intention is weaker than that of online monetary savings and online hedonic value.\nResearch limitations/implications: The study identifies the theoretical principles of the relationships among multichannel integration quality, online perceived value and online perceived value in multichannel context. Based on these theoretical principles, this study will help researchers to better understand consumers' online purchase intention and the creation of online perceived value in the integrated multichannel context.\nPractical implications: The findings of this study can provide retailers with useful strategies to increase online purchase intention depending on improvement of multichannel integration quality and online perceived value.\nOriginality/value: This study provides a first study to empirically assess various types of online perceived value attached to multidimensional properties of multichannel integration quality and the corresponding effects on online purchase intention. Overall, the results offer insights of how land-based retailers could manage their online performance by integrating multiple channels and improving online perceived value.\n",
"role": "assistant"
}
] |
When migrating stars fail to shine: Individual performance depends largely on the organization as a whole
|
[
"Recruitment",
"Retention",
"Employee development",
"Employee attitudes",
"Performance",
"Motivation",
"Teamwork",
"High flyers",
"Human resource strategies"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Peter Drucker wrote: \"Knowledge workers, unlike manual workers in manufacturing, own the means of production: they carry that knowledge in their heads and can therefore take it with them\" (Drucker, 1998).\nFemales' performance more portable than that of males: Not all of these talent acquisitions were doomed to failure; a lot depended on the similarity of the new firm to the old, the quality of colleague connections and even gender, with females' performance seeming to be more portable than that of males. However, these variables typically affected the length of the performance decline and not the decline itself. Those analysts who switched to firms with greater resources and capabilities and those who were able to move along with other team members (\"lift-outs\") exhibited no significant decline in performance.Perhaps predictably, stars whose performance dropped did not stay at their new firms for long. Within three years, 36 percent of the stock analysts left the hiring firm and a full 65 percent had quit within five years.In addition, more than just the performance of those migrating stars is negatively affected by such moves. When a firm hires a star performer, the performance of the teams that performers are placed in declines as well (Groysberg et al., 2004). Communication breaks down and personal conflicts develop as veteran employees become offended by the extra attention and surplus of resources companies bestow upon their new star. These colleagues often become highly demotivated about their prospects in the organization, feeling that they must look outside the firm if they want to grow professionally or move into leadership positions.It is important to note the limitations of the study. It investigated only one industry and care must be taken not to over-generalize the findings. However, similar observations have been made with surgeons in the operating room and players on the football field (Groysberg et al., 2008b). This research implies that simply acquiring star talent does not yield the outstanding results managers might suspect. Instead, organizations are best served by focusing inside the firm and employing a two-fold approach to talent management: first, keep current stars from leaving; and secondly, establish a talent-development process that creates new stars.In order to keep current stars, firms should increase collaborative work and seek to remove any silos that might inhibit cross-functional performance. Managers ought to consider what company-specific resources enhance employees' performance, such as computer systems or knowledge resources, and seek ways to increase their use. Organizations should also communicate the non-portability of star employees' performance by emphasizing how individual success is dependent on teamwork and organizational resources. If the culture highlights the individual instead of the team, this emphasis must be reversed.\nHow to develop new stars: In order to develop new stars, companies should increase their investment in training, development and on-boarding programs. The investment banks that were able to retain star analysts were typically those with the most extensive development programs in place. These development programs should be tailored to emphasize the individual's use of organizational resources and seek to create company-specific practices.Before looking outside the organization, firms must first consider what roles are actually portable. Managers should question whether the position involves extensive teamwork, buy-in from other colleagues and/or intensive knowledge sharing. Sometimes two jobs that appear similar on the surface might actually have varying degrees of portability. In addition, the portability of individual employees also varies as some people demonstrate a greater willingness to work with others across the organization.When a new hire is made, firms need to minimize the amount of time it takes that individual to get his or her bearings and accelerate the process of integration. Managers might consider allowing prospective hires to be interviewed by their potential teams in order to enhance acceptance.Drucker's observation about knowledge workers may still be accurate. Knowledge workers do not become less intelligent or lose their experience when they switch firms. However, the research makes a strong argument that performance is reliant on more than just the knowledge a knowledge worker has. Performance is highly dependent on organizational context. This dependency makes talent management more complicated than simply getting the right people on the bus. In order to maintain the competitive advantage provided by talented performers, companies are better served by growing their own stars.\n",
"role": "user"
},
{
"content": "Purpose: - Explores the myth of talent portability and offers implications for human resources.\nDesign/methodology/approach: - Summarizes the research on whether star performance can be maintained when performers move to new firms.\nFindings: - Advances the view that performance is largely dependent upon organizational context and, as such, does not transfer to new firms as well as is hoped.\nPractical implications: - Reasons that organizations should focus on developing talent internally while also taking steps to retain that talent.\nSocial implications: - Reveals how organizations - and so, indirectly, society as a whole - can benefit to the maximum from their top talent.\nOriginality/value: - Contends that organizations should abandon recruiting as a primary strategy for acquiring talent. Shows human-resource managers how to leverage this research to develop and retain talent.\n",
"role": "assistant"
}
] |
An empirical investigation on lean supply chain management frameworks in Indian manufacturing industry
|
[
"Framework",
"Indian manufacturing industry",
"Lean supply chain management",
"Validity and reliability"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Lean supply chain management (LSCM) is one of the strategies to bring down the cost of manufacturing products, and improve the quality and availability of the product (Vrijhoef and Koskela, 2000). Vitasek et al. (2005) defined LSCM as \"a set of organizations directly linked by upstream and downstream flows of products, services, information and funds that collaboratively work to reduce cost and waste by efficiently pulling what is needed to meet the needs of individual customers.\" The concept of LSCM was initially identified in the Japanese automobile industry as an integral part of the just-in-time delivery system (Shingo, 1988). The success of the Japanese automobile firms interested the western researchers as they wanted to understand the practices needed to achieve excellence in manufacturing operations. The US manufacturing industry was spending around 30 percent of the total cost of the product on logistics operations (Bigness, 1995). Logistics costs were estimated at $670 million in the USA on non-military products per year. This was almost equal to 11 percent of the gross national product of the USA. These factors generated an interest in the western researchers to try to understand the concept of LSCM that had already been proven in Japanese automotive industry. Western researchers have started to investigate standard and enhanced LSCM principles (Towill, 1992). To obtain productive results, LCSM has to be implemented in many departments of the organization instead of limiting it to one or two departments. It is recommended that implementation should progress through operations management, logistics, transportation and materials/distribution management, purchasing, as well as marketing and information technology to meet organizational goals (Giunipero et al., 2008).\n2. Literature review: According to Lamming (1993), the successful implementation of LSCM is possible with close relationships, open information systems and trust among partners. The sequence of implementation of the LSCM elements also plays a critical role in the success of LSCM philosophy in any organization. The significance of LSCM is increasing in the present global market conditions of competition and complexity. Many organizations have won in the market by delivering the product when the customer anticipates it and with better quality and cost. The LSCM philosophy suggests that companies should partner with customers and work to achieve mutual goals. Toyota motor company has created a similar environment in their organization to achieve excellence in LSCM operations. Before implementing any improvement in their system, Toyota always prefers to obtain feedback from their customers and suppliers about the improvement and also consider the partners' suggestions on how to achieve LSCM excellence (Sezen and Erdogan, 2009). However, compared with the organizations in developed economies, the Indian manufacturing organizations are seriously lagging in the implementation of LSCM practices. Strozniak (2001) has reported that around 20 percent of the Indian organizations practice LSCM principles. Many organizations are not able to implement LSCM principles due to lack of knowledge and improper sequence of activities being performed in the implementation process. To overcome these limitations, Indian manufacturing organizations need to follow a systematic methodology to implement LSCM principles in their respective organizations. To achieve excellence in LSCM implementation, many researchers have proposed various LSCM frameworks in the literature.\n3. Research methodology: A large number of research articles on LSCM principles are available in different publication portals and conferences. The study has collected research articles on LSCM principles from Emerald Online, Science Direct, Springer Link, Taylor & Francis, Sage publication and open access Google search portals. The study has communicated with experts, researchers and professionals to collect developed and practiced LSCM frameworks. The concept of lean production was coined in the book The Machine that Changed the World by Womack et al. (1990). Based on the 1990 citation, the study has selected 1990 as the year from which to collect the articles from various portals. The terminal date for collecting research articles was the end of April 2015. The study has found 864 articles on lean principles from all considered publication portals. Finally, the study has further filtered out the articles based on their focus on only part of LSCM principles. The study has identified around 165 research articles and gathered LSCM frameworks from books and various open source articles from Google search. The study has identified 30 LSCM frameworks from the existing literature.\n4. Results: Validity analysis was conducted to verify whether the existing frameworks could be recommended directly for further study or not. First, content validity was performed by administering structured questionnaire to six academicians and six professionals from industry. The experts were proficient in the field of empirical questionnaire design as well as in their knowledge about LSCM. The feedback received from the experts was incorporated in the final form of the survey questionnaire. Subsequently, the survey questionnaires were also sent to academicians in prestigious institutions and a pilot study was conducted in a premier auto component industry in India. The sample size of the pilot study was restricted to 30 respondents in middle- and top-level management of their organization. The observations and comments of the experts were taken into consideration to develop the survey questionnaire final format. Finally, after content validation, the questionnaire was sent to the targeted respondents in the Indian manufacturing firms.\n5. Conclusions: The present study performed validity and reliability analyses on 30 existing LSCM excellence frameworks in the Indian manufacturing industry. The questionnaire survey methodology was used to collect empirical data from the Indian manufacturing sector. The validity analysis revealed that nine LSCM excellence frameworks displayed unidimensionality with respect to the construct, i.e. the LSCM excellence it measures. The study has further checked the reliability of the selected nine LSCM excellence frameworks. The reliability analysis displayed high values of Cronbach's a, which indicated a high-level reliability. Finally, the study has used frequency distribution analysis to identify the most popular constructs in the final selected LSCM excellence frameworks. Most of the selected frameworks have dissimilar constructs with a certain amount of overlap among them. The study performed in-depth investigation on various key elements presented in the nine validated frameworks. The result shows that various key constructs were not incorporated such as top management commitment, just-in-time production, inventory management and visual management in the proposed LSCM frameworks. The aforementioned constructs played a critical role in achieving the LSCM in any manufacturing organization. LSCM frameworks of only three Indian researchers were included in the present sample of study. The present study has tried to identify suitability of the Indian researchers' frameworks as well; however, it found that none of the frameworks show the unidimensionality characteristic with respect to the constructs, i.e. the LSCM they measure.\n",
"role": "user"
},
{
"content": "Purpose: Supply chain management plays a vital role in deciding the final cost of products. A huge number of non-value-added activities are performed throughout supply chain process. Lean principles are useful to identify and eliminate waste activities across the supply chain management processes. The purpose of this paper is to examine the validity and reliability of the existing lean supply chain management (LSCM) frameworks in the Indian manufacturing industry through questionnaire survey methodology.\nDesign/methodology/approach: The present study collected empirical data from 180 top- and middle-level management personnel from the Indian manufacturing industry. Factor analysis was performed to check unidimensionality of LSCM frameworks by using empirical data. Cronbach's a value of each selected LSCM framework was calculated to find out its reliability. Finally, frequency distribution analysis was performed on the selected framework to identify and reveal critical constructs of LSCM.\nFindings: It was found that nine LSCM frameworks show unidimensionality. The study established that most of the selected frameworks exposed a high level of reliability. The frequency distribution analysis revealed that the larger part of the constructs had a high mean score and mode. It is concluded that there is a need for a new LSCM framework to fulfill the requirements of the Indian manufacturing industry.\nOriginality/value: The present study is focused on checking the suitability of existing LSCM frameworks in the Indian manufacturing sector. It is anticipated that the study will be helpful to the professionals who wish to execute appropriate LSCM framework in the manufacturing organization.\n",
"role": "assistant"
}
] |
Curriculum reform in remote areas: the need for productive leadership
|
[
"Leadership",
"Schools",
"Curricula",
"Curriculum development",
"Australia"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The continual underperformance of indigenous students throughout Australia remains one of the most pressing issues facing education in this country. Many states have tried to redress this imbalance through school reforms and curriculum change. The relationship between leadership practices and school performance has long been acknowledged as an important one. While direct benefits to student learning from school leadership have been hard to ascertain, few would question the need for leadership in the implementation of school, and in the case of this paper, curriculum reform. In this paper, we examine the implementation of the Bound for Success (B4S) curriculum document in schools in the Torres Strait and Cape York regions of Queensland, Australia (Zevenbergen et al., 2008). In remote contexts the importance of leadership is magnified due to distance, inaccessibility, and the difficulty of providing educational services to such regions. We therefore stress the importance of leadership throughout reform processes, particularly in remote contexts. We highlight, through our research, the vastly different experiences and perceptions teachers and principals had during the curriculum policy implementation process. Such differences have been previously acknowledged as being due to the different ideologies and values between teachers and educational administrators (Bishop and Mulford, 1999). However, as we demonstrate in this paper, there was widespread support for the policy itself on behalf of teachers and the principals, and yet the implementation encountered a number of problems. Research concerning school reform has provided numerous examples of the difficulties of policy implementation (Berends et al., 2002; Cuban, 1990; Fullan, 2003; Datnow, 2005; Reynolds et al., 2006). Our paper builds on this research by providing insights into the vast differences in the ways that each of the schools in our study implemented the new policy. These differences illustrate the dilemmas for school reform and ultimately the importance of leadership throughout reform. We consider the model of \"productive leadership\" put forward by Hayes et al. (2004) to be relevant to curriculum and pedagogical change in schools. In doing so, we actively examine the notion of leadership through the work practices of the principals who participated in our study.In the first section of this paper, we briefly introduce the context of the regions in which the schools are situated, as these contexts are highly important in the policy implementation process and need to be understood and incorporated into the change process. Following this we provide a brief background of the emergence of recent policy leading to the B4S curriculum document. The second section draws upon both qualitative and quantitative data consisting of survey responses and structured teacher and principal interviews. In this section we analyse the data through the lens of productive leadership to show the different ways that schools implemented the reforms. These data reveal the significant differences between the views and perceptions of the teachers and principals of the implementation process and a number of concerns from those upon whom the change was being enacted.\nBackground to the \"bound for success\" strategy: The continual underperformance of Indigenous students in Australia, particularly in remote contexts, is an area of serious concern to educators. Recent figures (MCEETYA, 2006) suggest that on a number of key indicators the gap in educational achievements between indigenous students and their non-indigenous peers has actually increased in recent years. In particular, the educational achievements and outcomes for indigenous students in both the Cape York region and the Torres Strait are characterised by much lower levels of enrolment and attendance, literacy and numeracy performance, retention into secondary school, and completion of schooling through to Year 12 (Education Queensland, 2005; MACER Report, 2004).The northern tip of Queensland is divided into two main regions - the \"Cape\", which is that region to the north of Cairns and which covers the Cape York Peninsula. The second major region is that of the \"Strait\", a collection of islands in the Torres Strait. These two regions tend to identify as two distinct groups of people who have very different histories, languages, and cultures. Both areas have significant numbers of indigenous students; indeed, per capita, the numbers of indigenous students is very high in comparison to the national average. Both these regions incorporate challenging terrain combined with a sparse population and these two factors make the delivery of educational services difficult, particularly in the wet season. Poverty and disadvantage are widespread through these areas with limited employment opportunities, high welfare dependency, poor health, high rates of incarceration, and low levels of literacy and numeracy (Fitzgerald, 2001).Over the last ten years the policy terrain in Queensland for addressing the educational concerns of Aboriginal and Torres Strait Islander students has been an active one. There has been recognition that the continuing poor educational outcomes of indigenous students need to be addressed. The B4S strategy was introduced in 2005 to target the poor outcomes of Indigenous students in remote regions of Queensland in the Cape York and Torres Strait Islands. Strategies specific to Cape York and the Torres Strait were developed in order to facilitate the appropriate delivery of services to these regions. This approach was in part due to the nature of the challenges facing the Cape York and Torres Strait regions and their indigenous communities. In 2007, Education Queensland began the implementation of the Bound for Success Scope and Sequence Statements (SSS) curriculum document in its Cape York and Torres Strait Island schools. The SSS curriculum document incorporates each of the seven Key Learning Areas for Years 1-10. In addition, the department also implemented an online toolkit aimed at increasing the effectiveness of monitoring students' progress and the development of individual learning plans. The Individual Learning Plan (ILP) toolkit and the Scope and Sequence Statements are significant aspects of the department's B4S strategy to target the outcomes of indigenous students in the Cape York and Torres Strait Islands. Within this context, our study (Zevenbergen et al., 2008) sought to identify and evaluate the implementation process and uptake of a policy reform that sought to bring about high quality curriculum in an environment characterised by remoteness and isolation.\nProductive leadership as a framework for analysis: The difficulties with and complexities of implementing reforms in schools have been documented for a number of years (McLaughlin, 1987; Odden, 1991; Fullan, 1993, 2001; Cooper et al., 2004). Recently, studies have attempted to examine policy implementation through sense-making (Spillane et al., 2002), micro processes (Coburn, 2006), and the enacting of professional development (Penuel et al., 2007). However, less understood and examined through empirical research are leadership practices of principals throughout the reform process. The leadership of the principal, in particular, is a key aspect of successful policy implementation (Sammons, 1999; Leithwood et al., 1999; Fullan, 2001), and especially so in remote contexts where support through additional personnel, communication, and face-to-face professional development is difficult. The leadership practices of principals thus become critical during the implementation of policy, and researchers (Taylor et al., 1997) have argued that the practices of implementation need to be viewed as constitutive of the policy itself. That is, \"policy is both text and action, words and deeds, it is what is enacted as well as what is intended\" (Ball, 1994, p. 10).In recent years, various models of leadership have been put forward. These models draw on the school improvement and effectiveness literature and aim to promote \"effective\" leadership. For example, styles include educative leadership (Grace, 1995; Webb, 2005), instructional leadership (Hopkins, 2003; Southworth, 2003), and pedagogical leadership (Sergiovanni, 1998). However, the view that the notion of leadership is both ubiquitous (Gronn, 2003) and inherently elusive (Leithwood et al., 1999) is important to note. In this paper, we work with the idea that leadership should be seen as being \"a set of relations practised by many teachers, principals and parents in a range of educational sites and in a number of informal as well as formal administrative positions\" (Blackmore, 1999, p. 6). While the direct effects of leadership on student learning have been shown to be limited (Hallinger and Heck, 1996), we believe that educational leadership has a significant part to play not only in fostering an environment in which sound teaching and learning exists, but also in facilitating the implementation of school reforms. Traditionally, there has been a plethora of literature, often from the corporate and business environment, that has sought to identify \"best practice approaches\" to leadership and that has often led to normative frameworks such as transformational and transactional leadership. Such leadership by design approaches (Gronn, 2003) are often descriptive and prescriptive lists of competencies and capabilities. We believe that adherence to the notion of productive leadership can be useful as it acknowledges the teacher as the most influential element in enhancing outcomes for students and therefore suggests that the work of those in formal leadership positions is to encourage sound classroom practices while mediating policy effects (Lingard et al., 2003).Hayes et al. (2004) describe their notion of productive leadership as consisting of the following characteristics:* A commitment to leadership dispersal that supports the spread of leadership practices and collaborative decision-making processes in building common vision and purposes.* Supportive social relationships within the school, between staff (teachers and others) and students.* \"Hands on knowledge\" about how educational theory translates into strategic action and is aligned with community concerns and relationships outside of the school.* A focus on pedagogy in which leadership in a school is focused on improving students learning outcomes and learning within the school as a whole.* Support for the development of a culture of care which encourages teacher professional risk taking.* A focus on structures and strategies in which leadership focuses on developing organisational processes that facilitate the smooth running of the school (Hayes et al., 2004, p. 524).While each of these dimensions in isolation does not present anything radically new to the field, as a whole, the productive leadership approach demonstrates support for the development of schools as communities focused on learning by emphasising professional learning communities and linking this with the notion of productive pedagogies (Hayes et al., 2004). The productive pedagogies framework evolved out of the Queensland School Longitudinal Reform Study (QSRLS) (Lingard et al., 2001), which outlined approaches to teaching that were based on the notions of intellectual quality, a supportive classroom environment, connectedness to the world, and valuing difference. In working toward this framework we concur with the work of Lingard et al. (2003) and Lambert (2000), who argue that leadership dispersal through the whole school community is important for the promotion of productive leadership. We therefore use the dimensions of productive leadership to analyse the interview data for the purpose of highlighting the point that a closer alignment between teacher and principal responses existed where aspects of productive leadership were said to be demonstrated.\nMethod: Data were collected during the final weeks of the school year. Timing was an important consideration since it was necessary to work with many teachers and administrative staff involved in the implementation of the B4S curriculum document before the staff completed the school year. In the region under investigation there is a considerable turnover of staff with as many as 50 per cent of staff changing schools at the end of a given year. Thus, the data collection was undertaken before the completion of the school year so that those staff involved in the implementation would be canvassed before they left the region. Data were collected via an online survey, face-to-face interviews, and telephone interviews. The selection of participants and the development of particular questioning protocols were undertaken in partnership with the Queensland Education Department.Online survey\nFindings: Initial response to the reforms\nConclusion: Overall, the interview data demonstrate a wide variety of approaches to implementing the B4S curriculum. These ranged from authoritative styles of leadership, or just \"do it\" approaches, to more collaborative whole school approaches. The over abundance of the former rather than the latter is cause for concern; the vast discrepancies between principals' and teachers' opinions and perceptions demonstrated within schools characterised by authoritative leadership can only be detrimental to the learning opportunities and outcomes of students. The qualitative data we present highlight the range of different \"leadership\" practices exhibited by principals and teachers throughout the reform process. Despite the inherent limitations of relying on a data set characterised by teachers' and principals' subjective perceptions of a reform process, the data indicate that processes that fostered a learning community with a dispersal of leadership to the curriculum leader in the school allowed for a closer of alignment of teacher and principal beliefs. The schools in which the school principal or administration staff told their teachers to just implement the new curriculum displayed a style of leadership that exhibited none of the aspects of productive leadership. It should come as little surprise that these particular schools were known to the state and regional education bodies to be under-performing and had very high levels of staff turnover. The principals of these schools demonstrated a clear hierarchical leadership focus with very little emphasis on student learning, pedagogy, and the creation of a supportive environment in which teachers could collaborate and discuss issues of concern with the principal.In contrast was the school in which productive leadership was present. The school principal actively encouraged staff to assume leadership responsibilities and organise teacher discussions around the curriculum document, pedagogy, and assessment. In other words, the principal supported a collaborative culture and collective responsibility for the implementation of the reforms. The alignment between the teachers' and principal's responses from this school indicates that these strategies were indeed taking place, whereas the vast mismatch in teachers' and principals' perceptions of the change process in the other schools indicates a lack of cohesion in the school support structures and the communication between administration and teaching staff. The comments from the teachers in these schools are alarming because they demonstrate that these teachers have clearly not been supported in the implementation of the new curriculum; if indeed the individual teacher has the most significant impact upon student learning (Lingard et al., 2001), particularly for students from minority backgrounds (Hayes et al., 2006), then it is highly doubtful that these schools can actively contribute to an improvement in their students' educational achievement. At a time when indigenous students' educational outcomes are still significantly behind those of their non-indigenous peers (MCEETYA, 2006), there is a serious need for schools to actively focus on leadership practices that are focused on enhancing student social and academic learning, otherwise they are contributing to the perpetuation of practices that are disadvantaging indigenous students.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to report on research into the challenges for leadership in implementation of a new curriculum in a remote region of Queensland, Australia.\nDesign/methodology/approach: - Data for the research were gathered through an online survey and semi-structured interviews with teachers and principals involved in the reforms.\nFindings: - Results show that there were substantial differences in the views of teachers and principals in their perceptions of the implementation process. The vast differences in the implementation strategies and leadership approaches between the schools suggested that the effect of the reform on leadership practices was more positive in schools in which elements of productive leadership were present.\nOriginality/value: - This research highlights the importance of leadership throughout reform processes, particularly in terms of the different experiences and perceptions teachers and principals have during the policy implementation process. As a result, this paper offers an important perspective on the complex dynamics of leadership processes and how curriculum reform is enacted at the local level.\n",
"role": "assistant"
}
] |
The nature of crowdfunding in China: initial evidence
|
[
"China",
"Entrepreneurship",
"Crowdfunding",
"Financing",
"MSME",
"New venture funding"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Micro, small and medium enterprises (MSMEs) are the main contributors to employment worldwide and are considered the backbone of economic activity. However, according to World Bank Research and Policy Report (Abraham and Schmukler, 2017), despite MSMEs significant contributions to the economy, access to financing and funding are relatively limited compared to large enterprises. They are less likely to have formal bank loans or other lines of credit because they tend to be young, informal and operate in unfamiliar business sectors which discourage the conventional financial institution from serving them (Bruton et al., 2014).\n2. A framework for analyzing crowdfunding in China: In reviewing the literature on crowdfunding in China, we develop a framework that ties together common elements of understanding crowdfunding industry based on institutional context or institutional environment by Bruton et al. (2014) which include elements such as the supply of capital, demand for capital, governance considerations and antecedents of this financing alternatives. As stated by Bruton et al. (2014, p. 2):\n3. Chinese crowdfunding market: China's rapid development of internet finance coupled with the prevalence of smartphone usage among the Chinese fuel the growth of crowdfunding. According to 2016 China Business Review, 86 per cent of Chinese internet usage is through cell phones. The vast online population, the diffusion of smartphones, the extensive use of social media and the use of online payment such as Alipay and WeChat Wallet drive the success of the Chinese crowdfunding industry. As of 2018, there were around 1.12 million mobile users in China, 751 million internet users and 911 million active mobile social media users, many of having electronic money (We are Social and Hootsuite, 2018). The World Bank estimated that approximately US$96bn in funding would be raised in developing countries by 2025 through crowdfunding, of which US$50bn would come from within China, especially given the rise of female investors and fundraisers (World Bank, 2013). There may be many factors driving the rapid development of China's crowdfunding market and certainly, the world's largest online population is one of the critical reasons (FSDC, 2016).\n4. Crowdfunding regulations in China: Unlike the rest of Asia, Chinese crowdfunding platforms have grown consistently since 2011. The rapid increase of crowdfunding operations because of the lack of specific regulations for crowdfunding processes enables platforms to begin their operations very quickly. As a result, substantial risks and challenges faced by investors, project initiators and regulators against illegal unlicensed fundraising have occurred. To address some of these issues, from 2015 to the end of September 2016, the central government has issued a series of policy announcements to encourage the development of crowdfunding.\n5. Types and features of crowdfunding in China: The differences between international and Chinese crowdfunding platforms are a manifestation of cultural differences. China's crowdfunding platforms are distinctly sale-oriented or commercial (Legrand, 2016; Xiao, 2015). Marketing campaigns are often run by platforms to drive product promotions and sales which is far from the ideal crowdfunding fostered through Kickstarter or Indiegogo. For example, a Chinese startup product like Mars (Bluetooth speaker) by Crazybaby raised more than US$800,000 on Indiegogo before even refining Mars and preparing it for mass production. The product was only put on the Chinese crowdfunding platform Taobao after having run a successful campaign on the international crowdfunding site Indiegoo and before launching Crazybaby into the domestic Chinese market. More than a thousand units of Mars were claimed by Taobao crowdfunding investors who earned 1277 per cent of the original target, surpassing it by more than RMB200,000 (Xiao, 2015). A similar example is the case of Czur Tech Co., Ltd., a hardware startup based in Shenzhen which produced the ET Smart Book Scanner. Who after successfully attaining funding through Indiegogo successfully launched their product on the JD crowdfunding website.\n6. Crowdfunding participants and the success of projects in China: The crowdfunding business ecosystem enables the demand and supply side of the investment process to interact directly without the intermediary functions that were provided by many financial institutions (Wang et al., 2018). The goal is to collect small amounts from many members of the community in order to realize the project (Egger et al., 2016). This process involves three types of participants: fundraisers as entrepreneurs or capital-seekers who provide the idea/project, crowdfunding platforms as an intermediary and investors as funders or backers.\n7. Antecedents of successful crowdfunding projects in China: A Crowdfunding project is successful if the funding goal has been met or exceeded the target amount within the set period. It is a failure if the funding goal is not met and the investors receive their investments back and the project can therefore not be realized and is scrapped (Babich et al., 2017). Hou et al. (2015) determined the common antecedents of a successful project using a dataset from Demohour. The results show that project duration, goal, initiator participation and experience, the number and source of followers and the number of visitors affect the success of the projects in China. Yip (2014) highlighted four feasible strategies for Chinese Social and Environmental Small and Growing Businesses (SESGBs) to use US-based crowdfunding platforms to gain access to financing:\n8. Conclusion, implications, limitations and further research: 8.1 Conclusions and implications\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to develop initial evidence about the nature and features of crowdfunding in China, given it is largely unregulated regulatory frameworks.\nDesign/methodology/approach: The paper used extensive desk research using data collected from the public and private sectors, after which the data was analyzed parallel to existing academic literature, that is, institutional context by Bruton et al. (2014). This paper uncovered patterns of development, profiling crowdfunding platforms, examining the regulatory landscape and providing antecedents of successful crowdfunding projects in China.\nFindings: When the traditional financial markets are hard to reach, micro, small and medium enterprises (MSMEs) were starved for capital. Crowdfunding can play a major role in funding and risk sharing. It is an innovative and dynamic vehicle for MSMEs as well as enthusiastic investors in China. Since its initial introduction to China in 2009, crowdfunding has gained substantial popularity in a relatively short period. Currently, there is still not an identifiable guideline on how to delineate the significance of the crowdfunding platform. The development of crowdfunding in China faces a few unresolved key issues. As researchers exploring this phenomenon in new ways, crowdfunding platforms can be enhanced in a manner that benefits the capital seeker, investors and society as a whole.\nOriginality/value: There is a dearth of information on start-up crowdfunding in Asia. With little data available to analyze, so this paper hopes to contribute to knowledge and provide valuable information to researchers and industry representations. Crowdfunding represents a potentially disruptive change in the way that new ventures are funded. This paper represents an initial analysis in the study of new ventures in China. Finally, the authors provide recommendations for entrepreneurs, investors and policymakers as well as researchers and practitioners with suggestions about yet unexplored avenues of research.\n",
"role": "assistant"
}
] |
An investigation of factors affecting knowledge sharing amongst UK academics
|
[
"Academic staff",
"Universities",
"Knowledge management",
"Knowledge sharing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: As knowledge-intensive organisations, universities play a critical role in knowledge creation through research and dissemination through publication. In addition to the knowledge possessed by individual lecturers and researchers, universities hold vast knowledge repositories (Rowley, 2000). Universities also hold a key role in transfer of knowledge by working with businesses and other organisations to support learning through teaching and research training programmes and promote innovation and enterprise through initiatives such as knowledge transfer partnerships and enterprise centres. Furthermore, it has been suggested that increased knowledge sharing specifically in higher education can initiate improved decision-making processes that could speed up curriculum development and research. It could also facilitate the retrieval and storage of institution-specific tacit knowledge from key individuals before they retire and assist with the formation of repositories for best practices in assessments and use of technology (Kidwell et al., 2000).\nLiterature review: Knowledge sharing is absolutely critical to the success of knowledge-management initiatives (Davenport et al., 1998; Al-Alawi et al., 2007). This is because knowledge sharing is a voluntary act and for the benefits of knowledge management to be realised, employees need to be persuaded that it is in their interests to share (Hislop, 2013). van den Hooff, B. and de Leeuw van Weenen (2004, p. 118) suggest that knowledge sharing is \"[...] the process where individuals mutually exchange their knowledge and jointly create new knowledge\". However, mutual exchange can certainly be limited by knowledge hoarding due to culture and values (DeLong and Fahey, 2000).\nResearch design: The present design draws substantially on the widely cited (Kankanhalli et al., 2005; Lin, 2007; Chow and Chan, 2008) research into knowledge-sharing behaviours, conducted by Bock et al. (2005). Thus, in terms of individual factors, beliefs, rewards, associations and contribution were based on Bock's items. However, in the light of previous literature on organisational cultures and the unique context of academia (Lee, 2007), it seemed appropriate to capture a different range of variables from those used by Bock et al. (2005). Therefore, this study augments Bock et al.'s (2005) research with items on leadership autonomy, affiliation to institution, affiliation to discipline and technology platform.\nResearch methodology: Research context\nFindings and discussion: Organisational factors\nConclusions and recommendations: One objective of this research was to develop a new measurement model that is appropriate to the context of universities, and this was achieved by an adaptation of Bock et al. 's model (2005). Similarly, organisational and individual factors that affect knowledge sharing in such a context were identified through the literature, a research model was constructed and the impact of such factors on sharing was measured by the use of SEM. As in other studies in private and public organisations, culture was found to have a significant influence on knowledge sharing, but in this particular study, individual beliefs were shown to have a stronger relationship with knowledge sharing. This is in keeping with the high level of autonomy afforded to academics.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to construct and investigate relationships between knowledge-sharing factors, attitude and the intention to share of UK academics, as research on knowledge sharing in higher education is extremely sparse.\nDesign/methodology/approach: A research model and hypotheses were constructed from individual and organisational factors that were identified to affect knowledge sharing. Questionnaire data were obtained from 367 academics concerning their attitude and intention towards knowledge sharing. This was then used in a two-stage structural equation modelling approach where the measurement model was used for confirmatory factor analysis. The structural model was used to measure and test the hypothesised relationships.\nFindings: Findings indicate that, in general, individual beliefs amongst academics were more influential on their knowledge-sharing attitudes than organisational culture. Furthermore, leadership was the most influential factor within the overall organisational culture whereas autonomy demonstrated the weakest relationship. Belief in the possibility of rewards through associations was found to be a highly significant individual factor. The relationship between attitude and intention was relatively weak although still statistically significant.\nOriginality/value: The research demonstrates that management should ensure that departmental leaders promote knowledge sharing and that valued rewards are linked to sharing within the department.\n",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.