title
stringlengths 6
301
| keywords
listlengths 0
23
| messages
listlengths 2
2
|
|---|---|---|
SMEs, environmental management and global warming: a fusion of influencing factors?
|
[
"Environmental management",
"Global warming",
"SMEs",
"Stakeholder theory",
"Internal information systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Firms are institutions and form part of the social community. The decisions made by management influence parties other than just shareholders (Suojanen, 1954). Management decisions result in business activities that utilise natural, human and financial resources. Corporate social and environmental responsibility has now moved into the mainstream of business agendas (Perrini, 2006). The adoption of pollution reduction and environmental sustainability into normal business activities is now accepted, particularly as environmental regulations are gaining pace, for example a carbon tax in Australia[1]. Any regulatory impositions intrude, directly or indirectly into the SME sector, as this sector forms part of any regional response to the reduction of business carbon emissions. Furthermore, some industries have a greater risk of pollution than others, and can be the focal point of regulatory burdens.\nTheoretical support and research propositions: Freeman (1984) suggested a stakeholder to be any individual or group who can affect or can be affected by the activities of the organisation. Firms rely on stakeholders for resources to continue their activities, their growth and ultimate survival. How the resources are used will depend on how the firm incorporates the rules and values of society into its daily activities - its licence to operate Dowling and Pfeiffer, 1975; Fontaine et al., 2006, p. 25). Freeman's (1984, p. 25) study identified the separation of stakeholder groups, the need to consider these groups and their concerns, and their associated influence on the business environment of the firm (Donaldson and Preston, 1995). Stakeholders can influence business decisions and operations (Swift, 2002; Seidel et al., 2009) as SME managers' constantly interact and engage with their individual stakeholders.\nLiterature review: The increasing importance of environmental issues to the business sector has generated research on the responsibility of SME managers and their commitment to environmental management (Gerrans and Hutchinson, 1998; Rutherfoord et al., 2000; Ahmed et al., 2003).\nResearch method and analysis: The main research method adopted by the majority of SME researchers is a survey, both mailed and telephone surveys (Dennis, 2003; Hillary, 2004). Accordingly a mailed survey was conducted and forwarded to a random sample of SME firms operating in Australia, with the self-administered questionnaire addressed to the attention of the SME owner or manager. The original mail out was to 2,024 firms listed in a relevant Dunn and Bradstreet database. A reminder mail out followed. The responses to the two mail-outs were 375. Oppenheim (1966) suggested non-response or a trend analysis that assumes that late respondents are basically non-respondents. Therefore, of the respondents were late in returning their questionnaire and thus were considered to be non-respondents. These, together with four inadequately completed forms, were removed and a usable response of rate of 360 SMEs or 18.4 per cent was determined. Although this response was disappointing, it is recognised that surveys have tendency to lower response rates (Acutt and Geno, 2000) and that low-response rates are prevalent in the small and medium size business sector (Macpherson and Wilson, 2003) Nevertheless, these can reduce the credibility of SME research findings (Dennis, 2003, p. 278; Goela and Bisman, 2002).\nDiscussion: In general the respondents are aware of their stakeholder concerns, and particularly those of their employee and customer stakeholders (Gadenne et al., 2009). The results indicated that these SMEs are aware of the environmental challenge and include stakeholder concerns, particularly employees and customers, when undertaking their business activities. They appear particularly committed to driving employee satisfaction and safety (Howes, 2008), as environmental issues are incorporated into their (OHS) guidelines, and the written procedures for their employees and contractors. However, the results are silent as to whether labour regulations pressured the SMEs to do so.\nLimitations: The results of this study add to both the academic and professional literature in relation to stakeholders and industry groups' fusion with environmental issues in the Australian SME sector. Practitioners and advisors to SMEs can also benefit from these results. However, the data is limited to this geographical location and thus the results cannot be applied to other geographical locations. Furthermore, respondents included 56 per cent from medium businesses that is, between 100 and 200 employees, and thus there is a size discrepancy (Goffee and Scase, 1995) towards the medium firms' business. However, larger firms are more susceptible to stakeholder pressure on environmental issues (Tilley, 1999). The scope of the study did not include the specific requirements of each individual stakeholder in relation to SMEs.\nConclusion: The purpose of this study was to explore environmental issues within SMEs, particularly their perceptions regarding their immediate stakeholder concerns, the incentives prompting their expenditure on environmental issues, the environmental issues within their information system and their view on how global warming may impact their activities. The results indicated that SMEs:\n",
"role": "user"
},
{
"content": "Purpose: Stakeholder theory provided the broad theoretical lens to explore environmental issues in small and medium enterprises (SMEs). The diversity in issues examined ranged from concern for immediate stakeholders, their industry group and the influence of global warming on their business activities, to the type of environmental information included within their internal information systems. The paper aims to discuss these issues.\nDesign/methodology/approach: A mailed survey was used to obtain data from SMEs operating in Australia. The focus was primarily directed to medium size firms.\nFindings: The findings indicate that SMEs were aware that their stakeholders, particularly their employees and customers were concerned with environmental issues. The respondent SMEs were also aware that global warming would influence their activities, for example, the design of their projects, occupational, health and safety, labour contracts and customer relations. Overall, the findings suggest that any tailored approach to regulate or self-regulate environmental management in SMEs, be industry and stakeholder driven.\nResearch limitations/implications: The limitations of this research are primarily those applicable to the survey method, SME response rates, and the geographical location covered by the survey. The focus of this study is primarily medium size firms, rather than micro small business.\nOriginality/value: This study gains insights into some of the practical aspects of environmental management in SMEs and in so doing adds to the growing body of literature in this under researched area.\n",
"role": "assistant"
}
] |
Can digital technologies speed up real estate transactions?
|
[
"Commercial real estate",
"Real Estate transactions",
"PropTech",
"Property passport",
"Blockchain",
"Automated valuation models",
"K25",
"L85",
"O33",
"R33"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Land shelters us, feeds us and provides us with space for recreation, shopping and work. With total global real assets valued at more than $ 238 trillion (Savills, 2017), land as an investment is a significant component of the global economy. At the same time, transacting real estate is a costly and time-intensive process. For example, the registration of a property in the United Kingdom (UK) takes, on average, 22 days, and costs up to 5% of the total property value (WorldBank, 2018).In this paper, we investigate the most important causes for delay in commercial real estate transactions in the UK, and we discuss how digital technologies could make this process faster and cheaper.For this purpose, we conduct a focus group interview with commercial real estate experts, who were asked to construct a process map of the steps required for a hypothetical transaction. More than 150 individual steps and seven major causes for delay were identified. In the second step, we assessed the potential of digital technologies to overcome these blockages, based on desk research and interviews with executives from established real estate firms and PropTech start-up firms.In summary, we could identify a lack of an up-to-date single pool of standardised property information to be the most critical cause for delay. However, technological solutions to effectively provide such information require many complementary investments and legal adjustments, which impede their introduction. In contrast, more specialised technologies focusing on individual inefficiencies in the conveyance process face less entry barriers and are, hence, more likely to be adopted.In the following section, we briefly describe the current conveyancing and registration process in the United Kingdom, before we present the identified causes for delay in Section 3. Section 4 discusses the potential of digital technologies to reduce these blockages and barriers to adoption. Section 5 concludes.\n2. Conveyancing and registration processes: A commercial land transaction broadly consists of five steps involving a host of actors. Sellers and their agents are responsible for (1) the preparation and (2) marketing required to establish an asking price of a property. Solicitors then organise the conveyance--the administrative process by which rights over land are transferred. This step includes all necessary legal works during (3) the due diligence period and (4) the post-exchange period. Finally, a land registry collects taxes and records the legal ownership during the post-completion period.The success of a property transfer can be measured by how closely the conveyancing process conforms to initial expectations, most importantly determined by the time it takes until settlement occurs. In practice, conveyancing processes in the current paper-based system are often delayed, as transaction times run beyond expectations. In the UK, for example, delays occurred in close to 40% of all transactions (NewStatesman, 2018).Improvements in the UK registration system have, so far, been hampered by a lack of digitalised (in other words, machine-readable) data: while the vast majority of service requests to Her Majesty's Land Registry (HMLR) are placed in a digitised format (HMLR, 2017), many of these requests are lodged as scanned PDF documents (digital copies of the original document). These scanned records still rely on manual interpretation to extract relevant information. The efficiency gains of using such digitised documents are, thus, limited.While many PropTech companies (Shaw, 2018) currently offer digitalised platforms which could contribute to more streamlined property transaction processes, many of them are siloed solutions for specialised problems, which do not take problems of software and data integration into account. As a consequence, the real estate industry is slow at adopting these new technologies.In order to make real estate transactions more efficient and to ease the conveyancing burden for firms and individuals, it was suggested that a single system connecting all the agencies that supply information involved in property transfers needs to be established (Abbey et al., 2017, p. 41). Such a single portal system is generally known as \"e-conveyancing\".However, e-conveyancing has so far not been successful in the United Kingdom. HMLR's pilot project Chain Matrix failed, as it suffered from poor software integration, and non-digital procedures slowing down the transaction chain. Similarly, VEYO, an e-conveyancing attempt from the Law Society has not been adopted widely (Solomons, 2016). In contrast, Australia's PEXA system provides a functioning e-conveyancing platform, after being mandated by the Australian government (Deloitte, 2018).\n3. Causes for delay in the current transaction process: Data are fundamental at every stage in a property's lifecycle: when buying an asset, investors need information about building operations, cash flows, maintenance and capital expenses, potential risks, market assumptions, tenancy schedules and more. These data need to be transferred efficiently from and to previous and current owners, legal advisors, property managers and others, but this process is commonly subject to inefficiency and delays.To reveal the information processing steps necessary for a real estate transaction, and to identify the bottlenecks in information transfer, we organised a focus group interview with real estate experts from HMLR, two major UK law firms and a leading real estate advisory firm in August 2018. We asked the participants to construct a process map outlining the individual steps required for a hypothetical transaction. The experts identified 150 individual steps and seven major blockages to efficient commercial transactions that take place in the five phases of a typical real estate transaction.Details on the process mapping exercise are provided in the supplementary materials (Section S1). The process maps itself can be found in the online supplement [1]. 3.1 The preparation period\n4. Applicable technologies and barriers to adoption: Having analysed the causes for delay in commercial real estate transactions, it is clear that removing these bottlenecks would speed up property transactions. Here, we explore the potential of individual technologies to reduce some of these delays. For this assessment, we draw on desk research and a screening of PropTech firms to identify technologies and solutions. Additionally, we identify the operational, regulatory and social barriers which need to be overcome before adoption, based on semi-structured interviews with real estate professionals and PropTech executives we have conducted from September to December 2018.Details on the interviews, including question guides and interview-notes are provided in the supplementary materials (Sections S2, S3, and S4). Section S5 lists start-ups working on relevant technologies. 4.1 Applicable technologies\n5. Discussion: 5.1 Summary\n",
"role": "user"
},
{
"content": "Purpose: This study presents a structured investigation of the most important causes for delay in commercial real estate transactions. It assesses the potential of digital technologies such as \"Blockchain\", \"Property Passports\" or \"Automated Valuation Models\" to make transactions faster and cheaper.\nDesign/methodology/approach: The authors conduct a focus group interview to identify the individual steps and the parties involved in real estate transactions. Subsequently, the authors discuss the prospects of digital technologies based on semi-structured interviews with real estate professionals and PropTech executives, and a comprehensive screening of technological solutions offered by PropTech firms.\nFindings: The lack of an up-to-date, single pool of standardised property information turns out to be the most critical cause for delay in real estate transactions. However, the most promising technologies to mitigate this problem, in particular digital property passports summarising all relevant building information, face substantial barriers to adoption. The real estate industry has so far not been willing to more openly share data, which is a pre-requiste for the successful introduction of property passports. In addition, the principle of caveat emptor makes a lengthy due diligence process essential for buyers.\nPractical implications: The authors conclude that industry-wide collaborations are necessary to help major efficiency gaining technologies to break through. Insurance products should accompany property data log books to guarantee the quality of data provided.\nOriginality/value: This study considers the potential impact of technologies in the wider context of the complete real estate transaction process. It identifies the major phases of that process and the associated bottlenecks. The authors gather evidence both from industry experts and PropTech executives and contrast their views regarding the potential of digital technologies to remove those bottlenecks.\n",
"role": "assistant"
}
] |
The effect of managerial coaching on salesperson's relationship behaviors: new evidence from frontline bank employees in China
|
[
"China",
"Relationship marketing",
"Coaching",
"Customer orientation",
"Sales performance",
"Adaptive selling"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: China's banking industry has undergone important changes since Deng Xiaoping started the Chinese economic reform in 1978. China's admission into the World Trade Organization (WTO) in 2001 has further accelerated and intensified the transformation process and the opening of China's banking industry to the outside world, moving it from an - originally - centralized monopolistic system to a more free and competitive market (Chen and Loke, 2012; Clemes et al., 2010; Wang et al., 2003).An increasing competitive market reveals that the Chinese banks face important challenges as they are no longer overwhelmingly protected by the Chinese government and they have to struggle in an environment that is now open for competition. As customers are not captive of domestic banks anymore, research shows that Chinese banks still lag behind foreign counterparts in modern management techniques and management experience (Wang et al., 2003). If Chinese banks can increase service quality, they can potentially increase customers' positive word-of-mouth (Choudhury, 2014), customer retention (Clemes et al., 2010) and sales performance (Karatepe et al., 2019). One way of increasing service quality is by developing relationship behaviors of frontline employees who deliver the service (Gerrard and Cunningham, 2004). Scholars suggest that the adoption of relationship behaviors - by both managers and frontline employees - can be instrumental in helping Chinese banks implement relationship strategies at the interface with their customers and be more successful in the marketplace (Pousa et al., 2018).In modern organizations, managers play a crucial role in leading and developing frontline employees (Mallin, 2018; Powers et al., 2010), and they are more effective if they adopt specific relationship managerial behaviors like transformational leadership, leader-member exchange and coaching (Mallin, 2018; Plank et al., 2018; Pousa et al., 2017).In the banking industry in particular, managerial coaching is a one-on-one exchange process between the manager and the frontline employee that can promote the development of relationship behaviors - like customer orientation - in the latter, leading to higher levels of performance (Pousa and Mathieu, 2014; Pousa et al., 2018).Banks and other financial institutions can reap important benefits from frontline employees who regularly use relationship behaviors - like customer orientation and adaptive selling - because these behaviors lead to increased customer satisfaction and retention (Lu-Ming, 2019; Paparoidamis and Guenzi, 2009), customers' trust in the banking institution (van Esterik-Plasmeijer and van Raaij, 2017), customer loyalty (Bhat and Darzi, 2016) and sales performance (Pousa and Mathieu, 2014; Pousa et al., 2018; Sengupta et al., 2015).The marketing and service literature suggests that customer orientation and adaptive selling are two of the most important behaviors that frontline employees can use to implement a relationship approach in their interactions with customers in order to achieve short- and long-term performance. In the banking literature, scholars are testing models that pair up managerial relationship behaviors - like coaching - with employee relationship behaviors - like customer orientation - but there is still a shortage of studies that include adaptive selling. In particular, there are almost no studies in the banking industry that integrate customer orientation and adaptive selling in the same model. In our view, this is an important gap because of the complementary nature of customer orientation - as a psychological state - and adaptive selling behavior - as a proactive process (Goad and Jaramillo, 2014).Given the importance of customer orientation and adaptive selling in the context of a relationship strategy and the potential synergy that managerial coaching can have in helping frontline bank employees improve their relationship behaviors and increase their performance, this paper tests the effect of a relational managerial behavior (i.e. managerial coaching) on frontline employee performance (i.e. sales performance) and the mediating effect of employee's relational behaviors (i.e. customer orientation and adaptive selling) in a Chinese environment.\nBackground and research hypotheses: Manager relationship behaviors: managerial coaching\nMethod: Research design and data collection procedures\nAnalysis and results: Sample\nResults: Test of main hypotheses\nDiscussion: Since the early 2000s, the Chinese financial industry has been facing increasing challenges and open competition from foreign competitors. Accordingly, Chinese banks should increase service quality and customer orientation in order to improve customer retention and sales performance. One potential solution to increase customers' loyalty and retention is to have frontline employees who can develop better relationships with customers and deliver more value to them. The literature suggests that frontline employee-customer orientation and adaptive selling are two critical constructs that managers can promote and develop through the use of managerial coaching.This study found that managerial coaching behavior is positively correlated with employee-customer orientation, customer orientation is in turn positively correlated with adaptive selling and adaptive selling is positively correlated with sales performance. Results suggest a full mediation model.Scientific contributions\n",
"role": "user"
},
{
"content": "Purpose: The purpose of the study is to test the effect of relationship managerial behavior (i.e. managerial coaching) on frontline employee performance (i.e. sales performance) and the mediating effect of employee's relationship behaviors (i.e. customer orientation and adaptive selling) in a Chinese banking environment.\nDesign/methodology/approach: Data were collected using a survey from 242 frontline employees working at a large commercial bank in Dalian (China). Measures on managerial coaching, customer orientation, adaptive selling and sales performance were adapted from the literature. Data were analyzed using structural equation modeling in AMOS 26.\nFindings: Results indicate that (1) managerial coaching positively affects employee customer orientation, (2) employee customer orientation positively affects employee adaptive selling behavior and (3) adaptive selling behavior positively affects sales performance. Bootstrap analysis confirmed the significance and stability of all the direct paths (suggesting that every mediator fully mediates the effect of its antecedent on the criterion) but none of the indirect paths found support (full mediation model).\nResearch limitations/implications: The study makes a contribution to the nomological network of managerial coaching by introducing a construct that has not been used previously (i.e. employee adaptive selling behavior) and testing its relevance in a commercial setting.\nPractical implications: The results suggest that (1) managerial coaching is a central managerial behavior for companies that would like to implement a relationship marketing strategy because it helps promote specific employee relationship behaviors like customer orientation and adaptive selling, (2) there seems to be little cultural differences in the banking industry between the Chinese and the Western banks, suggesting that coaching and other management tools can be transferable from one culture to the other and (3) that coaching is an effective tool to help employees achieve higher sales performance.\nOriginality/value: The paper contributes to the literature on the use of managerial coaching in commercial areas to increase frontline employee relationships behaviors. The identification of adaptive selling as a mediator is an original contribution because it has received little attention in scientific research. Additionally, the use of a Chinese sample of bank employees responds to recent call for more research in cross-cultural settings.\n",
"role": "assistant"
}
] |
Service quality and satisfaction in healthcare sector of Pakistan-- the patients' expectations
|
[
"Pakistan",
"Service quality",
"Patient satisfaction",
"Public health service",
"SERVQUAL",
"Private healthcare"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: With rising population (and healthcare issues) and a rising focus on improvement in each sector of the economy to achieve competiveness in national economy, the issues of patient satisfaction and service quality are seeking increasing interest among researchers and scholars all over the world. Developed countries are not only improving their healthcare sectors to serve their masses but also making money by promoting healthcare tourism (Lee and Fernando, 2015; Han and Hyun, 2015; Andaleeb, 2001). Developing countries can also reap these benefits by making available high healthcare quality at affordable costs (De Arellano, 2007; Horowitz et al., 2007; Lee and Fernando, 2015), which is a challenging task in countries like Pakistan where healthcare is not a priority, and issues like patient satisfaction and service quality are receiving more attention of academicians or scholars and less of policy makers or practitioners.\nLiterature review: Service quality\nHypotheses and methodology: Hypotheses\nResults and discussion: Table II is showing that more data have been gathered from private sector hospitals (54.8 percent) than from public sector hospitals (45.2 percent). More than half of the respondents are female (53.5 percent) as compared to males (46.5 percent). More data have been collected from the age group of 30-39 (45.2 percent), while least data filled up by the age group of 50 and over (8.5 percent).\nConclusion and recommendations: Healthcare sector is a neglected area in Pakistan, despite its growing importance in the developed countries, so the question of healthcare service quality in the country will continue arising again and again in academic papers, reports and news unless the government takes effective steps to lift up the healthcare sector of the country. The theme of this work may not be entirely novel, but the study highlights some important issues in order to improve the sanitary system in Pakistan. For instance, the results clearly indicated that the expectations of Pakistani patients are significantly higher from the public hospitals than from the private hospitals in a country where, according to the Constitution, the state is primarily responsible for providing free healthcare facilities to its citizens. The findings also underline the apparently irresponsible attitude of the state in fulfilling its constitutional commitment. One may also argue that most of the patients are well aware of the responsibilities of the state and thus their expectations from the public sector healthcare facilities are higher as compare to the private sector healthcare facilities. The findings guide the healthcare administrators and the practitioners by revealing on them the difference of expectations in term of the service quality dimensions. As reported in the results section of this paper, in public sector healthcare, the patients most strongly correlate their satisfaction with empathy and assurance and least with tangibility. On the other hand, in private sector healthcare, the correlation of patient satisfaction is strongest with responsiveness and tangibility and weakest with assurance. Thus, it can be said that when patients visit a public hospital, they expect from the staff to feel their pain and treat them with courtesy. However, when patients visit private hospitals, they expect to be treated quickly (without delays) and in a comfortable physical setup or infrastructure. Also, the results reveal on the policy makers that if they want to meet the expectations of patients from the country's overall healthcare sector, then they need to pay more attention toward empathy and assurance. This prompts the need to provide effective training to the hospital staffs and practitioners across the board because when patients visit hospitals, the first thing they encounter is people, not medicine!\nLimitations and implications: First, this study was conducted only in the hospitals of Lahore. The future study may be conducted in different cities of Pakistan as well in order to enhance the generalizability of the findings. In this study, patients having experience of both sectors evaluated both sectors. Future researchers can study by segregating the respondents in two distinct classes: private hospitals' patients and public hospitals' patients. The current study revealed that when patients who have experience of both public and private sector are asked to evaluate both sectors individually their prioritization of the service quality dimensions is almost similar for each sector. This finding seems interesting but cannot be generalized without further testing and confirmation (or disconfirmation) by future researchers. The unexpected negative relation between patient satisfaction and tangibility, despite the good fitness of the model, prompts the need to retest the model on different sample size either to verify the current findings or to present new insight on this relationship.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to assess the influence of patients' expectations from healthcare service quality on their satisfaction with nursing in public and private hospitals of Pakistan.\nDesign/methodology/approach: Data (n=456) were collected from three public sector hospitals and three private sector hospitals of Lahore, the capital of Pakistan's most populous province. Male and female patients who have experience of both sectors were surveyed using a self-administered questionnaire developed using the original SERVQUAL approach. Data were analyzed using the statistical techniques and the Laplace criterion.\nFindings: This paper attempts to explain degree of influences of five service quality constructs (empathy, responsiveness, tangibility, reliability and assurance) on Pakistani patients' expectations from the private and public sector hospitals and thus patient satisfaction. Further, this work can offer several intuitions into the effect of five constructs of service quality on patients' expectations of healthcare service quality and patient satisfaction with the service providers/nursing. The results reveal that the patient satisfaction is most strongly related to empathy in public sector and to responsiveness in private sector.\nResearch limitations/implications: In light of the previous studies and the current research findings, the study anticipates no apparently significant improvement in healthcare sector of Pakistan in near future considering various factors discussed in the study. The study will also help the service providers and the policy makers in understanding the deteriorating situation of the Pakistani healthcare sector and will guide them in identifying the areas by improving which not only the healthcare service quality in the country can be improved but also the image of healthcare sector among the masses and competitiveness of the healthcare sector can be enhanced.\nOriginality/value: The value of the study rests in its critical analysis of the current status of the healthcare sector of Pakistan with a view to suggest the areas that need to be worked on by the service providers and policy makers. Also, the study tries to settle a controversy within Pakistani healthcare literature concerning the question that who is producing more satisfied patients: private hospitals or their public counterparts?\n",
"role": "assistant"
}
] |
The current state and influential factors in the development of digital literacy in Thailand's higher education
|
[
"Thailand",
"Digital literacy",
"Thai universities",
"Digital literacy development",
"Factors for digital literacy development",
"Higher educational institutions"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Approximately two-thirds of Thailand's population currently has access to mobile broadband networks and information and communication technology services. There is also a huge increase in the use of internet - from 2.3 million people (3.7 per cent of the country's population) in 2000 to 41 million people (60.1 per cent of the country's population) in 2016 (International Telecommunications Union, 2016). Along with the growth in the use of inexpensive mobile devices, which allow continuous and easy access to the internet, Thai people now have the ability to create, use, manage, share and store information in digital formats. It is timely to begin thinking about educating the Thai population more systematically and coherently on digital literacy to ensure that the society is headed toward a healthy and successful integration of digital technologies.\n2. Research objective: The primary objective of this study is to describe the current state of digital literacy instruction in Thailand's tertiary institutions, and the factors affecting digital literacy development at these institutions.\n3. Definitions of the terms: Digital literacy means knowledge, skills and behaviors in the use of technology for communication, learning, working and leading everyday life. A digital literate person must be able to efficiently use digital technology depending on individual's competence and necessity.\n4. Research methodology: The quantitative research method was applied with a survey conducted using the self-developed questionnaire based on relevant academic concepts for posing questions that cover the research objective. There were three main items and 17 sub-items on the current state students' digital literacy development, and seven main items and 35 sub-items on factors affecting students' digital literacy development. The validity and coverage of the content of the questionnaire were tested by three experts, namely, one university academic administrator, one digital technologist and one information scientist, who evaluated and analyzed the Index of Item-Objective Congruence (IOC) from three levels of answers: +1 means suitable, -1 means not suitable and 0 means uncertain. The results showed IOC = 0.98, which is considered congruent to the objective at a high level. To test the reliability, the questionnaire was tried out with a sample group of 30 individuals in Khon Kaen University in Thailand whose characteristics were similar to the research population. The 30 individuals included 10 faculty and department academic administrators, 10 lecturers and 10 undergraduate students. The answers were analyzed to find Cronbach's alpha coefficient, which was 0.96. The survey questions were then slightly adjusted for better clarity.\n5. Research results: 5.1 Respondents' baseline data\n6. Discussion and recommendations: Twenty-first century education is very challenging, as society undergoes rapid changes that impact the country's socio-economic development. Development of technology and innovations have tremendous impacts on work and social life. In all environments, the use of internet technology has become a part of people's live (Singh, 2012). Therefore, curriculum management, instruction and preparing of graduates for this society must be in sync with global technological and digital changes. Developed nations have seen the importance of paying attention to digital literacy and have incorporated digital literacy in all levels of education with established standards for digital literacy (AASL, 2009;Martin and Grudziecki, 2006; Rosado and Bellsle, 2006; Visser, 2013). In Thailand, the results of this research bring to attention that a number of higher education institutions, roughly 60 per cent of the country's universities have established a policy or strategic plan for students' digital literacy development, either in the undergraduate program policy or in student development policy. This research shows that the majority of Thailand's higher education institutions are prepared for the changes and see the importance of students' digital literacy development. However, the remaining 40 per cent of the country's universities still are not quite ready to incorporate digital literacy into their curriculum.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to investigate the current state and influential factors in the development of digital literacy of the students in Thailand's higher education institutions.\nDesign/methodology/approach: The quantitative research method was applied with a survey conducted to collect information from the target group that consisted of administrators, lecturers and students of 116 Thai universities. Respondents included 81 administrators, 500 lecturers and 1,000 students.\nFindings: The findings revealed that 61.73 per cent of the universities set digital literacy as an undergraduate program policy, 30.86 per cent set it as a student development policy, while 12.35 per cent do not have any policy in this regard. In total, 40.74 per cent of the universities establish digital literacy as the required graduates' trait, and 22.22 per cent establish it as a graduate's identity. The study of factors related to digital literacy development showed that the respondents see that all of the factors affect students' digital literacy development at a high level, including infrastructures, human resources, students' awareness, external environments, development process, university's policy and academic management.\nOriginality/value: In Thailand, universities have seen the importance of information literacy development for students at a certain level. However, the concepts of information literacy and digital literacy differ. While information literacy mainly comprises technical skill and information management skill, digital literacy is composed of technical skill, cognitive skill and emotional-social skills. This paper presents the first research on digital literacy development in Thai higher education.\n",
"role": "assistant"
}
] |
Prevalence of Salmonella spp. and Listeria monocytogenes in different cheese types produced in Turkey
|
[
"Dairy products",
"Bacteria",
"Food poisoning",
"Turkey"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Cheese is one of the oldest human foods and is thought to have been developed approximately 8,000 years ago. Cheese is a fermented milk product with an extended shelf life and is more shelf-stable over the raw material used production (Fox et al., 2004). It is very important source of linoleic, linolenic, arachidonic acids and rich in proteins, calcium, phosphor, riboflavin, vitamin A (Tekinsen, 1997).Cheese manufacture is one of the classical preservation of the most important constituents of milk (fat and protein), which exploits lactic acid fermentation, redox potential reduction and salt addition. So, microbial spoilage in cheese is generally limited (Robinson, 2002). However, during ripening and storage, some pathogens including Salmonella spp. and Listeria (L.) monocytogenes are considered the most dangerous threats to the food safety of cheese; mainly due to poor sanitation, inadequate milk handling procedures, animal health issues and insufficient pasteurization and post-pasteurization contamination (Roberts et al., 1998).Salmonella spp. is one of the most important pathogenic genera implicated in foodborne bacterial outbreaks. It normally produces clinical disease that includes nausea, vomiting, septicemia, headache, diarrhea, abdominal cramps and fever; additional complications associated with septicemia and arthritis (Downes and Ito, 2001). In the last few years, several outbreaks of salmonellosis have been reported in France, Canada, USA, Italy, Finland and England due to the consumption of contaminated cheese (De Buyser et al., 2001).L. monocytogenes is a psychrotroph microorganism that is capable of growing at refrigeration temperature and at relatively low pH. These characteristics make L. monocytogenes particularly difficult to control in food therefore contamination could lead to high risk factor (Hobbs and Roberts, 1993). L. monocytogenes were linked to outbreaks associated with the consumption of Mexican style cheese, Vacherin Mont d'Or cheese, Camembert and Brie de Meaux cheese (Forsythe and Hayes, 1998).In Turkey, more than 40 varieties of cheese are produced. White cheese (Feta) is the most popular one that represents 60-80 percent of total cheese production. The other cheese varieties such as Kasar (similarly kashkaval in Bulgaria, caciocavallo in Italy, cascaval in Romania), Lor, Tulum, Dil (a string variety with fibrous texture) and processed cheeses (soft cheese) are well known that occupy a large place in national markets. The present study was undertaken to determine the prevalence of Salmonella spp. and L. monocytogenes in cheese samples obtained from retail markets and producers in southern Marmara region; (Bursa, Balikesir, Canakkale, Bilecik, Sakarya and Yalova) which has approximately an area of 45,200km2 (5.76 percent of the Turkish national territory) and a population of 4,738,864 million people. In addition this region is one of the richest parts of Turkey where people have high income and expenditure rates.\nMaterials and methods: Sample collection\nResults and discussion: Salmonella spp. and L. monocytogenes were detected in 0.7 percent and 2.5 percent cheese samples, respectively (Table II). According to Turkish Food Codex (TFC, 2006), the presence of Salmonella spp. and L. monocytogenes in 25g of cheese is not acceptable.Cheese samples have been examined in several countries for the presence of Salmonella spp. Akkaya and Alisarli (2006) collected 100 white cheese samples from Afyon, Turkey and 2.0 percent were found to be the positive for Salmonella spp. In another study conducted in Turkey, 42.8 percent of the 21 white cheese samples examined were found to be positive (Digrak and Yilmaz, 1996). Our findings showed similarities with their results. Contrary to this, the studies which had higher results (10.0 percent, 9.1 percent, 23.3 percent and 73.3 percent) than ours were reported by Erkan et al. (2007), Feitosa et al. (2003), Hayaloglu and Kirbag (2007) and Mendes et al. (1999) respectively. Inadequate pasteurization and post-pasteurization contamination can be considered as the source of high rates. In relation to other cheese varieties, studies performed for Salmonella spp. in kasar cheese (Kivanc, 1989), in tulum cheese (Ozalp et al., 1978), in carra cheese (Aygun et al., 2005) and in Minas Frescal cheese (Carvalho et al., 2007) found none.The prevalence of L. monocytogenes in cheese samples tested in this study (2.5 percent) was lower in comparison to those detected by Arslan and Ozdemir (2008) in Turkey (9.2 percent), Olarte et al. (1999) in Spain (5.6 percent) and McLauchlin et al. (1990) in England (64.0 percent). Similar or higher results in a lesser extent were reported by Ciftcioglu and Ugur (1991) in white cheese, Gulmez and Guven (2001) in cecil cheese and Manfreda et al. (2005) in gorgonzola cheese. The reason for contamination should be due to the use of raw milk and insufficient hygienic conditions during the manufacturing and storage. On the other hand, studies conducted by Keskin et al. (2006), Kara et al. (1999), Ashenafi (1990) and Comi et al. (1990), L. monocytogenes could not be detected in white cheese, civil cheese, ayib cheese and hard cheese, respectively.The cheese variety and manufacturing techniques may affect the types and amount bacteria present in the cheese. In the present study, the lowest rates of bacteria examined were found to be in dil cheese samples. Dil cheese is classified under the semi-soft cheese group (43-55 percent) with moisture content of 47.25 percent (Fox et al., 2000). It is made from raw milk, heated up to 32degC then inoculated with rennet. After coagulation, as done at kasar cheese, the curd is kneaded in hot water at approximately 75degC for about 1 min (Tekinsen, 1997). The high temperature applied to cheese and the low moisture content might be the reason why dil cheese had the lowest contamination rates (Kilic and Isin, 2004). Similarly kasar cheese had lower amount of bacteria examined than white cheese. Traditionally processed cheese is made by mixing and heating natural cheese, salts, calcium chelating salts (also commonly known as melting or emulsifying salts), water and fat up to 85degC (Lee and Anema, 2009). There was 1 (1.4 percent) L. monocytogenes positive sample in this type of cheese examined.In this study, the highest rates of Salmonella spp. (1.9 percent) and L. monocytogenes (4.8 percent) have been detected in white cheese samples. Turkish white cheese is soft variety based on its moisture content (55-65 percent). Also, it has a soft texture when fresh (Hayaloglu et al., 2002). Soft cheeses contain high levels of moisture and pH, particularly at the surface, increases during ripening; so that they are more prone to microbial growth than hard or semi-hard cheeses. The prevalence of pathogenic microorganisms in soft cheeses depends upon the quality of milk used, heat treatment of milk and general sanitation in the cheese plant. Because most soft cheeses are consumed fresh, it is important that the milk or other dairy ingredients used for soft cheese manufacture be adequately pasteurized (Robinson, 2002). Producing white cheese is different than kasar, dil and processed cheeses. Milk is first heated up to 77degC and kept at that temperature for 14 seconds, then cooled to 31degC. CaCl2 is then added to milk and rennet is inoculated (Tekinsen, 1997). Producers sometimes do not heat milk up to pasteurization temperatures and/or add raw milk into pasteurized milk. The reason for this is to avoid adding CaCl2, which means extra cost and to save energy and time. Because coagulation needs Ca++, with over heating Ca++ is precipitated and precipitated Ca++ is not used in forming curd, the yield goes down (Kosikowski and Mistry, 1997). The other reason for the high contamination rates for white cheese might be the brine. Pathogens may be transferred to cheese via contaminated brine.In conclusion, the result of this study confirmed that cheese may be contaminated pathogens such as Salmonella spp. and L. monocytogenes, which can cause great public health problems. The rate of survival and/or growth of pathogenic bacteria in cheeses depend on the ecological conditions water activity, pH, salt content, temperature of maturation within the cheese and/or brine. In this study, it was reported that especially white brined cheese samples might pose great public health concerns. In addition, consumption of immature white brined cheese is very common in Turkey, which also extends the possibility of health hazards of this type of cheese. White cheeses are manufactured with different protocols. The quality of the raw milk, heat treatment of the milk, activity of the starter culture and the salting process together with the storage in brine are the most important control points for the prevention of growth/survival of undesirable microorganisms during the manufacture of white-brined cheeses. Therefore, it is essential that applying good hygienic practice (HACCP, GMP, GHP) and developing standard procedure of production would be beneficial for controlling food-borne pathogens and enhancing the safety of food.\n",
"role": "user"
},
{
"content": "Purpose: - The present paper seeks to determine the prevalence of Salmonella spp. and Listeria (L.) monocytogenes in several cheese varieties.\nDesign/methodology/approach: - A total 280 cheese samples (105 white cheese, 70 processed cheese, 45 dil cheese and 60 kasar cheese) purchased from supermarkets in six provinces of Turkey were collected at intervals between March 2007 and February 2008.\nFindings: - The results showed that 1.9 percent of white cheese samples were contaminated with Salmonella spp, whereas the L. monocytogenes prevalence was 4.8 percent in this type of cheese sample. The prevalence of L. monocytogenes in processed cheese and kasar cheese were found to be 1.4 percent and 1.7 percent, respectively. No Salmonella spp. was isolated from these cheese varieties. Neither Salmonella spp. nor L. monocytogenes were found in dil cheese samples examined.\nOriginality/value: - Cheese is a ready to eat product that with a low incidence of contamination may pose great public health concerns. Microbiological evaluation of different cheese samples produced with different manufacturing methods and the possible effects of manufacturing methods on the microbiological quality of these cheese samples are the originality criteria of the study. The cheeses were also sold in a very restricted area of Turkey. The results of this study indicate that white cheese has the highest rates of contamination due to the lack of standardized procedures of manufacturing and ripening.\n",
"role": "assistant"
}
] |
A flow direction weighting scheme for facility layout in cellular manufacturing
|
[
"Cellular manufacturing",
"Facilities design",
"Group technology",
"Manufacturing systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Cellular manufacturing systems (CMS) can be defined as \"grouping dissimilar production equipment into machine cells, where each cell specialises in the production of a number of components or work part families\" Groover (2001a). It is an example of a mixed model production system which utilises the philosophy of group technology (GT) wherein \"similar components or work parts are identified and grouped together to take advantage of their similarities in design and production\" Groover (2001b). In terms of practical implementation, CMS can take many forms as classified by Groover (2001c), with two major sub-divisions being the presence or absence of automated equipment. In the case of automated CMS, they can be regarded as the current manifestation of a partnership between the practical implementation of techniques and concepts developed in the area of GT, computer numerically controlled (CNC) machinery and modern information and communications technology (ICT). Whereas, manual or semi-automated CMS on the other hand, may be viewed as a manufacturing system based on the principles of GT, but implemented via manual and/or low-cost automation which is controlled through the Japanese philosophy of just-in-time (JIT) production.In either case, the efficient and effective design of such manufacturing systems is a complex engineering problem, one which consists of many facets including, for example, aspects of overall system design, control and performance; together with component - cell formation, machine - cell layout and work-part scheduling. Over the years, a great deal of research has been carried out into these areas together with many other aspects of CMS and GT following the pioneering work of Mitrofanov (1959/1966) in grouping theory, Burbridge (1963) in production flow analysis (PFA), Opitz (1970) in component classification and coding and McAuley (1972) in cluster analysis. In particular, many authors have evinced an enduring interest in solving both the cell formation and machine layout problems which are associated with this type of manufacturing system. The cell formation problem can be defined as the identification from established shop floor data of compatible component families and processing machine sets such that they can be formed into suitably sized stand alone or linked manufacturing cells. Whereas the machine layout problem involves the order in which the identified machines in a cell are laid out to minimise overall work part movement within a cell or series of linked cells for the component set being processed.Various techniques have been utilised over the years in an attempt to successfully solve the cell formation problem and these may be sub-divided into two generally accepted approaches. The first can be designated as design oriented and is based on the design of the work parts to be processed and the machinery used to process them. As component geometry and machine capabilities in this approach are regarded as being fixed, then data relating to each can be used to identify suitable work part and machine groupings Dapiran and Manieri (1983). In the second more popular approach, designated as production oriented, information relating to the processing sequence involved in manufacturing a component is generally used to identify the appropriate work part and machine groupings. The techniques utilised in this approach are many and varied usually falling into one or other of the following seven classifications:(1) a graph theoretic approach as used by Rajagopalan and Batra (1975);(2) array-based methods as typified by rank order clustering King (1980);(3) heuristic procedures such as that proposed by Tabucanon and Ojha (1987);(4) non-hierarchical clustering methods (Chandrasekharan and Rajagopalan, 1986);(5) hierarchical clustering techniques (Gupta and Seifoddini, 1990; John et al., 2011);(6) mathematical programming as outlined by Lee and Garcia-Diaz (1993); and(7) artificial intelligence techniques including fuzzy logic (Xu and Wang, 1989), simulated annealing (Venugopal and Narendran, 1992), genetic algorithms (Uddin and Shanker, 2002), and neural networks (Solymanpur et al., 2002).In more recent years, academic research studies in the area of CMS have tended towards examining the problems associated with the practical implementation of such techniques and the limitations which often prevent their use in the effective design of CMS. Typical of these are studies in respect of constraints relating to:(1) A focus on the final product, machine capacity, utilisation and component scheduling (Al-Mubarak et al., 2003; Solimanpur et al., 2004).(2) The minimisation of inter and intra cellular work part movement or material handling, the layout characteristic or visual impact and processing unity, plus the associated layout flexibility and cost (Ngampak and Phruksaphanrat, 2011).(3) CMS design under conditions of randomness or uncertainty (Ariafar et al., 2011; Saidi-Mehrabad and Ghezavati, 2009).(4) Optimising system size or the number of cells which make up the system and the number of machines per cell, together with the effect and portability of system design techniques on their ability to solve large scale industrial problems and to employ more production data in the analysis (Mahadevan and Venkataramanaiah, 2008).(5) The evaluation of CMS design methods for efficacy via their assessment by varying performance measures and the solution of problems associated with CMS production planning and control (Arora et al., 2011).The second difficulty in the list being essentially the machine layout problem mentioned earlier, an aspect linked to CMS design which is sometimes neglected in certain techniques proposed to solve the question of cell formation. In essence, it may be regarded as the second part of the question of efficient CMS design and is often referred to as the group technology configuration problem (GTCP), deciding how to layout the identified component - machine grouping for minimum work part movement and hence the cost of inter and intra cellular transportation. It is this aspect of CMS design that this paper addresses via a modification to the technique originally proposed by Mukhopadhyay et al. (2000).In their original paper Mukhopadhyay et al. (2000) outlined, and effectively employed, a modified hamiltonian chain (MHC) graph theoretic approach to the GTCP. The Mukhopadhyay et al. (2000), algorithm, hereafter referred to as MBSMHCA, consists of a two-stage process. Stage I forms a graphical depiction of the component flow routes (from a given knowledge of the machine component incidence matrix (MCIM)), whilst Stage II generates a MHC (a linear arrangement of the facilities employed to process the example component set). The approach presented is straightforward to understand and to apply and, in their conclusions, Mukhopadhyay et al. (2000) show that the results obtained, in tests employing the suggested approach, compare favourably with other well known solution methods.In this paper, the MBSMHCA technique is re-examined with a view to its further improvement, together with the elimination of some of the difficulties arising from the use of this technique and their possible solution as identified by John et al. (2009). It suggests an alternative weighting scheme to that in the original methodology and to that proposed by John et al. (2009) which could further improve the overall solution quality for this type of problem. The suggested modifications are subsequently applied to the test problems previously outlined in the existing research literature by Mukhopadhyay et al. (2000) and John et al. (2009), and the results compared to those obtained by other published algorithmic procedures. Some concluding comments in respect of the comparison of the results produced by the basic method and its suggested modifications are then offered for consideration.The MBSMHCA technique\nConclusions: Reflecting upon the results contained in this paper, one obvious conclusion can be drawn regarding the techniques used. These appear to fall into two categories, those that are aimed at ensuring coherence in machine grouping and those with the primary aim of flow efficiency in the production system. Although the former is supposed to produce the latter, it does not appear to be the case in two out of the three problems examined here. This may well be a quirk of the problems chosen and more testing on additional examples is required before this statement can be proved or disproved. Another conclusion evident from the examination process is that methodologies which require demand or volume information as an input to the solution technique may well be at a disadvantage where production levels are not stable. The suspicion being that the result they produce may well be sensitive to fluctuations in this factor. That is a \"good\" result for a given set of demand values may be a \"bad\" result if those values change. In this study, only the heuristic pattern matching technique of Ho et al. (1993) fell into this category and could only be examined in one of the three problems. Here it produced a good sub-optimal answer in terms of inter-stage movement but a poor grouping efficiency of only 57.94 per cent. As with the previous conclusion, only more testing on a wide variety of problems can prove or disprove the robustness of demand linked solution techniques. Finally, a broader reflection upon this contribution and the results produced would have to enquire whether there exists any correlation between the grouping efficiency (see footnote 1) and similar criterion used as a measure of \"goodness\" when evaluating comparative machine clustering algorithms and the inter-stage flow measure used in this contribution to evaluate \"solution string goodness\". The FDWS modification to the basic MBSMHCA approach appears to provide good sub or optimal answers in the problems examined, providing a valuable economic and though-put time advantage, whilst returning a fair to good value of grouping efficiency in all three cases. This again requires further examination as with the previous conclusions, before its performance as an initial solution technique for perhaps more extensive simulation design studies on proposed real life manufacturing systems, can be said to be superior to other approaches. It may, however, be of interest to know that the FDWS technique produces for the main example problem used in this contribution, the same minimum work part transfer solution as that found via the use of the weighted similarity coefficient (WSC) approach proposed by John et al. (2011).Example calculation 2\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to re-examine the solution methodology to the facility grouping problem in group technology originally outlined by Mukhopadhyay et al. The aim of this re-examination is to attempt to resolve some of the problems associated with the original solution methodology and to provide the system designer with an improved technique for the creation of an efficient initial system design solution. Cellular manufacture by its very nature requires work-part transfer both within a machine cell and in some instances between machine cells. By utilising a design methodology which seeks to minimise work-part transfer, these non-value but cost adding tasks are kept to a minimum, thereby providing a faster throughput time and lower operational costs for companies who use cellular manufacturing systems in their production operations.\nDesign/methodology/approach: - In approach, the technique advocated parallels the possible alternative solution method suggested by John et al. to resolve the problem of edge ranking ties which exists in the original Mukhopadhyay et al. solution technique. Presented and discussed in this contribution is a revised flow direction weighting scheme (FDWS) which simplifies the solution input data requirements and overcomes a perceived disadvantage in the earlier scheme outlined by John et al. The methodology outlined in the paper is subsequently evaluated for its generic nature, applicability and effectiveness via the use of previously published synthetic production data, and the findings compared to the results determined by several alternative approaches to the problem.\nFindings: - The development and application of the FDWS to cellular manufacturing system design is outlined in the paper and the usefulness of the technique is subsequently examined and compared to the results found for several competing methods. The approach provides for the identification of improved solutions to the test problems examined in both of the earlier contributions by Mukhopadhyay et al., and John et al. In addition, the results show that irrespective of the assessment criteria used, the FDWS technique provides solutions that are comparable to or better than the alternative approaches examined.\nOriginality/value: - The proposed FDWS technique further contributes to the existing knowledge base on cellular manufacturing system design. It can provide good initial solution guidance for manufacturing system designers and thereby reduce the time required to achieve an acceptable and cost efficient design for a cellular production system.\n",
"role": "assistant"
}
] |
Studying the influence of strategic human resource competencies on the adoption of green human resource management practices
|
[
"Malaysia",
"Green HRM practices",
"Strategic HR competencies"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Increased concern about environment issues in recent years has made the \"go green\" concept a focus point by organizations. Subsequently, companies have taken several green initiatives such as green marketing, green supply chain, green management, and green human resource management (HRM). Among the green practices by companies, green HRM is gaining increased attention. Moving toward green practices in any organization requires HR support. The crucial role of HR professionals in green practices has been highlighted in earlier studies (e.g. Jabbour et al., 2010, 2012; Jabbour and Santos, 2008). Specifically, HR as a strategic partner and value-added function should recognize the importance of green practices, initiate environmental friendly activities, change the existing processes, and align its practices accordingly (Cohen et al., 2010; Mishra et al., 2014).\n2. Strategic HR competencies and green HRM practices: Green HRM refers to \"the systematic, planned alignment of typical HRM practices with the organization's environmental goals\" (Jabbour, 2013, pp. 147-8). In the literature, the extent to which HRM is green has been frequently studied in a continuum of all HRM practices, namely analysis and description of job positions, recruitment and selection, training and development, performance appraisal, and rewards (Jabbour, 2011; Jabbour et al., 2010; Renwick et al., 2013).\n3. Methodology: The sample of this study included 87 HR professionals in manufacturing and service companies in Malaysia. The samples were randomly selected from Federation of Malaysian Manufacturer Directory 2015. This study adapted an instrument from prior studies to get response from HR professionals. The scale consisted of 15 questions about green HRM practices and 20 questions about strategic HR competencies. A five-point Likert type scale was used ranging from 1 (\"strongly disagree\") to 5 (\"strongly agree\").\n4. Results and discussions: The demographic profile of respondents and responding companies are illustrated in Table I.\n5. Conclusion: The immense attention to environmental issues has stimulated a growing number of organizations to adopt green practices into their business. An extensive number of studies have been carried out and acknowledged the importance of HRM in organizations. However, studies on integration of environmental perspective in HRM - green HRM have been scarce. As such, the current study provided empirical evidences for the effect of strategic HR competencies on the adoption of green HRM practices in organizations.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the influence of strategic human resource (HR) competencies of HR professionals on the adoption of green human resource management (HRM) practices.\nDesign/methodology/approach: A quantitative research approach using survey was employed to get insights from 87 HR professionals from Malaysian manufacturing and service companies.\nFindings: The findings revealed that only strategic positioner and change champion are significantly related to green HRM practices. Specifically, strategic positioner associates positively with all green HRM practices including green analysis and description of job position, green recruitment, green selection, green training, green performance assessment, and green rewards. Additionally, change champion is positively related to green analysis and description of job position, and green rewards.\nOriginality/value: The extant literature suggests HR professionals' competencies can play a key role in adopting green HRM. However, examining the relationship between strategic HR competencies and each of the green HRM practices has been limited. As such, this study is timely to address the existing gap in the literature and provide fresh insights and implications in this regard.\n",
"role": "assistant"
}
] |
Value-at-risk performance in emerging and developed countries
|
[
"Value-at-risk",
"Artificial neural networks",
"ARCH models",
"Copula functions"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The topic of risk has been one of the main issues studied and discussed in recent years, perhaps because of the recurring crises that the market witnessed in recent decades. However, it was from the 1970s that the financial industry developed several controlling mechanisms to mitigate the risks inherent in the financial market. With the popularization of the capital market came the rise of speculators and growth volatility. What was once a market more restricted to financial institutions, has now become popular access.\n2. Theoretical background: In this section we present the theoretical concepts that govern market risk, starting with VaR and its estimation metrics and a review of the main studies on the subject.\n3. Research hypotheses: Considering the results of the studies of Palaro (2004), Aiube (2009), Coelho (2012), Pereira and Pereira (2009), Berger (2013), Berger and Missong (2014) and Han et al. (2017) we can develop some hypotheses regarding the best model for risk adjustment:\n4. Methodological aspects: 4.1 Data\n5. Results and discussion: Tables V and VI present the descriptive statistics of the returns of the main stock indexes of developed countries, followed by the heteroscedasticity, linearity and autocorrelation function (ACF) tests and partial autocorrelation function (PACF) in the period between 1995 and 2013.\n6. Final considerations: This study aimed to evaluate the predictive capacity of market risk models in developed and emerging countries. In order to do so, six investment portfolios of country indexes were tested to extract the VaR financial indicator applied to the daily returns of the portfolios. Four classes of models (traditional, multivariate ARCH, ANNs and copula functions) were used in the study, in the periods of the Asian crisis of 1997 and of the subprime crisis of 2008.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to evaluate the predictive capacity of market risk estimation models in times of financial crises.\nDesign/methodology/approach: For this, value-at-risk (VaR) valuation models applied to the daily returns of portfolios composed of stock indexes of developed and emerging countries were tested. The Historical Simulation VaR model, multivariate ARCH models (BEKK, VECH and constant conditional correlation), artificial neural networks and copula functions were tested. The data sample refers to the periods of two international financial crises, the Asian Crisis of 1997, and the US Sub Prime Crisis of 2008.\nFindings: The results pointed out that the multivariate ARCH models (VECH and BEKK) and Copula-Clayton had similar performance, with good adjustments in 100 percent of the tests. It was not possible to perceive significant differences between the adjustments for developed and emerging countries and of the crisis and normal periods, which was different to what was expected.\nOriginality/value: Previous studies focus on the estimation of VaR by a group of models. One of the contributions of this paper is to use several forms of estimation.\n",
"role": "assistant"
}
] |
Is co-creation really a booster for brand equity? The role of co-creation in observer-based brand equity (OBBE)
|
[
"Co-creation",
"Expertise",
"Intensity of integration",
"Non-integrated consumers",
"Observer-based brand equity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In 2011, McDonald's conducted a co-creation campaign in Germany with the theme \"Mein Burger\". About 116,000 \"concepts\" for future burger recipes were proposed by consumers via a project-specific Facebook platform - that is, an online form of co-creation. Some 1.5 million who had not necessarily taken part in that first stage of the promotion subsequently registered their \"votes\" for the competing concepts. Such participating customers, who either proposed concepts or voted for them, are described in the literature as being \"integrated\"; the majority who did neither can be classified as \"non-integrated\". Though the number of the former is impressive in its own right, it is nevertheless small in comparison with the roughly 2.5 million customers per day who visited a McDonald's restaurant in that year, noticing and consuming the results of the Mein Burger initiative (Herrmann, 2014). The literature shows that it is very typical of a co-creation campaign that the proportion of non-integrated customers in that daily footfall heavily outnumber the minority of integrated customers (Burnett and Buerkle, 2004; Takahashi et al., 2007; Madupu and Cooley, 2010; Fuchs et al., 2013). On average, one integrated consumer is facing a hundred passive ones (Carroll and Rosson, 2008). This phenomenon has been summed up as the \"90-9-1-principle\" (Nielson, 2006), noting that in customer integration, 90 per cent of consumers remain passive, 9 per cent are somewhat active in sharing or liking the co-creation aim, for example via Facebook, but are not actively engaged with the co-creation campaign, and only 1 per cent are active.\nConceptual background and development of hypotheses: The consumer-based brand equity is the sum of its parts\nEmpirical study: Pre-tests and procedure\nGeneral discussion: Theoretical and managerial implications\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to analyse the general effect of co-created products on the brand equity of observers (OBBE). The influence of different implementations of the co-creation approach on the OBBE is tested. It is also discussed whether co-creation can be a strategic method for companies to positively affect the OBBE in the mass market.\nDesign/methodology/approach: A between-subject experiment with a 2 (intensity of integration: democratically voted vs commonly created) x 2 (expert knowledge: no expert knowledge vs expert knowledge) design plus one control group (zero co-creation) is conducted for two brands to test the postulated hypotheses.\nFindings: Co-creation can have a weak positive effect on the OBBE. Integration intensity and expertise of integrated consumer also affect the OBBE only marginally.\nResearch limitations/implications: Further research might investigate whether the initial brand equity has a moderating effect. Also brand image and underlying product category could influence the relation between co-creation and the OBBE and would be valuable for future studies.\nPractical implications: Brand managers should aim to convert observers into participants, instead of setting the focus on the presentation of the user-designed product to the mass market.\nOriginality/value: This study is one of the few analyzing the effects of co-creation on observers in terms of brand equity. In addition to existing research, the concept of expertise in combination with co-creation and its influence on the OBBE is explored.\n",
"role": "assistant"
}
] |
Integrated management system: The experiences of three Australian organisations
|
[
"Environmental management",
"Management techniques",
"Quality management",
"Occupational health and safety",
"Case studies",
"Australia"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The implementation and certification of quality, environmental and occupational health and safety (OHS) systems has been a major activity for many organisations in light of increasing pressure from their internal and external stakeholders including the regulatory bodies, community, customers, employees, suppliers and the government, to name a few. Implementation and maintenance of systems/standards or making changes to existing systems requires the allocation of significant resources that may directly impact on the bottom line of the organization. Additional costs are incurred in obtaining certification by a third party that includes auditor and registration fees. Many small and medium-sized enterprises (SME) typically will not have the expertise in-house and will use the services of an external consultant to assist them in implementing the new system.Integration of the different management systems is the call of the hour, as it can result in significant benefits including cost savings and/or reduction in the use of valuable organizational resources. The literature review highlights the potential benefits for organizations from integrating their different management systems into a single system. There are few case study examples that have documented the experiences of organizations with integration of the different systems. In order to add to the existing literature, this paper presents the experiences of three Australian companies that have successfully undertaken the integration of their management systems. The aim of the paper is to understand the challenges, benefits and the critical success factors relating to integration of the various systems.The rest of the paper is structured as follows. The next section presents the literature review, followed by the research methodology used. The subsequent sections present the findings with respect to the benefits from integration, the challenges associated with integration and measures taken by organizations to overcome these. The major lessons learnt by the three companies from their integration approach are then identified and discussed. The final section presents the conclusions.\nLiterature review: Background to management systems and their integration\nResearch methodology and profile of the companies: The researchers opted for the case-study approach, in alignment with the scope and benefits offered by this methodology. In addition to finding answers or clarifying ambiguous or uncertain issues, the exploratory case studies also serve the purpose of narrowing the research topic or field. This aspect has been supported by a number of researchers including Wallace (1984), Eisenhardt (1989), McCutcheon and Meredith (1993), Greenwalt (1994), McGuire (1995), Simon and Sohal (1996) and Terziovski et al. (1997). Another advantage accompanied by case-studies is that its results, and though \"general [and] tentative\" in the view of Zikmund (1997, p. 108) can be easily coupled with explanatory and descriptive methods. In Yin's (1994, p. 12) view, the case studies try to uncover the answers to \"a decision or a set of decisions [for instance] why they were taken, how they were implemented and with what result\". Case studies thus provide a holistic view of the variables by going into the depth of the topic in consideration (for instance as in the case of interviews) by the use of \"why, when, who, what, where [and] how\" (Wallace, 1984, p. 181; Yin, 1994, p. 3; Zikmund, 1997, p. 38) questions. The results obtained following the case studies, though general in nature, can be presented in a coordinated and holistic manner.The three companies involved in this study were selected based on previous contacts with the managers from the respective companies. After re-establishing contact with the relevant managers and informing them of the objectives of the study, interviews were arranged to be conducted on-site. The interviews lasted between one and two hours and were taped, transcribed and written up as individual case studies. The case studies were then sent to the interviewees for validation. Any changes as required were made accordingly before each case study was finalized. Site visits and review of companies\" documentation also assisted in learning and confirming the content of the case studies. The finalized case studies were cross-analyzed for similarities and differences relating to the various aspects of integration. Conducting cross-case analysis has been supported by researchers including Eisenhardt (1989) and Ndonzuau et al. (2002, p. 282), as it increases the validity and reliability of the case-study methodology. The next section presents the findings of the case studies. To protect confidentiality, the three companies are referred to as Companies A, B and C.Table I presents a brief overview of the three companies interviewed. The Appendix presents a more detailed background on the three companies and their management systems.The three companies represent different industry sectors and are all multinational organizations. All three companies implemented their EMS, QMS and OHS systems at about the same time and also initiated the integration process at about the same time.The key driver for integration for all three companies was to make better use of resources.The following sections of the paper present our analysis in terms of the benefits of integration, challenges associated with integration and the key lessons learnt from the experiences of the companies involved in this study.\nBenefits from integration: Integration of separate management systems (QMS, EMS and OHS) into one operational system has provided the three companies with a number of tangible and non-quantifiable benefits. These are discussed below.Strategic planning\nChallenges/impediments accompanied by integration: Even though the three companies interviewed did not experience all the impediments listed below, these are nonetheless recognised as potential challenges commonly experienced by companies integrating their management systems. The challenges/impediments include:1. People's attitudes. Behaviour and attitudes of the employees within the organization can act as a determining factor for successful integration of the systems and its maintenance. It is common knowledge that nobody likes to change their old ways of doing things and implementing new systems or even updating existing procedures can involve substantial change. It is thus sometimes both challenging and frustrating for people to communicate and explain their system requirements to others and explain how their systems may be best integrated with the other system(s). Some resistance is thus to be expected during integration of the systems, which can generally be addressed by educating and training the employees. Some resistance may also be experienced from people who lose \"ownership\" of, for instance, forms or procedures resulting from integration of the systems.2. Lack of strategic planning. Companies that lack a formal long-term \"strategic plan\" can experience resistance and delays in the completion of the integration of their systems. Due to communication gaps between the top/middle management and employees of the aspects of the integration process, resistance would be experienced within the various management levels.3. Lack of expertise and use of consultants. One of the most common challenges faced by a number of companies when integrating its management systems is the availability and/or lack of qualified personnel to complete the job. Many times when the companies are able to find an expert in the area, paying them appropriate salaries becomes an additional challenge/problem for them due to lack of, or limited, resources, once again restricting them from recruiting such a person. Consultants are generally used by a number of organizations to implement their management systems. One of the problems associated with the use of consultants is the very high fees charged by them and many organizations, particularly SMEs, are unable to pay the consultation fees over an extended period of time. Another disadvantage of using consultants is that even though they can implement the systems, they cannot assist the organization in maintaining these, which has to be done by internal people. In the view of the interviewee, it is thus preferable and recommended that the system is implemented and maintained internally by the organization.4. Continually changing regulations and guidelines. Due to the continually changing regulations and guidelines, for instance in the environmental field, organizations are faced with the challenge of updating their procedures and systems. These changes consequently force the organizations to continually review their procedures, targets and objectives and update them accordingly.5. Reporting of results. To avoid delays in the finalisation and facilitation of the future improvement plans, the company's top management needs to receive and review the progress/audit reports promptly. For this to be successful, managers should ensure that there exists a fast reporting system across the organization and its national subsidiaries around the globe.6. Time-delays in integration. Some departments within the company could take more time than anticipated to initially understand and implement the integrated system. This could delay meeting the target completion date set by the management. Nevertheless, this possibility needs to be considered and taken into account when finalising the target dates.By being aware of some of the challenges experienced by companies within the same or different sectors, organizations contemplating integrating their systems can accordingly prepare themselves to meet and address these challenges, thus making the implementation of the systems and their integration faster and smoother.\nLessons learned during integration: The major lessons learnt by the three companies are identified below. These can be used as recommendations for other organizations currently undertaking or contemplating integrating their management systems:Top management commitment and management review\nConclusion: The paper presented experiences of three Australian-based organizations with respect to the integration of their quality, environmental and OHS systems. Based on the literature review and the experiences of these three organizations, it can be concluded that the \"integration of systems/standards\" is one of the major strategies for ensuring survival and savings (time, cost, resources) for the organizations in the twenty-first century. At the same time it is also strongly recommended that a multi-functional team be put together to manage the process. This would ensure that all areas of the organization are represented and consulted and the positive and negative issues relating to each function are considered in the integration process.The interviews revealed a number of quantifiable and un-quantifiable benefits experienced by the companies from operating one integrated system such as saving of dollars, better utilisation of resources and improved communication across the organization, to name a few. However, for the benefits to be realized it is essential that organizations are aware of the challenges and obstacles accompanied by integration of systems/standards. If these challenges are not addressed early in the process they can delay the completion of the integration process. Recommendations for other organizations contemplating integrating their management system include, obtaining commitment from the top management; having adequate resources to integrate the systems; having communication and training across the organization in aspects of integration, and last but not least having integrated audits. Implementation of these recommendations may vary from one organization to another, however, it would result in less resistance for the organizations following them. More case studies and quantitative studies such as questionnaire surveys need to be undertaken to document the progress being made in the area of integration of management systems. Research also needs to identify the long-term effectiveness of integrated management systems on the overall performance of the organization.\n",
"role": "user"
},
{
"content": "Purpose: - Management systems and standards have become a key part of the organisation's lifeline and a prerequisite for survival in the twenty-first century. Systems for quality environmental and occupational health and safety (OHS) now form the three main pillars of the organisation, the fourth one being financial accounting. In light of the increasing pressure and demands from different stakeholders, it is becoming necessary for organisations to adopt the different systems/standards. However, to achieve the benefits from the implementation and subsequently maintenance of these systems it is only a practical and logical step that the existing management systems/standards be integrated into a single system.\nDesign/methodology/approach: - This paper presents the experiences of three Australian-based organisations that have successfully undertaken the integration of their management systems/standards. Data for this paper were collected through in-depth interviews conducted with the managers responsible for quality, environment and OHS systems.\nFindings: - The interviews revealed a number of quantifiable and unquantifiable benefits experienced by the companies from operating one integrated system, such as saving of dollars, better utilisation of resources and improved communication across the organisation, to name a few. However, for the benefits to be realized it is essential that organisations are aware of the challenges and obstacles accompanying integration of systems/standards. If these challenges are not addressed early in the process they can delay the completion of the integration process.\nOriginality/value: - Recommendations for other organisations contemplating integrating their management system include: obtaining commitment from the top management; having adequate resources to integrate the systems; having communication and training across the organisation in aspects of integration; and, last but not the least, having integrated audits. Implementation of these recommendations may vary from one organisation to another; however, it would result in lesser resistance for the organisations following them.\n",
"role": "assistant"
}
] |
Determinants of wine consumption of US consumers: an econometric analysis
|
[
"Consumer behavior",
"Market orientation",
"Regression analysis",
"Wines",
"United States of America"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The driving forces of globalization have significantly altered the competitive landscape of the wine industry. While the demand for wine in the USA is growing, US wineries continue to lose domestic market share. In order to counter foreign competition as well as grow in this volatile environment, they must evolve by adopting a stronger market orientation, with consumers' needs at the core of their plans. This shift in orientation requires a solid understanding of their potential customers' habits and needs. To this end, this paper offers a fresh perspective on the consumption behavior of wine drinkers in the USA by providing a framework to identify and quantify determinants of wine consumption.We first introduce the US wine market and discuss the challenges facing US producers in selling wine domestically. We peruse extant research to distill what can be learned from prior survey-based analyzes and other relevant consumer studies. We then present our research methodology and goals. In particular, we estimate the determinants of wine consumption by employing an econometric procedure called categorical regression estimation (CATREG) for non-numeric response variables. Lastly, we use these empirical findings, combined with research from other sources to suggest how US producers may wish to enhance their marketing strategies to consider the demographic and behavioral patterns of US consumers or engage in additional research to better qualify their desired customers and, hence, market orientation.The US wine market and the mismatch between supply and demand\nResearch methodology: The study presents a statistical description of wine consumers, as well as an econometric analysis of the determinants of their wine consumption. Data on consumer characteristics and wine consumption were collected by student-administered surveys as part of a graduate marketing class project. Owing to certain resource and time constraints, we adopted a non-probabilistic sampling method to draw samples. That is, respondents were contacted at convenient locations, such as locations on a university campus and in the financial district of a Northern California city. Eventually, we collected 122 completed surveys from consumers in Northern California. The questionnaire used close-ended, multiple-choice questions to obtain information on certain demographic variables (age, gender, income, occupation, race), as well as behavioral variables (uses, benefits, influences, consumption volume, approximate monthly expenditures on wine), and knowledge level related to wine consumption. Using these data, we developed and empirically tested a model where various dimensions of consumers determine the consumption of wine. It should also be noted that we were concerned primarily with quantities consumed and average purchase price of bottles. While there is value in discovering correlations between demographics and other wine characteristics, such as generational or ethnic preferences for varietals or appellations of wine, such investigation was outside the scope of our analysis. We do not test formal hypotheses but explain and empirically explore how different characteristics of a consumer influences her wine consumption.Econometric framework: CATREG\nEmpirical findings: Sample characteristics\nDiscussion and conclusion: Despite certain limitations discussed below, our research contributes theoretically to the development of a framework for understanding consumption behavior of wine consumers. From the methodology perspective, we provide a roadmap to employ CATREG method, which is useful given a combination of nominal, ordinal, and interval-level independent variables. The CATREG method is further considered appropriate for measuring relative importance of different predictors. Although extensively used and quite useful for analysis with categorical variables, CATREG results are further verified by deriving both OLS and Tobit estimates. In that sense, this study does provide a quantitative starting point for additional and more in-depth future analysis of US wine drinkers. In addition, several managerial implications arise for wineries. Translated properly into plans and actions, these implications may help them ensure their survival and growth in the volatile and intensely competitive domestic wine market.First, the findings draw our attention to the significance of a drinker's knowledge of wine. Knowledge of wine positively influences a consumer's involvement in wine purchasing and her consumption of wine. In order to make consumers feel comfortable with their choice of wine, wineries must facilitate adequate information and education for the potential drinkers. Wineries must start with a clear understanding of consumers' needs and behaviors and devise appropriate communication and marketing strategies in order to help them overcome their knowledge deficiency in a practical and non-intimidating manner. This enlightenment is especially critical when consumers \"do find wine somewhat intimidating and mysterious\" (Heeger, 2006).The wine industry needs to overcome an enormous challenge when it comes to changing consumer knowledge and impressions around buying and drinking wine. Currently, 80 per cent of wine consumers are either \"uninvolved\" or uneducated about wine. Consumers feel intimidated by \"wine geek speaks\" on wine labels and have trouble remembering which wines they bought and liked. Researchers watching consumer behavior have noticed shoppers appear to be confused during the wine selection process. Customers have expressed that they want to be able to easily and consistently identify wines they will enjoy without having to solicit personal assistance in the store (Stallcup, 2005; Goulet, 2004). Thus, wineries should consider marketing strategies that increase consumer knowledge as a high priority.Secondly, wineries must also realize that as consumers become more knowledgeable, they may not uniformly purchase more expensive wines. This finding is consistent with Moulton et al. (2001)'s debunking of the myth that consumers will uniformly trade up to more expensive wines as they augment their wine knowledge. It is also consistent with Constellation's (2005) findings that 49 per cent of enthusiasts, self-described as the most knowledgeable segment, have not bought a bottle of wine over $15 in 6 months. Wineries that limit themselves to selling only wines priced over $10 are missing sales opportunities among both the frequent drinkers (per Table III) and the more knowledgeable (per Figure 2). Cholette and Castaldi (2005) show that US producers are in retreat from the lower price points, as New World imports, primarily from Australia, currently claim a majority share (51 per cent) of all wines sold through US supermarkets in the $6 to $9 per bottle price range. While higher-priced wines may offer both more prestige and higher per-bottle profit margins, our results suggest would be a mistake for US producers to abandon such a large and strategic part of the market.Thirdly, our study suggests that US producers may wish to consider demographics in their marketing plans. For example, we found race to have an influence in wine consumption. These results suggest that whites and non-whites should not be assumed to have the same consumer behaviors. This information deserves special attention from wineries because the wine consumer demographics are evolving rapidly. Research shows that Latinos have been developing a taste for wine in recent years. In 1998, the percentage of Hispanic adults who consumed domestic table wine was 12 per cent, which jumped to 22 per cent in 2003. A 2004 survey by the Wine Market Council also asked people if they were drinking more, less, or the same amount of wine as the year before. While wine consumption frequency increased by 11 per cent among whites, it rose by 31 per cent among Hispanics (Jung, 2005).The generalizability of our findings may be limited for the following reasons. The findings may be idiosyncratic to the specific sample we collected in a non-probabilistic fashion in Northern California. Future studies need to use random sampling method, cover additional geographic areas, and include a larger sample size. Second, the CATREG procedure can be further validated by examining the consistency with alternative methods such as non-linear principal components analysis, non-linear canonical correlation analysis, and correspondence analysis, all of which can handle non-metric data. Any extension will be more meaningful if it also includes additional variables that connect wine consumers' behavior with different wine related attributes and marketing mix elements. Further research could also study the simultaneity between knowledge and consumption by including regression analysis of panel data to capture the shifting dynamics of wine market. Given the current dynamics of wine market, it is essential that the consumption behavior of US drinkers be understood in more depth. In that sense, the findings of this study offer a solid starting point for future studies of US wine consumers.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to identify econometrically the determinants of wine consumption of US consumers.\nDesign/methodology/approach: - In empirically identifying driving forces of wine consumption, we used 122 survey responses from Northern California consumers.\nFindings: - The study found that even knowledgeable or frequent consumers of wine purchase across all price points. Further, a significant positive correlation exists between knowledge and volume of wine consumed. All three regression techniques applied in this paper indicate that knowledge remains the most important determinant in wine consumption.\nPractical implications: - The results emphasize the need for US wineries to better educate and connect with consumers by developing compatible positioning strategies and marketing programs that are as informative as they are appealing.\nOriginality/value: - As one of the few studies of the US wine market employing econometric analysis, this paper offers a fresh perspective on the consumption behavior of wine drinkers in the USA.\n",
"role": "assistant"
}
] |
Construction of international brand portfolios: impact on local brands
|
[
"Brand portfolios",
"Brand management",
"International",
"Reduction",
"Global",
"Local",
"P&G",
"International business"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Marketers have traditionally focused on managing their brands individually. Most recently however the emphasis has shifted and with a tendency for companies to consider these issues in the context of more generalised brand portfolio management.After several decades of exponential growth of their international brands portfolios, many multinationals firms today are leaning more towards rationalization and reduction. They usually choose a \"top-down\", approach consisting of taking decisions at the level of the international portfolio that impact on the local portfolios.Thus, since the early 2000s, Unilever has embarked on the disposal of 1,200 brands from its original international portfolio of 1,600, L'Oreal has built its success on only 17 international brands and Procter & Gamble has reduced its portfolio to 300 brands of which 17 earn $1 billion or more.Whilst this practice of rationalization seems to be justified by the profound changes taking place in the economic environment, it must be conducted with care to avoid risks.In the literature, the question of how the strategies of brand portfolios affect performance remains open and subject to contradictory developments (Bordley, 2003; Kumar, 2003).This article highlights the various steps involved in the process (identification, classification, implementation and assessment) as well as issues specific to each of these phases. It relies in particular on the results of a longitudinal case study conducted in collaboration with the marketing direction of the Procter & Gamble's European Business Unit \"Laundry and Fabric Care\" from 2004 to 2009.\n1. The evolution of international brand portfolios: The brand portfolio of a company consists of all brand names of products or services that this company offers worldwide. The concept of the international brand portfolio is therefore positioned at the intersection of, on the one hand, markets where the company operates, and on the other hand products/brands offered by it together with the countries in which the brands are present. Reflections on international brand portfolios are therefore based on both the analysis of strategic business units within the company, the markets/brand dimension and the international deployment thereof.A historical analysis of international brand portfolios of large multinational organisations, based on the work of Kapferer (2004), leads to the conclusion that two major periods followed one another: their growth and their reduction.1.1 The growth of international brands portfolios\n2. The analysis of international brand portfolios: Based on recently completed projects in large multinational and research studies that have been conducted (Kumar, 2003; Aaker, 2004), analysis of international brand portfolios can be presented as an iterative process which is accomplished in several steps: identification, classification, implementation and assessment.We will present in this part the strategic and operational movements that led to the reduction of the P&G European brand portfolio in the laundry category. We will then compare them with the current results of the company in this market.At the initial stage of thinking about its brand portfolio in the laundry detergent category, Procter & Gamble had seven brands in France. Starting from a base of four brands (Ariel, Dash 2en1, Vizir and Bonux) the group purchased two additional brands from Colgate-Palmolive in late 2003 (Axion and Gama) and then, in late 2004, added one further brand (Mr. Propre) which had been extended into the detergent category (Figure 1).2.1 Identification\nConclusion: The article highlighted the fact that following a period of relatively uncontrolled development of their international brands portfolios, companies then decided to take a much more rigorous approach to this highly strategic area of management. They initially focussed heavily on reducing their international brand portfolios, because of the opportunities in terms of economies of scale, profitability and value creation. However, they found that this practice was not without risks, the main threats are the sometimes unnecessary elimination of local brands, the implementation of potentially dangerous changes in brand names and the over extension of global brands.We believe that companies should make improvements in the management of their international brand portfolios and that this requires the establishment of a good balance between development and reduction in brand numbers, good coordination between the purely \"top-down\" or \"bottom-up\" practices and a good balance between strong local brands and global brands.As we have mentioned, the best way forward has not yet been specifically defined and many questions remain. These questions are definitely worth pursuing in the context of very promising research on brand portfolios.\n",
"role": "user"
},
{
"content": "Purpose: - In the literature, the question of how the strategies of brand portfolios affect performance remains open and subject to contradictory developments. This paper aims to highlight the various steps involved in the analysis of international brand portfolios as well as issues specific to each of these phases.\nDesign/methodology/approach: - This article relies on the results of a longitudinal case study conducted in collaboration with the marketing direction of Procter & Gamble's European Business Unit \"Laundry and Fabric Care\" from 2004 to 2009.\nFindings: - The authors present the strategic and operational movements that led to the reduction of the P&G brand portfolio in the laundry category. The authors then compare them with the current results of the company in this market to assess the performance of this strategy of rationalization.\nOriginality/value: - While the best way forward to construct international brand portfolios has not yet been specifically defined and many questions remain, this article provides an illustration of a methodology tested by an international company.\n",
"role": "assistant"
}
] |
Time rationalities: Complementing "the orientation from the present" and the focus on "short" and "long terms" in performance measurement
|
[
"Performance measurement",
"Time orientation",
"Time rationality"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The time orientation of performance measurement has been studied extensively in the accounting literature (Merchant, 1990; Dechow and Sloan, 1991; Chow et al., 1996; Van der Stede, 2000; Abernethy et al., 2013; Pugh, 2017). Much of this research relies on the ideas that the formal periods within the performance measurement systems such as the evaluation period (the period considered when evaluating a given manager's performance) and the overall rigidity of performance measurement are related to the length of the time orientation of those persons assessed by such systems. For example, Merchant (1990) and Van der Stede (2000) warn that rigidity in performance measurement systems can lead to short-termism. Van Rinsum and Hartmann (2007) show how managers' time orientation is generally positively impacted by the evaluation period in the performance measurement systems to which those managers are subject.\n2. Theoretical reflections: 2.1 Linear-quantitative view on time and \"the orientation from the present\"\n3. Method: This study relies on social constructivism (Berger and Luckmann, 1966) appropriate for studying multifaceted concepts such as time. However, referring to Latour and Woolgar's (1986) proposition of the redundancy of the \"social\" as any entity can be seen from a social point of view in such research, references to \"social constructs\" have been minimised, thus predominantly employing the term \"constructs\".\n4. Empirical investigation: In order to clarify the time constructs early on, Triplex is discussed before Singlex. Background information on each company is first provided, and thereafter an episode from the history of the company is presented to illustrate time rationalities. Interviewees' perceptions of financial market listing are also presented and it is shown that attitudes towards financial markets may be partial sources for the observed time constructs. Performance measurement in each company is then further analysed and tied to the time constructs.\n5. Discussion and conclusions: Previous research has perceived a fairly clear correspondence between the technical time orientation of performance measurement systems and the resulting time orientation of executives (Merchant, 1990; Dechow and Sloan, 1991; Chow et al., 1996; Van der Stede, 2000; Graham et al., 2005). In contrast, this paper shows time rationalities that can be rather more closely connected to performance measurement than the traditional long-term/short-term distinction. It could be said that we should not naively believe executives when they say they do not follow their performance measures in a short-termist way; however, this is not seen here as an issue of naivete as the point is not that the executives were trying to claim that they were long-term oriented despite the measures, but instead that they were looking at the entire topic of performance measurement through the lens of time rationalities, not predominantly through the short and long terms.\n",
"role": "user"
},
{
"content": "Purpose: The temporality of performance measurement systems has been claimed to affect actors' time orientation, such as that of listed company managers. The purpose of this paper is to explore this view.\nDesign/methodology/approach: The study uses constructivist data gathered from executives in one listed and one non-listed company.\nFindings: The study shows that the research on performance measurement is based on a linear-quantitative view on time that assumes that humans orient towards the future from one point, the present; this view excludes other time-related constructs, particularly the past, and highlights a choice between the short term and the long term, idealising the long term. It is shown that the performance measurement of non-listed company executives is constructed through past-based, present-based and future-based rationalities: executives acknowledge the past as a basis for present and future performance, present actions as shaping future performance and future plans and performance targets as bases for present actions. Listed company executives' performance measurement is constructed predominantly through the present-based time rationality.\nResearch limitations/implications: \"The orientation from the present\" and the \"short\" and \"long terms\" could be enhanced with time rationalities.\nPractical implications: The evaluation periods within performance measurement systems do not determine the time orientations of the actors subjected to those systems; time rationalities could be considered when designing such systems.\nOriginality/value: The paper provides a novel view on performance measurement and time.\n",
"role": "assistant"
}
] |
Hospitality quality: new directions and new challenges
|
[
"Customer services quality",
"Hospitality services",
"Hotels",
"Customer satisfaction",
"Leadership",
"Employees"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Hospitality organisations are in the relatively unique position of attempting to bridge two worlds - the domestic and the commercial - as they deliver on the service promise. Lashley (2001) described this as an attempt to \"square circles\" as employees at the behest of their managers must provide some degree of hospitality but balance that with the requirement to be efficient and profitable. Thus, while numerous writers have convincingly argued that the delivery of service is different from the production of goods, we argue that the delivery of hotel services is qualitatively different from the delivery of other types of services, and that this will influence the way that service is managed, delivered and perceived by the customer. This paper will focus on issues of quality in the hotel sector identifying gaps and highlighting new directions.\nHospitality: The word \"hospitality\" is often used to describe the rather broad field that incorporates lodging, food service, leisure, conventions, travel, and attraction (Ottenbacher et al., 2009). It may also be used descriptively to explain the way in which one person behaves to another. Pizam and Shani (2009) highlighted the distinction between hospitality and hospitableness, arguing that the latter refers to authentic kindness and generosity while the former refers to the creation of experiences. They argued that managing this tension between creating the feeling of being at home while simultaneously trying to create an extraordinary experience is a challenge faced by management. It may mean striving for the near impossible according to Sherman (2007), who determined that the ideal luxury hospitality experience may involve service beyond that of a mother, and require instead the services of an \"idealised mother\".\nCustomer's perceptions of quality: The hospitality industry, and hotels in particular, have witnessed increasing competition for high service quality and customer satisfaction (Parayani et al., 2010). Hotel organisations therefore need to understand with as much precision as possible what the guests want from the service experience. Doing so allows them to minimise wasted resources. Being precise becomes difficult, however, because of the heterogeneity of customers. A number of writers have attempted to answer the very important question: what type of service do hotel guests value the most?\nHospitality quality: Service and quality have been elusive concepts (Van Hoof, 2002), and many researchers have had difficulty defining them separately, let alone as a combined concept. Service quality is seen as an important strategy for gaining a competitive advantage in firms (Lewis, 1987; Hoffman and Bateson, 1997) and therefore the management of these unique features becomes important in the deliberations about the delivery of service quality (Bateson, 1995; Fridgen, 1996).\nThe hospitality employee: While early work in service quality management was steeped in Taylorism and saw the employee as an unthinking cog (Wyckoff, 1984), views have changed and the performance of employees in hospitality is seen as a critical dimension of quality (Wong et al., 1999). As far as the customer is concerned the employee\nManagement: Leaders of the organisation establish the framework in which employees and customers operate. Researchers have therefore attempted to understand how various leadership traits and behaviours influence service quality. Kim et al. (2009) have built on the work of others (Ahmed and Parasuraman, 1994; Hartline and Ferrell, 1996) and advanced the debate on the role of managers in the process of delivering quality service. They argued that management commitment to service through initiatives such as organisational support, rewards, empowerment, and training will lead to job satisfaction and ultimately the delivery of high quality service by front line employees. Moreover Clark et al. (2008) determined that a manager\nConclusion: The contribution of this research to the quality service management and hospitality fields is in the area of hotel service quality. Figure 1 shows the relationships that have been covered in this article, as well as what the authors argue is the key to achieving hotel guest satisfaction at the very minimum.\n",
"role": "user"
},
{
"content": "Purpose: This article aims to discuss issues related to service quality in the hotel industry. It highlights unique aspects of hotel work and the implications for service delivery, and discusses methods used to measure service quality and expectations.\nDesign/methodology/approach: Through an assessment of the extensive body of literature on quality service in general and within hotel contexts, some key themes emerge which have been explored in detail. Much of the analysis focuses on how the body of knowledge has developed over time and points to strengths and weaknesses in the literature.\nFindings: There has been much debate about what constitutes quality service and more specifically how this applies to the hotel industry. The industry is a dynamic one in which the guest dictates the pace and type of service, and in which increasing competitiveness has resulted in satisfactory service being the minimum expectation of guests. Different measurements of service expectations have been proposed for the hotel industry but the most promising is the quality function deployment (QFD) approach, which treats service as a process in which the guest\nPractical implications: The highlighted differences between hospitality, hotel and other service organisations indicate where managers and researchers need to place their emphasis in order to enhance the level of service quality.\nOriginality/value: The paper adds new urgency to the need to develop effective measures and understanding of the hospitality and in particular the hotel experience.\n",
"role": "assistant"
}
] |
Distance, online and campus higher education: reflections on learning outcomes
|
[
"Blended learning",
"Summative grades",
"Education",
"Online education",
"Sweden",
"United Kingdom",
"Learning methods"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The paper investigates the use of information and communication technology in education. The use of information technology is today an integrated part of higher education (Ryan et al., 2000; Hillman and Corkery, 2010; Lockwood and Gooley, 2001; Moloney and Oakley, 2010; Rovai and Downey, 2010). Between the 1970s until after 2000 conditions in teaching and learning changed and adapted to demands from students and from teaching institutions for greater flexibility (Lockwood and Gooley, 2001; Ausburn, 2004). Teaching students utilising virtual learning environments (VLE) is becoming normalised and accommodated by both on-campus and off-campus learners (Moloney and Oakley, 2010). As a result, teaching in higher education is fluid, marked by a transformation process where online education is becoming more prevalent for many academic disciplines. However, some studies imply that distance learners are often \"better\" students (Means et al., 2009) it is not known if this type of education is cost effective, or if it actually improves the learning experience for either student or lecturer (McPhee, 2009; McPhee et al., 2010). In the literature, numerous studies comparing student achievement indicate no differences between study modes (e.g. Kessler, 2007; Russell, 1999; Schoenfeld-Tacher et al., 2001). Extensive and rapid technological developments in the field create new educational conditions for both on-campus and off-campus courses, suggesting a need for continuing research (Dixon et al., 2005; Harlen and Doubler, 2004; Oliaro and Trotter, 2010; Zapalska and Brozik, 2006). For example, Solimeno et al. (2008) suggests that the application of cooperative and collaborative teaching models in online education is related to new technologies facilitating social interaction. Garrison (2009) claims that rapid technological development brings forth beneficial changes in teaching design and implementation. A more inclusive and engaging two-way communication between learner and institution have massively improved the early self-instructional correspondence packages. Moreover, research comparing different study modes has not fully acknowledged longitudinal studies and studies that address development issues (e.g. Caywood and Duckett, 2003; Oliaro and Trotter, 2010; Phipps and Merisotis, 1999).Drawing on two cases, one from Scotland and one from Sweden, the paper contributes to quality issues in distance, online and campus education by taking into account, in the first case, different student performance in the same course over a longer period and in the second case, changes over time within the same educational programme. Drawing on two cases, we consider three themes: first, differences in students' performance by study mode; second, differences in students' performance by length of study; and finally comparing performance by study mode between modules in Scotland with an entire programme in Sweden.This paper draws on empirical data from distance, off-campus and on-campus study groups in Scotland and Sweden with exactly the same module syllabus. We define online education as learning accessed exclusively through the internet, while distance education is a blended mix between face-to-face and online education. Evaluating different study modes is complicated by the fact that both courses and study modes vary (Oliaro and Trotter, 2010). This means that examples from the different empirical settings will be used to illustrate and discuss the questions addressed in the paper. The University of the West of Scotland has been offering flexible postgraduate programmes in Alcohol and Drugs Studies online since 1999. Umea University has been offering distance and flexible courses and programmes since the 1980s in Sweden and in this paper the human resource management study programme will be discussed.\nOnline vs campus education: The predecessor to online education, distance education, has often been seen as the second best alternative to traditional university education (Forsyth et al., 2010). This tension between distance and on-campus modes of teaching and learning has created a debate whether distance or online education is as effective as face-to-face campus learning (Oliaro and Trotter, 2010; Schoenfeld-Tacher et al., 2001). A rudimentary ingredient for this discussion is the absence of paralinguistic cues in online education, which some consider essential for learning to occur (Tolmie and Boyle, 2000). The absence of physical proximity can make it easier for less vocal students to communicate since perceived or actual barriers to face-to-face conversations are absent or minimised in online communications whether synchronous or asynchronous. But this absence of paralinguistic cues could also potentially contribute to negative educational experiences. Price et al. (2007) indicate that online students perceived their online learning as less engaging and effective than on-campus students. They conclude in their study that both lecturer and student require training to compensate for the lack of communication cues in online learning contexts.Historically the spread of distance education contributed to the increase in access of part-time students in higher education which raised quality assurance issues and a concern with the deterioration of standards in academia (Caywood and Duckett, 2003). However despite such concerns, many studies conclude that there were no significant differences between campus and distance education concerning learning outcomes. For instance, Russell (1999) reviewed 355 studies on distance education produced between 1928 and 1998. Some of the early studies examined correspondence courses, but most compared instruction over videotape, interactive video or satellite with on-campus, in-person study programmes. The comparisons were based on test scores, grades or performance measures unique to the study and on learner satisfaction. The use of the internet has contributed to a shift from distance to online education which has penetrated into all levels of education. Research focusing online and on-campus education in general terms tends, similarly to the review conducted by Russell, find few significant differences in outcomes and satisfaction ratings between on-campus and off-campus learners (e.g. Edwards et al., 1999; Duffy et al., 2002; McPhee et al., 2010). This is also noticed by Kessler (2007) who compared 176 students studying online and on-campus found no significant differences in grade scores between study modes. Similarly, Caywood and Duckett (2003) found no differences between online and on-campus in regards to learning in a course of behaviour management. Schoenfeld-Tacher et al. (2001) illustrated in their study that online learning was at least equivalent to on-campus course concerning content learning opportunities. On the other hand Connolly et al. (2007) found that online students performed better than on-campus students in their study. Likewise, Harlen and Doubler (2004) concluded that online students understanding of the science content in a course designed for using science enquiry skills was significantly greater than the on-campus students. While some meta-analysis studies, principally the analysis by Phipps and Merisotis (1999), concluded there was no significant difference, they did find a significant variation in the outcomes of distance education and face-to-face education. However, they also point out that any evaluation research in this area tends to focus on one small part of an entire study programme (Phipps and Merisotis, 1999). Recent research of the campus vs online dichotomy made by Tsai (2009) explored college students' conceptions of learning in general terms with their conceptions of web-based learning. The result from the study illustrated that conceptions of web-based learning were more sophisticated and comprised higher order thinking. Zhao et al. (2005), found that low instructor involvement led to less positive outcomes for distance education but more positive outcomes as instructor involvement increased. Likewise, Shaw and Young (2003) found that lecturer activity tends to increase if equivalency is to be achieved. In line with Hrastinski and Keller (2007) argument that more research is needed on different modes of learning this paper will make a contribution to the research by investigating quality issues between study modes (distance, online and on-campus) and between countries (Sweden and Scotland).\nMethods: The findings presented in this paper are based on empirical data from higher education in Scotland and Sweden. The aims, content and assessment demands were similar in both cases. The Scottish case is based on evaluation of a postgraduate course in addiction studies where online study groups access exactly the same module syllabus as their on-campus counterparts. Students on both modes of study are taught on the same traditional 15-week trimesters as students on the on-campus version. They are assessed in exactly the same manner.In the Swedish case, the study programme in human resource management is a three-year long education programme examining personnel, staff and organisational issues. The programme is an on-campus programme with VLE support. Three different study groups (the distance group) located at three different places in northern Sweden pursued full-time studies together with the on-campus group.Data collection in Scotland\nFindings: Results from Scotland\nDiscussion and conclusions: The findings show that, despite a rapid technological development which the latest years has changed teaching and learning conditions, both the Swedish and the Scottish data tentatively suggest that it is not possible to claim that campus education is a better educational option than distance or online education. In general, there are no differences between the grades or test scores between campus students with face-to-face education and distance students with electronically (VLE)-mediated education. The findings are consistent with the literature regarding whether there are any differences between distance, online and on-campus education. The Scottish data show that there are not any differences between different study groups reading the same course online or on-campus over the years. This may be due to strong attempts to create a comparable learning experience between the online and on-campus student cohorts (see McPhee, 2009). These findings imply that the online learning environment, as constructed in this particular case, serves as a suitable, comparable setting for students to learn and achieve grades appropriate to their abilities, when compared to the traditional on-campus supported learning environment. Concerning the Swedish case, the findings show that there were differences between distance and campus students in the beginning of the human resource programme. However, these differences did not occur in courses after one year of the programme. The distance students performed equally on the examinations as the campus students.The result from the Swedish case indicates that a comparison between different study modes has to consider several courses to cover the whole study programme. A sample taken initially in a programme to investigate different study modes might give results that may not be related to the study mode. For example, the distance students were on average older and had less experience of studying in higher education as the on-campus students. The less study experience and also that there are several years since the studied might be one factor that explains why distance students performed less well than the on-campus students in the beginning of their education. When the less experienced distance students become more familiar with studying in general and the distance format the initial differences disappears.In this paper we considered three themes: first, differences in students' performance by study mode; second, differences in students' performance by length of study; and finally comparing performance by study mode between modules in Scotland with an entire programme in Sweden.There were no significant differences in student performance either by study mode, length of study or between countries.To conclude, both cases indicates that utilising a distance learning model for students does not adversely influence grades scores, and this suggests that it is appropriate using both on-campus and online delivery methods without any negative impact on student performance.Limitations and future research\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to discuss performance in postgraduate education in Sweden and Scotland. Drawing on two cases, the paper considers three themes: differences in students' performance by study mode, differences in students' performance by length of study, and finally comparing performance by study mode between modules in Scotland with an entire programme in Sweden.\nDesign/methodology/approach: - The empirical setting from Scotland builds on an evaluation of online and on-campus study groups with exactly the same module syllabus. The Swedish setting is also based on an evaluation of distance and on-campus study groups with exactly the same module syllabus. The data compiled in both countries arise from student performance scores and grades.\nFindings: - The results indicate that students in both countries foremost use the virtual learning environment (VLE) as a forum for accessing information, to access asynchronous postings in the forums and access streaming-synchronous online lectures which are also accessed asynchronously in the VLE. The results show that there are no differences between the grades or test scores between campus students with face-to-face education and distance students with electronically (VLE) mediated education. These differences and similarities will form the basis of these reflections in this paper.\nResearch limitations/implications: - The study is limited because the examples given are only a few cases and small samples and there is a need to more rigorously investigate different educational programs in different academic disciplines.\nOriginality/value: - The paper contributes to quality issues in distance, online and campus education by taking into account, in the first case, different student performance in the same course over a longer period and in the second case, changes over time within the same educational program.\n",
"role": "assistant"
}
] |
The employability skills of business graduates in Syria: Do policymakers and employers speak the same language?
|
[
"Higher education",
"Syria",
"Employability skills",
"Business graduates"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: For most countries, the dilemma of university graduates' employability becomes a critical issue involving different social, economic, cultural and national aspects (Asonitou, 2015; Pavlin and Svetlik, 2014; Garrouste and Rodrigues, 2014; Farcnik and Domadenik, 2012; Gokuladas, 2010). To satisfactorily reach a level of employability that is nationally accepted, local governments adopt national educational strategies directed either towards market needs, national needs or a mix of both types of needs. This conflict produces, subsequently, conflicting beliefs in other aspects of educational policy such as national admissions policies to universities, educational providers, curricula issues and university graduates' employability skills.\n2. Review of literature: Different studies have referred to the difficulties and challenges facing higher education systems with regard to preparing employable graduates (Natalia et al., 2015; Tarvid, 2015; Abou-Setta, 2014; Sung et al., 2013; Clarke, 2008; Holmes, 2008; Little, 2008; Wilton, 2008). In exploring the relationship between skills developed in undergraduate programmes and those subsequently used in employment, particularly in managerial careers, and drawing on data drawn from a questionnaire survey of 1,999 graduates, Wilton (2008) raised concerns about the extent to which higher education is able to adequately prepare graduates for employment.\n3. Syrian higher education priorities for employability skills: In accordance with Article No. 29 of the Constitution of the Syrian Arab Republic, education is granted for all people. Education is free for all levels and is obligatory from primary level to preparatory and secondary levels. There are two ministries concerned with education in Syria. The Ministry of Education governs the educational process up to the secondary level, and the Ministry of Higher Education governs tertiary education and lifelong learning.\n4. The study approach: Based on the study aims and questions raised earlier, the authors employed the main source of qualitative data through interviews. Interview data are one of the important sources of data (Heu et al., 2015; Yang et al., 2015; Silverman, 1999). Silverman (1999) argued that words are important simply as a jumping-off point for the real analysis; he suggested that where texts are analysed, they usually are presented as \"official\" or \"common sense\" versions of social phenomena. In relation to the aims of this study, the authors depended on the written priorities of the higher education sector in Syria and the written sector priorities as they appear in the official state plan. This stage of the documentary data was useful to the authors as it helped in drawing up the initial understanding of the research questions. This stage also provided wider insight to the research, which has been added to its final stages.\n5. The results of the interviews: Interviews with senior managers in the private sector\n6. The matrix of business graduate type: Based on the previous results and in response to the third research question, the study indicates that while higher education policymakers believe that business graduates should have skills that are mostly socially and ethically oriented, managers from the private and business sector believe in acquiring individual and technical skills that may benefit the business. The previous scenario points to the traditional dispute between public thinking and private thinking, which is reflected clearly in the managers' opinions regarding the skills of business graduates. However, the state plan which was identified earlier may draw up a compromise scenario that is neither public nor private. To compare the previous two sources of data, the authors developed the matrix of business graduate type (see Figure 1).\n7. Conclusion, and implications for future studies: This study explores the two perspectives of higher education and the business sector with regard to business administration graduate requirements. The study ends up with the development of a qualitative comparative scenario that aligns the two perspectives by identifying four styles of graduates: the leader style, the collective-manager style, the technical-manager style and the trainee-manager style.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to compare the desired employability skills of business graduates in Syria from the perspective of both higher education policymakers and employers in the private sector.\nDesign/methodology/approach: Interviews were conducted with 12 higher education policymakers and managers from the business sector. Content analysis was utilized to analyse the content of the interviews and the strategic priorities of the higher education sector in Syria.\nFindings: Results revealed that although higher education policymakers focus more on societal, public and thinking skills for business graduates, the business sector focusses more on individual, private and practical skills. Accordingly, a comparative tool that aligns the two perspectives was developed in the study. The tool, based on the contradicting employability skills, identified four types of business graduates: leader, collective, technical and trainee.\nResearch limitations/implications: The study is limited by data collected before the current political instability in Syria in 2012. The data were collected only from official documents and interviews with policymakers and employers. Students were not part of the study.\nPractical implications: The managerial tool developed at the end of the study will help both policymakers and the private sector to statistically allocate business graduates for better planning. The study provides recommendations to the different stakeholders in the higher education sector in Syria.\nOriginality/value: Although the majority of the previous literature raises the voices of the business sector, this study is one of the first studies that aligns the discrepant perspectives of the higher education and business sectors. The managerial tool developed in the study is original and usable by policymakers and the business sector, and it is subject to further development.\n",
"role": "assistant"
}
] |
"Green" wine packaging: targeting environmental consumers
|
[
"Consumer behaviour",
"Wines",
"Packaging",
"Marketing strategy",
"Green marketing",
"United States of America"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The last two decades have witnessed a dramatic shift in environmental consciousness worldwide. Consumers have begun to change their behavior integrating environmental considerations into lifestyle choices. This change includes consumers' purchasing decisions based upon how well products satisfy their needs and how these products affect the natural environment (GFK, 2007).As people are becoming more aware of the damage caused to the environment, environmental concerns of industry have been identified as a critical issue that packaging companies (Amorim, 2009; Nomacorc, 2008), lodging (Manaktola and Jauhari, 2007; Mensha, 2006), tourism destinations (Weaver and Lawton, 2004), and wineries (Baughman et al., 2000; Marshall et al., 2005; Chiffoleau 2005) must contend with. However, the reason behind consumers' adoption of sustainable practices, attitudes, and intention to purchase toward environmental products remains largely unexplored, particularly in the context of wineries. For example, the wine industry is perhaps the most vulnerable of any food and beverage producer to \"carbon criticism,\" due to its historical reliance on heavyweight glass packaging and its slow movement toward more environmentally sensitive packages (Rocchi and Stefani, 2005).Yet, the environmental performance of the wine industry does not receive as much media attention as industries often characterized as \"dirty.\" These issues include the application of toxic pesticides, herbicides, and fertilizer; the use of scarce water supplies; the creation of contaminated wastewater run-off; organic wastes; and the consumption of non-hazardous packaging materials (Marshall et al., 2005; Baughman et al., 2000).At the same time, the global wine industry is growing more competitive, and marketers are seeking ways to differentiate their brands to consumers. How consumers perceive wine as a product and its characteristics is an essential factor in the decision process and to a brands success. However, with the choice of wines numbering in the thousands, when consumers begin selecting a particular wine over another is still unknown. Companies promoting increased environmental awareness believe information will lead to increasing environmental knowledge, thus changing attitudes and buying behaviors (Peattie, 1995). Arcury (1990) and Bazoche et al. (2008) have suggested that knowledge, information, and attitudes are important for changing human actions toward the environment.Barber et al. (2009) stated using segmentation through selective marketing techniques, following Inskeep (1991), may attract environmentally oriented consumers who respect the environment and who are conservation-minded. Such segmentation considers research studies that used variables such as gender (Barber et al., 2009; Mitchell and Walsh, 2004), \"geographic origin\" (Guagnano and Markee, 1995), age (Dietz et al., 2002), knowledge (Barber, 2008), and attitudes (Bredahl, 2001; Eagly and Chaiken, 1993; Frick et al., 2004).Therefore, based on the preceding discussion, the aim of this study was to profile consumers who expressed an intention to pay more for environmentally friendly wine packaging and to elaborate marketing strategies that arise from this consumer profiling. This study contributes to the understanding of wine consumer expressed purchase intention in two ways. First, by investigating which characteristics of consumers are related to environmental behavior and second whether consumers express their intention to spend more for wine that is environmentally friendly. Developing a better understanding of what causes individuals to pay more for environmentally friendly products; the greater the opportunity marketers will have to develop strategies specifically targeted at these consumers.\nLiterature review: Consumers appear to have mixed opinions about the environment. According to a 2002 Gallup Poll (Carroll, 2002) Americans were closely divided about the quality of environmental conditions, but were more definitive when asked about the future, with 54 percent of adults, saying the quality of the environment is getting worse, while 40 percent say it is has improved. Three years later, a Gallup Poll (Saad, 2005) of residents in three countries described the overall environmental quality as \"only fair\" or \"poor,\" rather than \"excellent\" or \"good.\" Combined results for only fair/poor figure was 67 percent in Great Britain, 58 percent in the USA, and 52 percent in Canada.Consumers have exhibited preferences for environmental services and products, either directly through polls or surveys, or indirectly, by participating in outdoor activities, environmental causes, undertaking conservation through recycling, or purchasing environmentally friendly products (e.g. boxed wine in recycled material) (Laroche et al., 2001; Torgler et al., 2008). However, when consumers are confronted with decisions to purchase environmentally friendly products, the results were mixed. Bazoche et al. (2008) found in their study of French wine consumers, that wines with an environmental signal were valued similarly as conventional wines, while Loureiro (2003) found by testing Colorado wines consumers in some cases were not willing to pay more for the environmentally friendly wine when quality was a perceived difference.On the other hand, Brugarolas Molla-Bauza et al. (2005) found in their study of Spanish consumers that the average premium consumers expressed a willingness to pay for an organic wine was 16 percent above the price of a standard wine, while D'Souza et al. (2006) found in their study of Australian consumers, that despite 60 percent of the respondents having the perception environmentally safe products would be more expensive than conventional products, nearly 70 percent indicated they would purchase these products even if they were more expensive than the alternatives. Finally, in a New Zealand study, nearly 73 percent of respondents specified they would be willing to pay more for an environmentally sustainable wine and 33 percent stated they were prepared to pay a premium of up to 5 percent for sustainably produced wines (Forbes et al., 2009). Thus, consumers may allege to be \"green,\" but uncertainty exists if they are truly willing to purchase goods based solely on environmental grounds, particularly if quality is an issue.Vermeir and Verbeke (2006) suggest buyers are engaging in decisions about everyday consumption as well as sustainable purchase consumption. Everyday consumption is driven by convenience, habit, value for money, personal health concerns, and individual responses to social and institutional norms. Whereas sustainable consumption is based on a thoughtful decision-making process that considers social responsibility in addition to individual needs and wants (Vermeir and Verbeke, 2006). One possible opportunity of incorporating environmental responsibility into consumer behavior and related package design and planning is to attract consumers who are fundamentally interested in protecting the environment and consequently behave in ways that lead to a smaller ecological footprint.Environmentally conscious consumers\nMethodology: Design of the study\nResults and discussion: Data analysis\nDiscussion and managerial implications: The results of this study provide theoretical and practical contributions to the understanding of determinants for consumers' intention to pay more for environmentally friendly wine packaging.Impact of attitudes\nConclusion: The major contributions of this study were the investigation of the profile of consumers expressing their intention to pay more for environmentally friendly wine packaging and the subsequent discussion of the implications for marketers. The results of this study are a stepping stone toward future research in a field growing in theoretical and practical significance. The world-wide increase in media coverage of the environment and the issues facing the planet, the marketplace expansion of environmentally harmonious products, and the integration of ecological issues in educational and political systems suggests the need for more research on defining who the environmentally conscious consumers are.Consumers expressing an intention to pay more for green products in general, and specifically wine packaging, have suggested that our ecological problems are serious and that behaving in an ecologically favorable manner is important and not inconvenient. On the contrary, consumers expressing they would not pay more for green wine packaging reported they do not consider ecological issues when making purchases. These consumers attributed a lower score to the severity of ecological problems and to the importance of being ecologically friendly. Thus an important approach would be developing new strategies for ecological marketing by redirecting consumers' needs and wants toward environmentally friendly wine products, such as organic or biodynamic produced wines, and reorientation of the product mix through repackaging and re-labeling. For example, Robert Sinskey Vineyards uses bio-dynamic wine processes yet does not promote this in their general advertising strategy. It could be, as stated by Schlecht (2009), that there is a perceived quality issue. Also, given the state of New Hampshire's new programs to \"green\" the retail stores, particularly given the control state liquor commissions exert over retailers, following their example could add real educational value to the environmental movement.Thus, the logical implications for marketing managers suggest they should appeal to the attainment of a collective environmental goal. What this means to the wine industry specifically, or other hospitality service industries in general, is that a different focus is needed in marketing, with a media approach directed toward different market segments, such as younger wine consumers. The idea of a \"generic advertisement\" approach may not capture the different environmental attitudes or gender markets; nor begin to expand through re-direction, the wine industry toward more environmentally friendly consumers.Opens in a new window.Table I\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to profile consumers who expressed an intention to pay more for environmentally friendly wine packaging and to elaborate marketing strategies that arise from this consumer profiling.\nDesign/methodology/approach: - The population under study was the general adult population of the USA. The sample was randomly selected from a consumer database and provided a link to an anonymous on-line survey. Reliability and factor analyses were performed to identify whether the measurement items reflected the constructs of attitudes, behavior, values, and intention to pay more. To determine significant differences, t-tests on each construct were performed, along with stepwise discriminant analysis.\nFindings: - The results indicate that importance of being environmentally friendly, considering environmental issues when making a purchase, and collectivism were all very good predictors of consumers' intention to pay more for green wine packaging.\nResearch limitations/implications: - The intention to purchase does not always translate into actual purchase behavior when consumers are confronted with a purchase situation.\nPractical implications: - This paper contributes to the understanding of environmental wine purchase intention by investigating relationships of consumer characteristics to environmental behavior. By segmenting consumers in this manner, it is possible to better understand the importance of environmentally friendly wine packaging, thus aiding wine producers, retailers, and packaging companies in directing their marketing and advertising efforts.\nOriginality/value: - This paper examines the significance of personality variables in detail using measurement scales that consider customers' perceptions and intention to purchase.\n",
"role": "assistant"
}
] |
Sorry for your loss, now when will we be paid? Customer service after death of a customer
|
[
"Customer service management",
"Financial services",
"Brand equity",
"Services marketing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: While there is near-universal knowledge of the adage \"there are no sure things except death and taxes,\" companies too frequently ignore the certainty of death in establishing customer service plans. One would indeed expect life insurance companies to have clear procedures and policies for handling claims. Similarly, financial services organizations that routinely establish accounts with ownership such as \"payable on death\" should have knowledgeable employees ready to help beneficiaries settle estates and reallocate funds. While any business that conducts transactions with consumers will be faced with the death of some of its customers at some point, too few recognize this as a marketing opportunity, instead choosing to focus on the cost side of the equation.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to explore how retail organizations serve heirs attempting to settle estates, suggesting that companies providing minimal customer service risk driving away future business, while those companies that seek solutions to ease the estate settlement process can acquire new customer relationships from heirs and beneficiaries.\nDesign/methodology/approach: - The paper provides examples of companies that adhere to strict rules, minimal service, or emphasis on cost reduction, comparing them to companies with more flexible and compassionate approaches to estate settlement.\nFindings: - An approach to customer service centered on meeting customer needs can enhance an organization's reputation. By contrast, those reliant on rigid policies risk losing long-term business opportunities.\nPractical implications: - Companies can gain new customer relationships by improving the way they handle estate claims and provide service to executors, administrators, heirs, and beneficiaries. Financial services companies in particular stand to acquire new customer relationships from the forecasted large generational wealth transfer.\nOriginality/value: - The paper provides guidance for improving customer service when settling estates, and argues that companies that do not seek positive ways to ease the difficulties in dealing with financial matters at death ignore the basic directives of a marketing perspective.\n",
"role": "assistant"
}
] |
Operations and supply chain management: The role of academics and practitioners in the development of research and practice
|
[
"Career development",
"EurOMA",
"Research competences"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: As a scientific discipline, operations and supply chain management (O&SCM) continues to look for practical relevance and theoretical impact for its research. Problems and opportunities in practice are diverse and call for theoretical bases and methodologies - often from different domains and schools of thought. O&SCM, systems and practices are evident in the design, operation and performance of manufacturing firms, service providers and public organisations. While practitioners and researchers develop a discrete understanding of this evidence separately, each needs the other in the generation of a shared understanding. It is here that there is a challenge. To dichotomise, both parties work to different timescales, abstract differently and understand each other's practice differently. Each has to train and to educate others both to think and to apply that thinking systematically and even creatively to the design, running and improvement of operations.\n2. The Young Scholars Workshop: The EurOMA has developed a clear strategy for the development of its members (largely academic) as contributors to the field of O&SCM. In particular, the education of its PhD students and the support of newly appointed faculty (Young Scholars) in their first academic appointments are central activities in the EurOMA agenda.\n3. How to increase the usefulness and relevance of O&SCM research?: Wickham Skinner is regarded as one of the founding fathers of modern operations management thinking. In his seminal Harvard Business Review (HBR) article, Skinner (1969) was the first to make the link between corporate and manufacturing strategy. Skinner based these observations on his employment over a decade with Honeywell. He was driven by a desire to improve the practice of OM by solving the problems that he witnessed first-hand. Over 40 years later, in his 2010 keynote address at the Decision Science conference, Skinner expressed concerns that OM was losing its industrial relevance, as academics focussed on publication (\"taking from the system\") and not solving the big problems of today. To quote Skinner (2010):We need to move away from a culture of extraction (taking from the system) and concentrate on building things and leave a legacy [...]. We need to focus on improving methodology and apply it to solving the \"big\" problems of today.\n4. Designing, conducting and publishing collaborative research: Useful and relevant O&SCM research is often collaborative. Young scholars have choices as they design, conduct and publish their collaborative research. For many, publishing from a recently defended dissertation may be on an immediate horizon filling up with new teaching assignments and research proposal development. It is in the development of new research proposals that a major choice emerges. Should the young scholar plan to research alone, with other researchers only, with practitioners only or with both researchers and practitioners? This last option opens the possibility of collaborative management research.\n5. Taking theory to practice - taking practice to research: Up to this point in the paper, the proposition is that young scholars may be given choices and take the opportunity to develop an academic career where they teach and research in O&SCM, possibly as collaborative researchers. Alternatively, they may opt for an industrial career and take all of their knowledge and skills to industry. This part of the paper focusses on the latter option, and exemplifies the dynamic interplay between theory, research and practice. The discussion is organised around four cases in which one of the YSW contributors has been involved over the past 20 years. The cases are summarised in Table VI and the interplay is illustrated in Figure 3. Except for case 1, the cases are set in the electronics or, more specifically, the semiconductor industry.\n6. Consolidation: 6.1 Implications for theory, research and practice\n",
"role": "user"
},
{
"content": "Purpose: Based on the combined experiences of operations and supply chain management (O&SCM) scholars and a reflective practitioner, the paper compares, contrasts and reconciles the competences needed to research O&SCM practice and to practice O&SCM research. The purpose of this paper is to locate these competences for young faculty in relation to their ambitions and career choices.\nDesign/methodology/approach: The paper is based upon the contributions made at European Operations Management Association 2014 Young Scholars Workshop (YSW). The theme and programme of the workshop was \"Operations management - research and practice\".\nFindings: The paper outlines first the concept of the YSW, the evolution of themes and the specific focus of the 2014 workshop. It concludes with a reflection on the career development of O&SCM scholars, their potential role, as academics or practitioners, in the development of O&SCM theory and practice, and the role of collaborative research in that development.\nPractical implications: This paper shows what it takes for O&SCM researchers to engage with \"the world around us\" involves and, vice versa, how \"doctorate\" OM and SCM practitioners may successfully solve practical problems and engage with the O&SCM world surrounding them in doing so.\nOriginality/value: The paper presents a collection of viewpoints of O&SCM scholars and a reflective practitioner on the competences needed to research OM practice and to practice OM research.\n",
"role": "assistant"
}
] |
Flexibility and crisis resistance: quantitative evidence for German establishments
|
[
"Labour market flexibility",
"Financial crisis",
"Flexible firm",
"Numerical flexibility",
"Labour",
"Labour market",
"Finance",
"Flexible organizations"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The impact of the global financial crisis (GFC) on the German economy, the subsequent reactions by companies, social partners and policy makers, as well as the relative resilience and stability of the German economy compared to other economies since 2008, have been extensively discussed elsewhere and have led to a debate about a second German Wirtschaftswunder or \"economic miracle\" (Burda and Hunt, 2011; Moller, 2010; Schneider and Graef, 2010).While there is an emerging academic and policy debate about the lessons learned from the GFC concerning macro-level policy options (European Commission, 2012), we call attention to micro-level policy and the potential conclusions that companies might draw from the GFC experience in order better to prepare for future crises. With respect to crisis resistance and sustainable competitive success in an uncertain environment, the extant literature has emphasized the role of labour flexibility both at the macro level (OECD, 1986a, b, 1989) and at the micro level (Atkinson, 1985; Atkinson and Meager, 1986), while more recent work emphasizes agility (Nijssen and Paauwe, 2012) and resilience (Hamel and Valikangas, 2003; Lengnick-Hall and Beck, 2005; Lengnick-Hall et al., 2011). Yet, with respect to the claims that flexibility is a precondition for agility, resilience and crisis resistance, Kaiser et al. (2005) note a lack of empirical evidence.In this paper we seek to analyse the following research question: to what extent does (labour) flexibility contribute to crisis resistance at establishment level? More specifically, we analyse the determinants of variation in the extent to which establishments showed resistance to the GFC, i.e., the extent to which they were affected by the crisis, focusing on an available secondary dataset related to organizational, industry level, and (numerical) labour flexibility.We believe that the identification of the factors associated with crisis resistance during the GFC may then inform the development of an organizational and human resources architecture that helps companies to enhance their adaptive capacity as a precautionary measure in the face of future environmental changes.The first part of the paper introduces, compares and contrasts existing theoretical and analytical approaches, focusing on labour market flexibility, the theory of the flexible firm, and recent perspectives on agility and resilience. Drawing on the German Job Vacancy Survey (IAB-Erhebung des Gesamtwirtschaftlichen Stellenangebots, EGS), the subsequent empirical part of the paper first describes the variation among German companies affected by the crisis in the spring of 2009 (when the adverse effects of the GFC peaked in Germany). It then analyses and discusses potential reasons that help to explain crisis resistance. The final section discusses the implications of our findings for both management and policy makers.\n2. The crisis, flexibility and resistance: theoretical considerations: Crises such as the GFC represent external shocks that alter the parameters within which companies navigate and operate. The initial shock wave of 2008/2009, as well as the continuing dynamism and volatility of the GFC, created a high degree of uncertainty within which governments and businesses take decisions on strategic and operational issues (Streeck, 1987).As to the operational challenges attributable to the crisis, first and foremost is the decline in product demand, with its associated decline in labour demand and subsequent labour surplus. Financial problems, such as those related to difficulties in obtaining or extending credit, may lead to downward pressure on (labour) costs. Labour surplus, as well as pressure on labour costs, requires companies to initiate and implement adjustment policies.From the management perspective, uncertainty as well as pressure to reduce (labour) costs at short notice has effects on the desirable organizational and human resource architecture at company level, putting a premium on labour flexibility (Atkinson, 1985; Atkinson and Meager, 1986), agility (Nijssen and Paauwe, 2012) and resilience (Lengnick-Hall and Beck, 2005; Lengnick-Hall et al., 2011). These concepts will be introduced and discussed in the subsequent sections, with distinctions between their macro- and micro-level dimensions.2.1 Macro level: labour market flexibility\n3. Flexibility and crisis resistance in face of the GFC: quantitative evidence: 3.1 Research question, dataset and statistical procedures\n4. Conclusions: This paper set out to contribute to the investigation of reasons why some establishments were either resistant to or not affected by the GFC while others were affected adversely, as reported by Heckmann et al. (2009). Drawing on macro- and micro-level perspectives on labour market flexibility (OECD, 1986a, b, 1989) and labour flexibility (Atkinson, 1985; Atkinson and Meager, 1986), as well as organizational agility (Nijssen and Paauwe, 2012) and resilience (Lengnick-Hall et al., 2011), we analysed the impact of different organizational characteristics, such as work-force characteristics, and industry characteristics as well as factors related to product markets and finance, on the likelihood that an establishment might be affected by the GFC in April 2009, at the peak of the crisis in Germany. The focus of the empirical analysis was on different types of flexibility, with a strong emphasis on numerical labour flexibility, as this was a factor considered relevant from all of the perspectives mentioned above.Admittedly, whether or not an establishment is hit by the crisis results from a complex array of factors, a comprehensive analysis of which is beyond the scope of this study. Yet, while the limited number of variables and the cross-sectional nature of the data imposed restrictions on the quality and interpretation of the analysis, the estimation results give an indication concerning the relevance of both labour market flexibility and establishment-level flexibility. Our findings suggest that measures of numerical flexibility are not significantly associated with crisis resistance, except for a 10 per cent significance of the proportion of employees outside social insurance. Instead, organizational and industry-related flexibility as well as problems with product demand and establishment finance represent active factors. Despite all the imperfections of our study, these findings are both relevant and, for a number of reasons, intriguing.First, our findings indicate a limited role for numerical flexibility in assisting crisis resistance at establishment level, especially in comparison to organization-related and industry-related factors. However, at this stage it is important to mention that the small effect size of our numerical flexibility measure may indicate that it is not so much the \"total level\" of numerical flexibility that is important as the \"optimal level\". Second, problems with product demand and finance at the start of the GFC have the highest odds ratios, indicating the general importance of the (pre-existing) economic situation for crisis resistance, and a greater magnitude in the effect of the economic situation by comparison with organizational and workforce-related factors. The relative importance of these dimensions (organization, workforce, product, finance) should be scrutinized further, as this may help management and public policy to identify priorities for increasing crisis resistance and resilience.Future empirical research should investigate, in addition, the link between resilience as \"the capacity to recover quickly from difficulties\" (Oxford Dictionaries, 2012a) and resistance as \"the ability not to be affected by something, especially adversely\" (Oxford Dictionaries, 2012b). So far as research into numerical flexibility is concerned, it may be advisable to look into types of flexibility other than fixed-term contracts and temporary agency work, and to probe into what kind of employment is hiding behind \"employment outside social insurance\".Until it is over, the GFC will continue to produce uncertainty for business. With reference to the theories discussed above, and in line with our empirical findings, companies and public policy makers may seek more flexible, adaptable, and adjustable organizational and institutional arrangements. With respect to the practical aspects of increasing establishment-level resistance as a precautionary tool, our study also has important practical implications. For management, the development of measures to improve crisis resilience should focus additionally upon product-related and finance-related factors rather than simply prioritizing workforce flexibility. Crisis resistance and resilience are comprehensive concepts that require cooperation and coordination across the different functions within organizations. The lesser importance of numerical flexibility may also be relevant for trade unions and workforce representatives in their discussions of, and their negotiations about, non-standard employment in the organization.At the public policy level, we argue that a discussion of the lessons learned from the GFC should also elaborate on company-level developments, and avoid focusing exclusively on macro-level policies. Public policy may otherwise miss out on important developments at company level, with serious implications in terms of the capacity to coordinate economic activities brought about by further deregulation, decentralization, and flexibilization of the labour markets. We recommend not only an investigation into the role of workforce characteristics (especially numerical flexibility) as drivers of crisis resistance and resilience, but also a broader and more comprehensive perspective which will take into account, in addition, matters relating to organizational and industry characteristics as well as products and finance. Lastly, it could be useful not only to focus on flexibility but also to consider attitudinal, behavioural and contextual matters relating to agility, capability, resistance, and resilience.\n",
"role": "user"
},
{
"content": "Purpose: - The research question which this paper aims to address is: To what extent does (labour) flexibility contribute to crisis resistance at establishment level? More specifically, the authors seek to analyse the determinants of variation in the extent to which establishments showed resistance to the global financial crisis (GFC), i.e. the extent to which they were affected by the crisis, focusing on an available secondary dataset related to organizational, industry-level, and (numerical) labour flexibility.\nDesign/methodology/approach: - Based on a unique cross-sectional dataset of 8,000 establishments in Germany, the authors use binary logistic regression to assess the link between organizational characteristics, industry-specific factors and workforce characteristics, and the fact that some establishments were affected by the GFC while others were not affected.\nFindings: - Establishment size, being located in western Germany and business problems before the crisis were positively associated with being affected by the crisis. The sector of economic activity also played a significant role. Contrary to predictions relating to the strategies employed by flexible firms, the extent to which they made use of temporary agency workers or of fixed-term employees showed no significant association with crisis resistance.\nPractical implications: - The dependent variable measures the management respondent's (subjective) perception of being affected by the crisis, although it does not specify the ways in which a company has been affected, for example by a drop in demand or by difficulty in extending credit. The set of independent variables permits several tentative conclusions regarding numerical and functional flexibility, but it does not take alternative forms of flexibility into account.\nOriginality/value: - Using a unique and representative dataset, the findings suggest a less important role for numerical flexibility in establishment performance and crisis resistance when compared to other variables, at least within the authors' research framework and its exceptional external economic circumstances.\n",
"role": "assistant"
}
] |
Turkish foreign direct investment and peace in Somalia: a new political stabilization policy
|
[
"FDI",
"Conflict",
"Infrastructure",
"Peace building",
"Post-conflict economy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Somalia is one of Organization of Islamic Cooperation (OIC) member states that have been suffering from prolonged conflicts. Africa is associated with armed conflicts and most of the conflicts are centered on resources. Similar to other African conflicts, Somalia's conflict is believed to be linked to conflict of resources (Raleigh, 2010; Bradbury, 1994; Elmi and Barise, 2006; Osman, 2007; Shortland et al., 2013; Ingiriis, 2012; Voors et al., 2017). Following the state collapse in Somalia, physical resources became limited, infrastructure remain ill developed, education opportunities are scarce, access to health services is difficult and to achieve a stable daily food is a dream for many households (Ahmed, 1999; Lindley, 2009; Francois and Sud, 2006; Akpinar, 2013; Ozkan and Orakci, 2015). In the context of Somalia, the conflict persisted for decades and peacemaking efforts failed due to the devastating economic conditions (Menkhaus, 1996, 2003, 2009; Akpinar, 2013; Elmi and Barise, 2006; Fernandez y Garcia, 2017; Malito, 2017). Different means have been employed to make peace in Somalia but neither of them provided lasting peace in the country (Donais and McCandless, 2017; Fernandez y Garcia, 2017; Kumuyi, 2017).\n2. Literature review: FDI is an investment made by foreign individuals or/and organizations into a country with the purpose of improving profit and enhancing shareholders' wealth by either boosting revenues or reducing costs (Jones, 2017). Any economy that does not participate in the process of globalization or formulate poor policies may relatively run the risk of becoming less competitive in the global economy (Jordaan, 2005; Kalai and Zghidi, 2017). Thus, FDI is an essential part of the huge private investment that has been driving global economic growth in the last two decades (Dabour, 2000; Addison et al., 2001). Moreover, various studies have analyzed the relationship between FDI and economic growth and the extent to which FDI affects economic development. Though there are methodological and conceptual differences, the general assumption to most of FDI studies is that it stimulates economic growth which in return improves social welfare (Gohou and Soumare, 2011; Ostby et al., 2009; Murshed and Tadjoeddin, 2009; Addison et al., 2001; Buhaug and Rod, 2006).\n3. Methodology: The aim of this study is to identify the extent at which FDI contributes to peacemaking. Research in this area is very limited; so, the availability and access to economic data that examine the relationship between FDI and peacemaking is difficult. A qualitative research method was adopted to collect relevant data from field experts who are believed to have extensive knowledge about the field of study as well as the environment.\n4. Results and discussion: During the interview, different issues have been raised and various solutions have been proposed for the security and challenges facing Turkish investors and their business interests in Somalia. On the other hand, respondents recommended that a national policy must be devised which provides guidelines to investors from OIC member states including Turkish investors. The results of the interview are presented in accordance with the research objectives: economic opportunities created by Turkish investors and contribution of Turkish investors to the process of peace building in Somalia.\n5. Conclusions: The aim of the study is to identify whether Turkish investments can create more economic opportunities and how these opportunities will contribute to peace building process in Somalia. The study found out that Turkish FDI can create more economic opportunities in Somalia if Somali Government produces a national policy that guides Turkish investors to appropriately allocate their funds to the different sectors of the economy while protecting domestic investors. Furthermore, Turkish FDI in Somalia is expected to create new jobs, bring new technologies to the country, teach new skills to the domestic workers, enrich private sector, develop economic infrastructure of the country and provide adequate CSR. Alternatively, the study revealed that Turkish FDI in Somalia possibly will lead to conflict reduction and peace building.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the extent at which Turkish foreign direct investment (FDI) contributes to the process of peacemaking in host country.\nDesign/methodology/approach: This study utilized a qualitative method and used a focus group discussion and an in-depth interview.\nFindings: This study found that Turkish FDI contributes well to the process of peace building in Somalia.\nPractical implications: A well-balanced FDI is expected to contribute to the process of peace building in host country because FDI is associated with job creation, technology transfer, private sector enhancement, infrastructure development and social responsibility. These factors increase household income of the host country and contribute to human capital development. By doing so, resource-centered conflicts are reduced and peace is established.\nOriginality/value: This study is considered as one of the unique studies that have been conducted in the context of post-conflict economy of Somalia. The study investigated the extent at which Turkish FDI contributes to the process of peace building in the host economy and found some interesting evidences, which makes it an appealing work that should be pursued further. This study is not only investigating a novel area of research but also contributes very much to the ongoing policy debate on the recovery and development of post-conflict economies.\n",
"role": "assistant"
}
] |
An integrated Bayesian-Markovian framework for ascertaining cost of executing quality improvement programs in manufacturing industry
|
[
"Manufacturing industry",
"Cost of quality",
"Quality improvement program"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Nomenclature\n1. Introduction: The dynamics of manufacturing industry particularly product-centric manufacturing has evolved from the era of mid-to-late twentieth century (Spring et al., 2017). Between the period of 1960s-1990s, original equipment manufacturer (OEM) used to get themselves involved at all levels of value chain in which OEM's role ranged from one end of the spectrum as a component manufacturer to another end as an assembler. However, with increased competition among OEMs, they were forced to focus on their core activities (e.g. assembling and selling automobiles in the case of automobile industry), thereby outsourcing much of the important but auxiliary activities, for example, component manufacturing to suppliers (Cirvegna et al., 2013). Easier technology availability also played a pivotal role in ensuring that suppliers in manufacturing industry developed their technical expertise so as to meet the demands of OEMs both in terms of quality and volumes. Infact, in today's era, be it automotive industry, construction equipment industry or any other product-centric manufacturing, there are often very few critical items that OEMs manufacture in-house; rather, they rely on suppliers possessing specific design and manufacturing expertise. A case in point is the example of Bosch that supplies fuel injection pumps (FIPs) for engines to major automotive OEMs across the world (Bosch Automotive Catalogue).\n2. Literature review: The problem that we are addressing through our research pertains to modeling for supplier quality management. The Bayesian theory and Stochastic process, specifically Markov chain, are other two important dimensions that provide inputs to the framework that we intend to develop. Against this backdrop, we discuss some recent and relevant research literature. Although studies related to ascertaining the cost of quality including the cost of executing process improvements are relatively sparse, the related literature presented in this research is further segregated in three broad categories, namely, case-study-based research, empirical investigation based, analytical and theoretical studies.\n3. Problem setting: The need for a critical assessment of the quality improvement programs in terms of accompanying efficacy and effectiveness has been aptly discussed in the study of Antony and Karaminas (2016). This study also emphasized that quality improvement programs should be viewed from a resource-based perspective as well such that appropriate trade-offs can be established between cost and quality. Therefore, in this backdrop, within our problem environment, an OEM needs to ascertain the expected cost that it would have to bear to support the process quality improvement initiatives at a particular supplier's facility. In the backdrop of the OEM supporting the supplier for implementation of new technology at the supplier's facility, we assume that the initial state of process quality level at supplier's facility is QLUL. If the underlying problem is modeled along the stochastic process (more specifically Markov chain), QLUL (initial state) can transition into different states, i.e. QLMAL, QLMOL, QLMIL and QNL with associated transition probabilities. The rationale for modeling the problem along the Markov chain is grounded essentially in the fact that transition probabilities involving how the process improvement program will evolve in the future would depend only on the current state of the Markovian process. As the process improvement program at the supplier's facility would be driven by the OEM's prior experiences in executing similar PQIP with other suppliers, we model the problem in such a way that there would be m number of subject matter experts (on the basis of their prior technical experiences executing similar programs) providing inputs in ascertaining the initial transition probabilities. As these transition probabilities cannot be ascertained directly and also to mitigate subjectivities, we employ the pair-wise comparison and subsequent Bayes theory-based probability determination approach to ascertain the transition probabilities (Chin et al., 2009; Goswami, 2018).\n4. Illustration using a real-life case: We in this research consider a case of automotive industry in India in that we consider an example of India's premier automobile company whose managerial leaders have had significant number of man-years driving the SQIP at different types of suppliers be it component focused, sub-assembly focused or engineering system focused. Most of the supplier's facilities where the improvement programs were executed have their own in-house design and manufacturing capabilities as opposed to purely possessing only manufacturing capabilities.\n5. Results and discussions: Referring to the steady-state probability values, it can be ascertained that based on qualitative inputs from experts and using the quantitative analysis grounded in Bayesian theory and Markov chain, the most likely quality level that the supplier would be transitioning from process quality standpoint is near perfect level of quality (QNL). However, inspecting the associated steady-state value corresponding to this quality level also establishes that the probability of transitioning to this quality (from early quality level) is not very dominant vis-a-vis transitioning to other quality levels. Furthermore, even after executing the PQIP at the supplier's facility, there is around 20 percent chance that supplier's process quality level might not improve. We sought to rationalize this aberration in that most likely steady-state quality level is QNL, while, at the same time, significant chances are that supplier process quality level would be rather the lowest level. One of the possible justifications for this anomaly is that the particular supplier had rather heterogenous mix of technical workforce in that a certain fraction of workforce was technically very proficient, while on the other hand other chunk of workforce was technically relatively deficient as well as unmotivated. This heterogeneity, however, got reflected in the supplier's capability assessment by experts in that few expert's judgment would have been dominated by the share of technical workforce having high proficiency level, and on the other hand, few expert's judgment would have been dominated by the share of technical workforce possessing the lower proficiency level and/or motivation. However, this varied judgment would be mitigated to an extent because mean transition probabilities are being considered in our devised model.\n6. Managerial implications: For an OEM, continuous improvement for themselves and associated suppliers' technical and process capabilities are an important tactical decision for which OEMs often need to earmark significant budgetary provisions out of the annual anticipated operational costs. Anecdotally speaking, except for a few large global OEMs, most OEMs often take the decision pertaining to the level of expected budget for supplier process quality improvement in a rather unscientific and unstructured manner. Typically, quality improvement managers either rely on historical data or on their own empirical judgment for ascertaining the expected budget for executing the quality improvement programs. A key limitation of such unscientific and unstructured approach is that managers often end up overestimating or underestimating the expected cost associated with carrying out such programs at suppliers' facilities. A key feature of our evolved Bayesian-Markovian anchored framework in this research is that by adopting our framework, OEMs can minimize deviations either considering overestimation or underestimation. From a practitioner's perspective, our study augments the extant literature in a number of ways. By incorporating a Bayesian-based pair-wise comparison grounded approach, we are able to ascertain relatively accurately various transition probabilities as far as different quality levels are concerned. This is a significant departure from often unstructured ways in which managers ascertain the quality levels associated with suppliers. Furthermore, this approach also negates the presumption that after carrying out certain quality improvement program, the process quality level would automatically improve to certain other higher quality level in a sure shot manner. Furthermore, our approach is rather unconvoluted and pragmatic in that once all possible quality levels are identified, pair-wise comparisons from all experts are ascertained, cost coefficients are estimated, solving a number of simultaneous equations fetches us the steady-state probabilities associated with the respective quality levels. These steady-state probabilities in conjunction with the cost coefficients associated with respective quality levels ultimately yield the expected cost of carrying out the quality improvement programs.\n7. Conclusion and future research direction: In this research, we have undertaken an important problem of ascertaining the expected cost of carrying out suppliers' process quality improvement programs that are driven by OEM. The mathematical framework devised in this research is based on the Bayesian theory and Markov chain.\n",
"role": "user"
},
{
"content": "Purpose: Typically, the budgetary requirements for executing a supplier's process quality improvement program are often done in unstructured ways in that quality improvement managers purely use their previous experiences and pertinent historical information. In this backdrop, the purpose of this paper is to ascertain the expected cost of carrying out suppliers' process quality improvement programs that are driven by original equipment manufacturers (OEMs).\nDesign/methodology/approach: Using inputs from experts who had prior experience executing suppliers' quality improvement programs and employing the Bayesian theory, transition probabilities to various quality levels from an initial quality level are ascertained. Thereafter, the Markov chain concept enables the authors to determine steady-state probabilities. These steady-state probabilities in conjunction with quality level cost coefficients yield the expected cost of quality improvement programs.\nFindings: The novel method devised in this research is a key contribution of the work. Furthermore, various implications related to experts' inputs, dynamics related to Markov chain, etc., are discussed. The method is illustrated using a real life of automotive industry in India.\nOriginality/value: The research contributes to the extant literature in that a new method of determining the expected cost of quality improvement is proposed. Furthermore, the method would be of value to OEMs and suppliers wherein the quality levels at a given time are the function of quality levels in preceding period(s).\n",
"role": "assistant"
}
] |
From the stakeholder viewpoint: designing measurable objectives
|
[
"Organization objectives",
"Strategy",
"Strategic planning",
"Stakeholders",
"Behavioural outcomes",
"Measures",
"New paradigm"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. An illogical mess: Organization objective setting has long been a core ingredient in the diet of CEOs and senior management teams. It is a necessary condition for developing clear and concise organization strategy. As the saying goes: if you don't know where you're going, you're never going to get there.Yet in spite of objective-setting's centrality to strategy design and, indeed, to performance measurement, the method for developing objectives hasn't progressed beyond a seat-of-the-pants response to the question \"What are our objectives?\".The neglect of the this core function can be seen in strategic management textbooks, for example Hill et al. (2007), Hubbard et al. (2008), Johnson et al. (2011) and Fitzroy et al. (2012). As a result, managers of organizations and directors on boards are left to their own devices as to how to go about designing measurable objectives.The purpose of this article is to describe a method that has been field tested for more than a decade in a vast range of organization and industry settings and with senior executive teams from all sectors - private, public, not-for-profit.\n2. Typical objective-setting session: Let's look at what usually takes place when it comes to setting organization objectives as part of a strategic planning exercise.A group composed of the CEO and several managers will already have deliberated at length on the organization's mission, vision and values. Then a member of the group exclaims: \"Well, what are our objectives?\". \"Our\" doesn't refer to the group as managers, but to the organization - be it in the business, government or not-for-profit sector. It's apparent that the managers have rightly adopted the persona of the organization. The organization has become a \"person\", which is completely consistent with company law (Baxt, 2009).To address the question, the CEO, a senior manager or a professional facilitator, will walk over to a flipchart or whiteboard and write the word \"Objectives\" on it. Everyone will pile in with suggestions and, in no time at all, a list something like this will probably appear:* to grow the business by 10 percent per annum;* to operate with integrity;* to increase productivity;* to get suppliers to deliver on time;* to attract better staff;* to increase profitable revenue;* to be a good corporate citizen;* to be innovative;* to change the foyer decor;* to be number one in the industry;* to decrease employee turnover;* to increase funding; and* to provide products diametrically in the opposite direction to the competition.The group looks at its handiwork and thinks \"good!\". But it also thinks \"not so good\". The good bit is that the managers have a list. They have many items. Everyone has participated. The not-so-good part is that the list is a real jumble - frankly, an illogical mess. Some items are quantified, others are vague and general. Certain ones are a repeat of the organization's mission, vision and values. As a result the managers are not really satisfied with the outcome. But it's the way they've always done it. Hardly sophisticated, the managers will readily acknowledge, but as they know no other way, they continue the practice.It's quite alarming that in spite of all the advances in strategy design through the works of Michael Porter (Porter, 1980, 1985) and others, this very central element in the strategy development process - objective-setting - has remained so sterile for so long. Before I outline a fresh objective-setting technique, I review where we are at present.\n3. Objectives in strategy: It's widely recognized that it was Igor Ansoff's 1965 book, Corporate Strategy, that formalized business strategy (Ansoff, 1965). As Henry Mintzberg has described it:The publication of the book, Corporate Strategy, by H. Igor Ansoff was a major event in the 1965 world of management. As early as it came in this literature, the book represented a kind of crescendo in the development of strategic planning theory, offering a degree of elaboration seldom attempted since (at least in published form) (Mintzberg, 1994, p. 43).Ansoff provides an extensive discussion on objective-setting and its pivotal place in strategy design, an emphasis that has been brushed over by subsequent authors. (Ansoff wrote other books on strategy - e.g. Ansoff, 1979, 1984 - but neither of these dealt with objectives to the extent of his 1965 book.)Before reviewing Ansoff's contribution, I'd first like to consider what another of modern management's forefathers, Peter Drucker, had to say about setting objectives.Drucker's seminal contribution came a decade earlier than Ansoff's with the publication of The Practice of Management (Drucker, 1955). This bestseller was distributed in training programs and was widely read by managers. It also had a major role in disseminating \"management by objectives\", which refers to how individual managers are tied to organization objectives through a process of cascading objectives from the top of an organization to the bottom. The purpose of the method was to assist every manager to develop his or her objectives so that each would know his or her contribution to overall organization objectives.It's the latter, of course, that I'm concerned with here, and Drucker devotes a whole chapter to the subject: \"The objectives of a business\". But here's the problem: he doesn't provide any \"how to\" at this level. While he nominates certain \"areas in which objectives of performance and results have to be set\" he leaves it to practitioners to move ahead from there (see also Drucker, 1976, also in Drucker, 1981).Drucker did, however, suggest a pattern in classifying objectives and developing attendant measures (see also Drucker, 1964). Today this is commonly conducted under the rubric \"key result areas\" or some similar heading, the most recent manifestation of which is the Balanced Scorecard (Kaplan and Norton, 1996, 2001; Schneiderman, 1999), with its four \"perspectives\" - or, as some managers have described them, the \"four buckets\" (Ittner and Larcker, 2003). But all of these remain ad hoc methods for developing the objectives themselves. In the Balanced Scorecard's case, for example, executives are interviewed for their off-the-cuff suggestions (Kaplan and Norton, 1996, pp. 302-5).So for some fundamental thinking on the subject we really do have to return to Ansoff.In Corporate Strategy, Ansoff devotes two chapters to objective-setting - one called \"Objectives\", the other \"A practical system of objectives\". He starts the latter thus:We shall approach practical objectives through a series of approximations. Keeping the maximization of the rate of return as the central theoretical objective, we shall develop a number of subsidiary objectives (which the economists call proxy variables) which contribute in different ways to improvement in the return and which are also measurable in business practice. A firm which meets high performance in most of its subsidiary objectives will substantially enhance its long-term rate of return. (The defect in our approach is that we cannot prove that the result will be a \"maximum\" possible overall return.) As will be seen, this road has its own obstacles: the difficulties of long term maximization are replaced by the problem of reconciling claims of conflicting objectives (p. 47).Clearly Ansoff is addressing the needs of firms here and not necessarily the requirements of government and not-for-profit organizations. But \"the problem of reconciling claims of conflicting objectives\" is common across all sectors.Within the chapter, Ansoff provides a section called \"The process of setting objectives\". Here you might think is the \"how to\". But no, the author remains at a system and structure level for the \"firm\", and hence it is far removed from managerial action.To understand better what Ansoff is driving at, readers need to visit his early chapter, \"Objectives\", and here I find myself parting company with him. (Readers may wonder why the concentration on Ansoff. It is because of his pivotal role in the objective-setting literature and the lack of methodological development since his contribution; for an academic review of the literature, see Shinkle, 2012.) In the \"Objectives\" chapter Ansoff rejected early \"stakeholder theory\" (Abrams, 1954; Stewart et al., 1963) as a foundation for setting organization objectives because, as he wrote, \"the theory maintains that the objectives of the firm [emphasis added] should be derived by balancing the conflicting claims of the various 'stakeholders' in the firm: managers, workers, stockholders, suppliers, vendors\" (Ansoff, 1965, p. 39). In this interpretation, organization objectives become what stakeholders want from the firm. Viewed this way a stakeholder framework for developing objectives looks like \"all give and no take\". Little wonder, then, that Ansoff was not enamored of a stakeholder approach. Yet in my formulation (Kenny, 2001, 2005) a stakeholder framework in strategic planning and objective setting is both give and take (Freeman, 1984). In fact, and in practice, organization objectives are what the organization wants from its key stakeholders.\n4. Stakeholders and objectives: Organization objectives to do with profitable revenue, for example, are based on obtaining this from customers, one of the stakeholder groups. Organization objectives concerned with improving productivity and increasing innovation are what a firm wants from another stakeholder: employees. And so it goes. This rethink about the very nature of organization objectives makes a logical connection between what stakeholders want from an organization, which I call \"strategic factors\", and which are the basis of an organization's strategies, and what the organization wants from its stakeholders, labeled \"organization objectives\". Both sides together provide a true strategy framework called the strategic factor system (Kenny, 2001, 2005).I now take a close look at the lists of objectives developed by senior management teams following the conventional means, such as that provided at the beginning of this article. What's hidden beneath them is indeed a stakeholder structure. In other words, unless the items were merely a restatement of the organization's vision, mission and values in a different form, or an action, each objective on these lists is related to an organization stakeholder:* to grow the business by 10 percent per annum (customers);* to operate with integrity (a value);* to increase productivity (employees);* to get suppliers to deliver on time (suppliers);* to attract better staff (employees);* to increase profitable revenue (customers);* to be a good corporate citizen (a value);* to be innovative (a value);* to change the foyer decor (an action);* to be number one in the industry (part of vision statement);* to decrease employee turnover (employees);* to increase funding (government); and* to provide products diametrically in the opposite direction to the competition (part of mission statement).So if organization objectives are underpinned by a stakeholder structure, why not design a more systematic approach to objective setting by making the stakeholder structure explicit, right at the outset? In other words, identify an organization's key stakeholders first, such as customers and employees, and then methodically develop objectives, stakeholder by stakeholder.What then becomes clear is that designing an organization's objectives means laying out what it needs to achieve through its key stakeholders (Kenny, 2001, 2005). This is also what an organization wants as inputs from its key stakeholders, for example funds or productivity. Getting those inputs is strategy. And this is the link between the strategy of an organization and its objectives.I came to unravel and reconstruct the objective-setting problem in this way several years ago. It is of note that it's completely consistent with the modern systems view of everything - organizations, climate change, ecology, etc. - in which one thing depends on another. At the time, however, I felt that there was a much needed prior step to get to a truly fundamental objective-setting method.It goes this way. If an objective is an expression of what an organization wants from a key stakeholder, it must involve behavior by a stakeholder and, as far as the organization is concerned, a behavioral outcome. For example, if an organization wants revenue from customers, the behavioral outcome is having customers buy from the organization. If we recognize these behaviors first, I thought, we're in a much better position to develop measurable objectives once these behaviors are translated into organization objectives.\n5. A streamlined approach: 5.1 Identify key stakeholders\n6. Time for a new paradigm: The method of objective setting outlined in this article is sound theoretically and leads to the design of clear and measurable objectives. Once developed, these objectives will shape an organization's strategies effectively, leading to sustainable success. It's important to remember that targets on organization objectives shape organization strategies; they also provide a means of assessing the effectiveness of those strategies. Finally, without clear and quantified objectives, any strategy will do!\n",
"role": "user"
},
{
"content": "Purpose: - This article seeks to review how objectives are conventionally designed for organizations before detailing a new and proven approach based on the author's many years of research and consulting experience. The practical situation on which the article aims to focus is strategic planning, a process in which CEOs, middle managers and directors on boards find themselves frequently involved.\nDesign/methodology/approach: - The outlined objective-setting methodology is the result of more than two decades of action research. The author has worked with organizations across all sectors - private, public and not-for-profit - to develop meaningful and measurable objectives during the course of designing their strategic plans.\nFindings: - Most organizations flounder when it comes to objective setting. They come up with lists of objectives that are a mixture of quantified and non-quantified items or restatements of mission, vision and values; they are often vague and general and are assembled as one undifferentiated block, i.e. not segmented according to an organization's key stakeholders. No wonder managers find the conventional process of objective-setting both frustrating and ineffective. In contrast, the author has found that the approach outlined in this article engages participants and produces clear and measurable outcomes.\nOriginality/value: - Organizations in all sectors - business, government and not-for-profit - can benefit greatly by re-thinking their current practices and changing the way they go about designing their objectives. The approach outlined herein has solved many conceptual and practical problems managers face and led to the design of organization objectives that really shape precise strategies.\n",
"role": "assistant"
}
] |
Relationship among school size, school culture and students' achievement at secondary level in Pakistan
|
[
"Schools",
"Students",
"Pakistan"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nAbout Pakistan: Pakistan is located in South East Asia. It comprises four provinces: Punjab, Sindh, North West Frontier Province (NWFP), Balochistan and other regional territories called Federally Administrative Tribal Areas (FATA), Federally Administrative Northern Areas (FANA), and Azad Jammu and Kashmir (AJK). Besides this, there is a federal territory, Islamabad, the capital of the country. The estimated total population of the country is 140 millions (Retallick and Mithani, 2003). Punjab is the most populous province, comprising around 55 per cent of the entire population of the country. Balochistan is the least populous province, but the biggest one in geographical area. About 96 per cent of the population is Muslims; 2 per cent Christians and 2 per cent other minorities including Hindus, Sikhs, Parsis, Ahmadies etc. In each province, apart from the national language, many local languages are also spoken; the most popular are: Pashto in NWFP, FATA and FANA; Balochi in Balochistan; Sindhi in Sindh, Punjabi in Punjab and AJK. The medium of instruction is Urdu at elementary and secondary level, but at higher education level both English and Urdu are used as a medium of instruction and assessment throughout the country. In the province of Sindh, however, there is an option, either to take Sindhi or Urdu at primary level. The most popular categories of public sector schools are primary (grades 1-5), elementary (grades 1-8), secondary (grades 1-10 or 6-10) and higher secondary (mostly grades 6-12). All constitute about 157,391 with students' enrollment of 17,279,100 (Government of Pakistan, 2003).The literacy rate in Pakistan is less than 50 per cent. The participation rate at all levels is low and drop out rate is high, especially at primary level, and over 5.5 million children (5-9 year age) do not attend schools (Government of Pakistan, 1998). The drop out at primary level is 45 per cent, while at middle/elementary and secondary level is about 30 per cent. Low literacy rate is mainly due to poverty and non-statutory primary education. But appreciating thing is that education is free up to primary level all over the country; in Punjab and Sindh it is free at middle and secondary levels. It is hoped that the other provinces and regions will also undertake this initiative and gradually extend the free education from primary to middle and secondary levels.\nBackground: The learning process in the individuals starts with their birth. It begins at home by imitating the elders and observing their activities. Though the child learns many basic skills at homes, but he/she needs to learn more to face the complexities of life and perhaps educational institutions are the best sites for his/her further development. The schools are the first institutions of socialization of individuals which help them in developing certain competencies and find ways for their adjustment in the society. Early child enters the school with its peculiar intellectual, social and emotional characteristics wherein he/she finds teachers of distinct intellectual capabilities, professional competencies and experiences, which interact with each other during the educational process.School environment is the result of interaction of teachers, students, head teachers, and learning activities with the prime goal of preparing young generation for having successful future. But a number of factors affect the learning process or success interactions. Saeed (2003) states as \"school, teacher, pupil and parent/community-related - all affect the teaching learning discourse\". It is imperative that these factors should properly function for the quality of education. Although educational researchers and policy makers consider a number of variables in operationalizing quality of education (Bossert, 1988), thus the ultimate criterion for assessing the effectiveness of any school is the extent to which it improves the academic achievement of the students.The term \"culture\" is a system of shared meanings, symbols, experiences and practices, shared values, beliefs and sense-making (Morgan, 1986). Deal and Peterson (1990) includes in culture \"deep patterns of values, beliefs, and traditions that have been formed over the course of the school's history\". In view of McBrien and Brandt (1997) school culture and climate refers to the sum of values, culture, safety practices, and organizational structures within a school that cause it to function and react in particular ways. Heckman (1993) perceives school culture as \"the commonly held beliefs of teachers, students and principals\". Phillips (1993) states that \"school culture is the beliefs, attitudes, and behaviours which characterize a school\". Barth (2002) perceives it rather a complex phenomenon, as he states \"a school culture is a complex pattern of norms, attitudes, beliefs, behaviours, values, ceremonies, traditions, and myths that are deeply ingrained in the very core of the organization\".There are many factors which affect the students' learning in schools. These include: teachers' qualifications and experiences, teachers' guidance to students, availability of teaching learning resources, physical facilities, students' own cognitive and other abilities, and their socio-economic backgrounds. Saeed et al. (2005) found that parental qualifications, family size, school liking and teachers' guidance have impact on the achievement of primary grader students. Warwick and Riemers (1992) in their study at Harvard found that teachers' qualifications and subject knowledge had a strong correlation with students' achievement. Khan and Shah (2002) quote that \"family environment and parents' involvement in educational activities of their children had a positive impact on their achievement\". The studies conducted at the Academy of Educational Planning and Management (AEPAM) (Khan and Shah, 2002; Habib et al., 2004) revealed that teachers' academic and professional qualifications had a positive impact on the students' achievement. AEPAM study (Habib et al., 2004 also found a positive impact of parental education on students' performance.There are factors related to school culture and climate that influence the students' achievement. Deal and Peterson (1999) state \"school's culture\" is one of the important factors that influences academic achievement of students. Bulach et al. (1995) found a significant difference in students' achievement between schools with a good school climate and those with a poor school climate. Saeed (1997) in the context of international comparisons investigated that school prestige, principal's leadership style, monetary rewards, better working conditions for teacher, and parental encouragement were major predictors for secondary school teachers' job satisfaction which in turn had a positive impact on students' learning. Erpelding (1999) and Hirase (2000) found that schools with a positive climate had higher academic achievement of students.In schools, there are however, factors that affect the school culture and climate; one of the most prominent one is the number of teachers and students in a school. Tucker (1997) found small schools have a better learning climate than large schools because in smaller schools there is a greater likelihood that faculty and students know each other. And this could affect the levels of openness, trust, cooperation, atmosphere and other culture and climate variables that can directly or indirectly affect the students' achievement. Pittman and Haughwout (1987) found that smaller school size is to be strongly correlated with a composite measure of school climate including student participation, interaction with faculty, sense of cohesion and infrequent discipline problems.At secondary school level, most evidence about students' performance on achievement tests also points out to the benefits of smaller schools for most students (Eberts et al., 1984; Fowler, 1995; Mok and Flynn, 1996; Texas Education Agency, 1999; Howley, 2002). Fowler and Walberg (1991) found that school size was inversely related to test scores in mathematics and writing. Smaller schools were associated with higher passing rates on writing portions of the state's High School Proficiency Test and reading portions of the state's Minimum Basic Skills Test. Walberg and Walberg (1994) while using the data of National Assessment of Educational Progress (NAEP) (1990) examined relationships among size, expenditures and achievement. The study revealed that larger schools tended to score lower on the NAEP mathematics assessment.There are, however studies which found that there is no impact of the factor that either the school is larger, medium or smaller; even in many cases smaller schools found low students' achievement (Cotton, 1996; Greenwald et al., 1996; McMillen, 2007). Lindsay (1984) found no meaningful relationship between school size and academic ability. In a study carried out by Jewell (1989), it was found that there is a significant relationships between high school size and either ACT or Scholastic Achievement Test (SAT) scores. Gardner (2001) investigated that there is no relationship between school size and students' achievement. Lee and Smith (1997) found that students from medium-sized high schools demonstrated higher achievement than students in either smaller or larger schools. The study of Mirza and Hameed (1994) revealed that primary sections of middle and high schools were most effective in promoting cognitive achievement and molding students' behaviour.In view of the above literature review, it is evident that controversial results were revealed in regard to relationships among school size, school culture and climate, and students' achievement. There is no large scale study available in the context of secondary schools in Pakistan, therefore the researchers intended to investigate the relationships among school size, school culture and students' achievement.To achieve the objective of the study \"to determine the relationship among school size, school culture and students' academic achievement\", following three null hypotheses were developed against the three major variables of the study: school size, school culture and students' achievement.Ho1. There is no significant relationship between school size and school culture.Ho2. There is no significant relationship between school size and students' academic achievement.Ho3. There is no significant relationship between school culture and students' achievement.\nMethod and procedure: Population and sample selection\nResults: Demographic features\nDiscussion of results and conclusions: The intent of this study was to determine the relationship among school size, school culture and students' academic achievement of secondary schools in Pakistan. Findings from the study indicate that school size is one of the factors that contribute to the positive school culture and higher academic achievement of the students. The findings of this empirical study clearly indicate that the tangible and intangible elements of school culture are given attention in smaller and well-performing schools in comparison to larger and poorly performing schools. This leads to the conclusion that school size is directly related to both school culture and the academic achievement of secondary grader students. Previous research carried out in secondary schools supports this key finding of the study in that smaller schools are more associated with better school culture and higher academic achievement (Eberts et al., 1984; Fowler and Walberg, 1991; Mok and Flynn, 1996; Howley, 2002).There may be many reasons for this claim that small schools having a better school culture academically performed better than medium and large schools. One of the most probable reasons is that in smaller school there is a greater likelihood that faculty and students know each other, opportunities to talk to each other, to know each other's names and interests are more likely to occur. This could affect the levels of openness, trust, cooperation, atmosphere and others culture and climate variables that directly or indirectly affect the students' academic achievement.Another factor associated with school improvement and effectiveness is the parental involvement. Berlin and Cienkus (1989) state \"education is more productive when parents, students, and teachers are bonded in their educational goals\". Other research also supports this conclusion. Thomas (1987) found \"strong parent-teacher relations as being positively correlated with students' achievement\". Research strongly supports to more parental involvement tends to be greater in smaller schools than in larger schools (Walberg, 1992; Meier, 1996). Walberg (1992) stated that small schools tend to have greater physical and psychological proximity to students' homes than do larger schools, and that such proximity increases parental involvement because parents feel more comfortable approaching a smaller school. Meier (1996) took a similar approach arguing that a culture of respect between parents, students, and teachers was necessary for parental involvement. This culture comes from parents and teachers knowing each other personally, and is, therefore, more likely to develop in smaller schools. Green and Barnes (1993) also found similar results that administrators in small schools relied more heavily on verbal communications and held more conferences with parents than administrators in large schools. The findings of this study also support this claim of researchers as parents were frequently involved in school activities and they could check their children's progress easily in smaller schools as compared to larger schools.Personal social relations (e.g. teacher-teacher, teacher-student, student-student) and opportunities for more varied approaches to instruction and assessment may play a mediating role in creation of positive school culture and high academic achievement (Lee et al., 2000; Wasley et al., 2002; Darling-Hammond et al., 2002). So in smaller schools, there may be a greater possibility to develop personal social relations and having better opportunities of professional growth of teachers than medium and larger schools. In smaller schools, teachers generally have better opportunities to know each other, share ideas and materials with each others, and the head teacher can easily guide and provide proper feedback to teachers that can contribute in the betterment of academic achievement of students.Disciplinary problems in large school may also be another factor of average or poor performance among students. Stockard and Mayberry (1992) found that the increase in behaviour problems in larger schools was sufficient to abolish any benefits that might occur from having larger schools. In small schools, administrators can control disciplinary problems easily as compared to large schools. The increase in disciplinary problems in large schools tends to create negative changes in administrators' behaviour. Gregory (1992) found that the control of students' behaviour becomes a primary concern of administrators in large schools. As a result, school policy tends to become restrictive and disciplinary actions highly punitive (Meier, 1996).To sum up the discussion, if we see to what extent the results of this study could be applied to other geographical settings; for that might be needed more research as teachers' qualification, socio-cultural norms, and economic situation differ in other geographical regions, especially in the context of such studies where to investigate school culture. But as throughout Pakistan, national ideology and school curricula are the same and to some extent economic standards are alike, therefore it can be assumed that similar findings may be emerged from other provinces and regions of the country. But it might not be appropriate to exactly generalize the results of the study at large scale owing to small sample of 79 schools and 540 teachers. But as these results are in line with the previous studies, so these may help the education departments and other policymakers to at least consider that what should be the optimum size of a secondary school? So that all students may find better school culture and show better academic performance. The researchers recommend further research on a national level sample so as to make proper decisions before policy formulation and execution in the field.\n",
"role": "user"
},
{
"content": "Purpose: - This paper seeks to determine the relationship among school size, school culture and students' achievement at secondary level in Pakistan.\nDesign/methodology/approach: - The study was descriptive (survey type). It was conducted on a sample of 90 secondary school head teachers and 540 primary, elementary and high school teachers working in the government boys secondary schools of Punjab province. Data was collected through three sources: first, statistics on education from Education Management Information System (EMIS) Punjab; second, annual results of grade 9 and 10 students of Boards of Intermediate and Secondary Education (BISEs) Punjab; and third, a questionnaire which contained 39 items at five-point rating scale and ten items in yes/no form. The validity and reliability of the questionnaire were ensured through experts' opinions and pilot testing in early 2006; the overall reliability was established at 0.967.\nFindings: - There was a significant correlation between school size and school culture, and school size and students' achievement. Small schools revealed positive school culture and performed better than medium and large schools.\nOriginality/value: - The policy makers, administrators and managers, and teachers at secondary level may improve school culture by bringing schools to a reasonable size, which may improve the students' achievement in Pakistan. The findings may be useful for other countries of almost similar socio-economic status to improve the quality of teaching-learning at secondary level.\n",
"role": "assistant"
}
] |
Determining factors of success in internal communication management in Spanish companies: The influence of social media
|
[
"Employee relations",
"Internal communications",
"Social media",
"Companies’ success"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The advent of the Web 2.0 era has changed the style of interaction inside organizations by creating an open corporate space where collective intelligence accrues added value on top of the mere sum of individual expertise possessed by the organization's employees (Men, 2015). In the past decade, social media channels have revolutionized internal communication by their ability to create and strengthen ties among employees, collect and exchange information, foster teamwork, generate trust and loyalty, build lasting relationships and facilitate the role of workers as ambassadors of their own company (Cao et al., 2012; Mazzei, 2014a, b; Morgan et al., 2004; Skeels and Grudin, 2009; Wattal et al., 2010; Zoonen et al., 2014). Companies that invest in improving the effectiveness of their communication practices deliver better market share, more shareholder value, increased engagement and lower employee turnover (Yates, 2006).\nThe strategic role of the communications director: Internal communication is crucial in order to ensure that employees carry out their activities and that the company increases its competitive advantage. Its integration within the managerial structure is one of the keys in achieving a higher organizational effectiveness (Chen, 2008). In this sense, the principles underlying the widely accepted theory of excellence in public relations (Broom and Dozier, 1986; Dozier and Broom, 1995; Grunig, 1992) claim that those responsible for managing communications in business organizations: must be involved in the company's strategic management; must be part of the dominant coalition; must have a managerial point of view, expertise and professional experience; and must resort to bidirectional communication. The authors of the theory of excellence likewise claim that the organization's structure and culture have an influence on the effectiveness of work done by public relations professionals and advocate a \"communication-driven culture\" where participation, symmetry and equal opportunities are encouraged.\nImportance of social media channels in communicating with employees: Today, organizations have an array of communication channels accessible to disseminate messages to employees, including company website, intranet, blogs e-newsletters, meetings and print materials. Each media conveys information and engage receivers in different ways. According to Daft and Lengel (1984), media capability is differentiated by the availability and speed of feedback of the channel, the use of multiple cues and natural language to facilitate understanding and the personal focus of the message.\nCommunication activities conducted through the internal social media: The daily functioning of companies nowadays requires access to knowledge and experience that are often diverse, disperse and instantly needed (Riemer and Scifleet, 2012). The complexity of their activities and the globalization of firms generate a growing number of situations that demand the collaboration of employees from several departments, working in different geographic locations and sometimes having different working hours. Aware of this need, social media platforms have in the last few years veered toward business practice, while there has been a steady rise in the demand for business-oriented social software in order to enhance organizational efficacy (Bughin and Manyika, 2007; McAfee, 2009).\nMethodology: In order to test this model, an online questionnaire was designed to be responded by communications professionals working for Spanish firms. As a prior step, we built a database including a total of 4,300 professionals from Spanish business organizations of several sizes in several sectors and operational scope. Selecting a representative sample of communications professionals involves a number of limitations because of the unavailability of accurate data about the complete population. However, the number of respondents does provide a starting point to analyze the practices of internal digital communications in Spanish companies.\nFindings: The results show that 41.40 percent of the Spanish communication practitioners evaluate internal communication as highly successful in their companies, 32.48 percent as successful, and the remaining 26.11 percent as unsuccessful. The model contrasted to explain which factors influence the level of success of internal communication is that of the ordered probit regression, using the usual likelihood ratio (kh2) statistic. The regression model is significant since the kh2 statistic is 68.26520 and the significance level is 0.0000. Additionally, the obtained goodness-of-fit observed by using the Nagelkerke's coefficient (0.56) is adequate and the results can therefore be considered valid (see Table IV).\nConclusion and discussion: The purpose of this study has been to analyze to what extent the role of the communication professional, the channels used to communicate with employees and the activities conducted through social media platforms influence the quality of internal communication within the organization. Ongoing changes at the workplace with respect to the evolving expectations of the employees make internal communication more important (Argenti,1998). The findings of the current study provide important implications for public relations scholars and professionals.\nPractical implications: Pragmatically, the findings of the current study provide important implications for public relations professionals on what (i.e. the content) and how to (i.e. the channels) communicate within an organization. In general, internal communicators should move from historical roles as information producers and distributors to advisory roles in strategic decision making. Additionally, the implementation and use of ISM should be carefully revaluated. Communication practitioners should examine and address the difficulties involved in choosing the correct channels, devoting the necessary time to their adequate management, analyzing employees' feedback and improving the engagement. Creating the opportunity for social interaction using the latest technology, organizations will foster the sense of community and will improve the financial performance. Organizational leaders need to analyze costs and benefits and must provide the company with mechanisms that encourage the adoption of new technologies that ignite the passion for knowledge and reduce barriers to creativity.\nLimitations: The current study has several limitations that should be noted and addressed in future research. The main limitation lies in the fact that the dependent variable - success in communication - is entirely based on communication professionals' perceptions. Replication studies can be conducted to cross-validate the results obtained from this study using business outcome metrics to measure the communication effectiveness. A further limitation is related to the sampling procedure. Obtaining a representative sample of communications professionals in Spanish companies poses a number of limitations due to the impossibility of having accurate data on the total population. It would be highly interesting to replicate this research with a bigger sample and to include the employee's perception. In addition, future research should analyze the particularities and influence of each social media platform: social networks, blogs, microblogs, wikis, etc. It is therefore necessary to continue analyzing the current use of social media tools at the workplace in order to contribute toward the success of communication with employees. Finally, future researchers could focus on developing a comprehensive theoretical framework in predicting factors that might contribute to the success of internal communication.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore the factors that influence the success of internal communication in Spanish companies, specifically the importance of the channels used, activities conducted through ISM and the role played by the communication professional.\nDesign/methodology/approach: The model employed argues that to obtain success in internal communication, three factors are considered necessary: first, communication professionals must participate in strategic decisions; second, they have to use together social media channel and classical channel to communicate with employees; and finally they must choose which activities they carry out during the use of social media. These elements are developed in the literature review. To research practices of internal digital communications and level of success of internal communication across types of organizations, a quantitative survey between professionals from Spanish companies was conducted.\nFindings: The results revealed that the hierarchical level, participation in the strategic decision process and dedication of the communication practitioner, alongside with the use of face-to-face and online communication channels, and the perceived importance of communication activities conducted through social media platforms are the key factors that influence the quality of internal communication. However, no significant relationship between the level of success and the use of social media channels was found.\nResearch limitations/implications: The current study has several limitations that should be noted and addressed in future research. The main limitation lies in the fact that the dependent variable - success in communication - is entirely based on communication professionals' perceptions. Replication studies can be conducted to cross-validate the results obtained from this study using business outcome metrics to measure the communication effectiveness. A further limitation is related to the sampling procedure. Obtaining a representative sample of communications professionals in Spanish companies poses a number of limitations due to the impossibility of having accurate data on the total population.\nPractical implications: The findings of the current study provide important implications for public relations professionals on what (i.e. the content) and how to (i.e. the channels) communicate within an organization. In general, internal communicators should move from historical roles as information producers and distributors to advisory roles in strategic decision making. Additionally, the implementation and use of ISM should be carefully revaluated. Communication practitioners should examine and address the difficulties involved in choosing the correct channels, devoting the necessary time to their adequate management, analyzing employees' feedback and improving the engagement.\nOriginality/value: The study shows that internal communication practices in Spanish companies are changing. The model used in this research can be applied in an individual organization to evaluate what factors improve the communication of its employees and carry out additional research in other countries or types of organizations to identity new challenges.\n",
"role": "assistant"
}
] |
Shaping knowledge management: organization and national culture
|
[
"Knowledge management",
"National cultures",
"Organizations",
"Cross‐cultural management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Competitive advantage has been shown increasingly to rely on the effective management of knowledge (Bukowitz and Petrash, 1997; Stewart, 1997). This is particularly relevant for multinational firms, which may adapt not only the organizational structure of the subsidiary to the host country, but also its knowledge management practices in expanding abroad. Indeed, Drucker et al. (1997) have identified \"harnessing the intelligence and spirit of people at all levels of an organization to continually build and share knowledge\" as a top priority for firms wishing to succeed in today's competitive environment (Chow et al., 2000). Considering that the perceptions and behaviors associated with the management of knowledge greatly depend on those who put it into practice, i.e. knowledge workers, KM can no longer be solely attributed to the firm's management, but should take into account local personnel who may create it and apply it in its daily tasks.Turner and Makhija (2006) suggest that \"any given control mechanism has the capacity to affect both the nature and flow of knowledge in a firm by the manner in which it processes particular attributes of knowledge\" (p. 213) and they contend that these control systems influence each stage of the knowledge management process. In addition, organizational climate and organizational structure have been recognized as shaping knowledge management (Chen and Huang, 2007). In fact previous research has shown that the very organization of the firm (Magnier-Watanabe and Senoo, 2008) and national culture (Moller and Svahn, 2004; Martinsons and Davison, 2007) constrain and influence the nature of KM.This research builds on the framework introduced by Magnier-Watanabe and Senoo (2008) in which organizational characteristics - specifically structure, membership, relationship, and strategy - affect knowledge management - namely knowledge acquisition, storage, diffusion, and application. Their research acknowledges the strong role of management in the firm's handling of knowledge and suggests that practitioners can increase the yield of KM when it is integrated upstream into the elementary business processes rather than when left voluntary (Magnier-Watanabe and Senoo, 2008). This framework has already received a qualitative validation based on the case study of a Japanese company, and this paper now seeks to confirm it using a questionnaire survey. This research falls under the category of intermediate theory (Edmondson and McManus, 2007) proposing relationships between established constructs, and it is therefore suitable for hybrid data analysis consisting of both qualitative and quantitative assessment.\nTheoretical background: Organization and knowledge management strategies\nConcepts and hypotheses: Knowledge management, organizational characteristics, and national culture\nQuantitative analysis: Methodology, sample selection and data collection\nDiscussion: Organizational characteristics and knowledge management\nConclusion: Knowledge management has long been considered from the onset in the firm's business strategy and has been shown to have a positive effect on innovation. However, other factors beyond the deliberate initiatives taken by the firm also affect knowledge management. Globalization for instance requires that multinationals consider the perceptions and behaviors of local staff in the management of knowledge. Existing research has promoted knowledge to the status of valuable strategic asset and demonstrated the benefits of a strong link of its management with organizational characteristics. Furthermore, some recent studies have highlighted the importance of national culture in decision-making styles and thus in the tailoring of knowledge management systems as well.This framework linking organizational characteristics and national culture to knowledge management had received a first justification using a case-study approach with a qualitative comparative method (Magnier-Watanabe and Senoo, 2008) and has now been confirmed with a quantitative approach. Among the predictors of knowledge management beyond the realm of deliberate measures within the firm, the data shows organizational characteristics to exert a stronger influence than national culture. Hence, changes in organizational characteristics such as structure and relationship in particular, rather than adjustments in the composition of employees' nationalities, will have a stronger impact on the resulting management of knowledge. Furthermore, because the nature of relationships (systematic or ad hoc) among workers can be attributed to the type of organizational structure (vertical or horizontal), practitioners should focus on structure, which appears as a worthy starting point to positively affect knowledge management.One limitation to this research lies in the selection of organizational characteristics since other researchers could identify other factors in the firm's organization affecting the management of knowledge. In addition, as this paper centered on a single company in the pharmaceutical industry, future research will attempt to confirm the validity of this framework in other industries.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to confirm quantitatively the previous finding that organizational characteristics influence knowledge management, and to assess whether the national culture of knowledge workers equally affects the management of knowledge.\nDesign/methodology/approach: - Based on data gathered from a questionnaire survey of a Japanese pharmaceutical company's 14 foreign subsidiaries, the effects of organizational characteristics and national culture on knowledge management were tested using multiple regression analysis.\nFindings: - Although organizational characteristics and national culture were found to affect knowledge management, the data showed organizational characteristics to be a stronger prescriptive factor compared with national culture.\nResearch limitations/implications: - Because this research centered on a single company in the pharmaceutical industry, future research should attempt to confirm the validity of this framework in other industries.\nPractical implications: - Changes in organizational characteristics, such as structure and relationship in particular, rather than adjustments in the composition of employees' nationalities, will have a stronger impact on the resulting knowledge management.\nOriginality/value: - This framework linking organizational characteristics and national culture to knowledge management had received a first justification using a case study approach with a qualitative comparative method and has now been confirmed with a quantitative approach. Among the predictors of knowledge management beyond the realm of deliberate measures within the firm, the data show that organizational characteristics exert a stronger influence than national culture.\n",
"role": "assistant"
}
] |
A case study of a simultaneous integration in an SME: Implementation process and cost analysis
|
[
"ISO 14001",
"ISO 9001",
"Cost analysis",
"Small and medium-sized enterprise",
"Management system integration"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In recent years the studies analyzing the implementation and integration of management systems (MSs) have increased (see, e.g. Douglas and Glen, 2000; Karapetrovic et al., 2006; Salomone, 2008; Bernardo, 2014; Domingues et al., 2015; Bernardo et al., 2017). This is mainly because of the success in the implementation and certification of the quality MS (QMS) ISO 9001 and the environmental MS (EMS) ISO 14001. According to the last available data (ISO, 2016), the growth rate of certifications for ISO 9001 has decreased 0.2 percent and for ISO 14001 has maintained an 8 percent as in the previous year (ISO, 2016).\n2. Literature review: This section is divided into three subsections. First, the different aspects of the integration process are analyzed. Second, the existing studies of IMS in SMEs are reviewed and finally, the cost analysis in IMS.\n3. Research methodology: In order to achieve the research objective a case study of a single organization was done. The organization analyzed is a family business with ten employees. Its activities are related to the construction sector, specifically of heating and air conditioning. It is located in Mollet del Valles, in Barcelona (north of Spain).\n4. Results: In this section, the implementation of the IMS is presented first, then its cost analysis and finally, the interviews with the manager and the consultant.\n5. Conclusions: The aim of this paper was to analyze the simultaneous implementation of an IMS in a SME, together with the analysis of the costs incurred. The single case study presented allows posing the following conclusions, although they should be taken with caution because a single case was realized.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze the simultaneous implementation of an integrated management system (IMS) in a small and medium-sized enterprise (SME) and its impact on costs.\nDesign/methodology/approach: Based on a case study, the paper presents and analyzes the implementation of the ISO 9001 and ISO 14001 management system (MS) standards. The organization analyzed is an SME of the heating and air conditioning sector that had no MSs implemented (common in the sector) and decided to implement an IMS based on both function-specific MS standards (MSSs) and also achieving the separated certification. The analysis of expected and real costs is also presented.\nFindings: The results show that a simultaneous integration is possible and the cost analysis evidences the improvement of the organization's efficiency. The certificates for both MSSs were also achieved. These findings help in confirming theoretical statement posed in the literature of the integration of MSs: integration strategy, methodology and level. The organization values this integrated implementation as very positive internally and externally, as this is one of the first organizations of the sector implementing and integrating the ISO 9001 and ISO 14001.\nOriginality/value: Although studies analyzing empirically the integration of MSs have been widely published in recent years, to the best of the authors' knowledge, this is one of the first papers analyzing the implementation of an IMS directly and analyzing its costs. In addition, this paper provides evidence that those organizations implementing integration later than others, can learn from the beginners' experience.\n",
"role": "assistant"
}
] |
C.U.L.T.U.R.E.: marketing education in the age of cultural diversity
|
[
"Cultural diversity",
"Marketing education",
"Assessment practices",
"Learning approach",
"Student attitudes",
"Student perception"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: It is essential for marketing educators in Universities to prepare students with the core competencies required to compete in the global marketplace (Ackerman et al., 2003; Davis et al., 2002; Kelley and Bridges, 2005; Walker et al., 2009). Marketing educators are expected to provide students with a skill base which, beyond specific disciplinary expertise, allows them to learn, retain and disseminate information more efficiently and effectively. This is achieved through critical thinking, problem solving, analysis, synthesis of ideas and the drawing of inferences and conclusions (Ackerman et al., 2003; Davis et al., 2002; Gapp and Fisher, 2012; Jackson and Chapman, 2012; Kelley and Bridges, 2005; Walker et al., 2009). Compounding the modern educational challenge is the dramatic increase in education globalisation, as evidenced by the increase in the number of ethnically and linguistically diverse University marketing students in Australia and other western countries (Pennycook, 2005; Singh and Doherty, 2004; Teo et al., 2012).\n2. Literature review: Learning approach\n3. Research methodology and procedures: Context\n4. Results and discussion: After focus groups were compared and contrasted it was evident that students generally agree that assessment practices at tertiary institutions bring their own benefits, constraints and challenges to studying in a culturally diverse environment. By applying Hall's (1976) low-context/high context cultural typology and Hofstede's (1986) cultural dimensions as a lens for exploring similarities and differences between students, themes encompassing learning environment, learning approaches, assessment tasks and their effectiveness emerged.\n5. Recommendations and conclusion: Based on the findings presented, it is proposed that clear guidelines be adopted in order to implement effective assessment practices and foster a vibrant, inclusive learning environment which meets the educational needs and wants of a culturally diverse student cohort. It is recommended that the principle of C.U.L.T.U.R.E. be integrated into learning approaches and assessment practices across Universities at the tertiary level:\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the role Australian University marketing students' cultural backgrounds play in their learning and their perceptions of assessment and explores whether current assessments appropriately address the differing needs of a culturally diverse student population.\nDesign/methodology/approach: The paper adopts a qualitative approach utilising five focus groups, each comprised of 12 students.\nFindings: Results indicate learning environments, learning and assessment approaches and assessment tasks each bring their own benefits, constraints and challenges to studying in a culturally diverse environment. Principles are presented for adoption by marketing educators in order to foster a vibrant, inclusive learning environment which meets the educational needs and wants of a culturally diverse student cohort.\nResearch limitations/implications: The number of students representing different global regions or countries limited this study. With the exception of students from Australia and the Asian region, there were minimal students representing other cultural backgrounds despite every attempt being made to be culturally inclusive across global regions.\nPractical implications: The paper presents the principles of C.U.L.T.U.R.E. and recommends their integration into learning approaches and assessment practices across Schools and Faculties at the tertiary level.\nOriginality/value: This paper fulfils an identified need to study a culturally diverse student cohort's perceptions and attitudes towards learning approaches and assessment practices and their perceived relevance to the provision of core graduate business and generic skills necessary for employability in the global marketplace.\n",
"role": "assistant"
}
] |
Lean information management: the use of observational data in health care
|
[
"Lean production",
"Problem solving",
"Quality improvement",
"Continuous improvement",
"Operations management",
"Manufacturing systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Lean is a term that was first coined Womack, Jones and Roos (Womack et al., 1990; Womack and Jones, 1996) to describe the Toyota Production System (TPS) and the steps to continuously improve the efficiency and effectiveness of a system by driving out waste. Womack and Jones identified that in order to meets its customers' needs an organisation must identify what its customers think of as value. Once this is clear, the organisation can work to eliminate non-value adding process steps or \"waste\", make the remaining steps in the process flow, implement pull systems where flow is not possible and work continuously towards perfection.Taicho Ohno (Ohno and Bodek, 1988), an executive at Toyota, led the development of the TPS and the concept of the Gemba Walk. The Gemba Walk is an opportunity for staff to stand back from their day-to-day tasks to walk the floor of their workplace to identify wasteful activities. Waste might include producing too much too early, duplication or rework, mistakes and errors, waiting, unnecessary stock, transportation of people or materials over long distances, or unergonomic work environments. The first step is to be able to see the waste and then to eliminate it.Without data it is difficult to identify the root cause of a problem and to make accurate decisions to reach a solution. Ohno demonstrated repeatedly that there is no substitute for observing the problems yourself. This idea was repeatedly advocated by the quality geru W. Edward Deming (2000) who is credited with saying:... in God we trust - everyone else must bring data.In order to make well founded and informed clinical or managerial decisions good quality data needs to be readily available. For the treatment of individual patients this might include access to diagnostic results, healthcare records, drug charts and appropriate consultations with specialists when necessary.Making decisions that impact the way that an organisation functions or meets the continually changing needs and expectations of patients requires a different kind of data. This data needs to be conducive to a detailed analysis of how the current system works, the outcomes delivered and the areas that offer an opportunity for improvement. Accuracy, timeliness, quality and availability are important standards, which traditionally within healthcare settings are not met.The purpose of this paper is to compare and contrast traditional data collection methodologies employed in health care with more practical observational methods which are closely aligned with Lean thinking. The paper shows that, when combined with problem solving, observational approaches achieve demonstrable improvements in clinical outcomes, productivity and efficiency. It describes how some of the obstacles to obtaining good quality data can be overcome, as well as some of the limitations of this alternative method of collecting data. The paper describes the changes in mindset and behaviour required to adopt the observational methods.\n2. Approach: The paper's approach is to describe and evaluate case study examples on the use of observational data in the National Health Service in the UK. These cases are then used to derive generic principles about the wider application of observational data in health care.\n3. Traditional data collection methodologies employed in health care: Table I shows some of the systems currently in use to collect data in the National Health Service in the UK.Countless other systems are in use in finance, human resources, occupational health, estates and other support services.\n4. Limitations of traditional data collection methodologies: Data tend to be reported in table format which makes visual analysis of trends or the current position difficult. Averages and simple comparisons from one year or month to the next distort the true picture.Data are often reported a month or two after it is collected which makes it difficult to identify why problems occurred and to take action immediately to prevent problems from happening again. There are numerous committees and meetings where data is discussed, the same problems arise and no action is taken.There is a target-driven culture of reporting data upwards that is requested by the Department of Health. Often the front-line staff do not understand why they are collecting and reporting the information, and receive little or no feedback on their performance. This kind of approach discourages front-line staff from collecting and monitoring the data and taking responsibility for improving their service.The way in which data are reported is often reactive. It is common to hear about services that are overspent or complaints after the event. The authors worked with an ophthalmology service which was constantly under scrutiny by the trust board for the complaints it received. The result of this was that the manager spent about half of their working week formally responding to the complaints. Most complaints were about waiting times in the department. Through observation of the workplace with dedicated time, staff within the unit were able to redesign processes to reduce waiting times by 50 per cent and reduce the level of complaints.Wards and departments collate data in different ways and many departments collect the same data in different ways. This causes duplication and a lack of clarity about what data is available, its accuracy and where it can be found. Some specific examples of the limitations of traditional data collection methods include:*Accuracy of length of stay data - 58 patients on a ward with 26 beds because nurses had not discharged them on the computer.*Waiting for the results of blood tests - not knowing where a particular sample is in the system and when results may become available.*Unable to track images - the diagnostic image results are on a server but clinicians are unable to access them leading to delays in diagnosis.*No usage data for procurement - the only available data is spend, to find out the amount of a specific product purchased the hospital must contact the supplier and ask them.*Budgets not representing actual costs or spend - no uplift on non-pay spend for eight years combined with a 3 per cent year on year cost improvement programme, inflation around 3 per cent for eight years, budget decreased by 3 per cent every year at 2000 values whilst costs increased. This led to massive overspends.*No data on the time the patient entered the theatre, the time the operation started or finished.*No good performance metrics within theatres indicating whether organisations are efficient or productive. The key metric is theatre utilisation which is time used to operate divided by time available to operate. For example a surgeon operates for three hours of a possible four providing utilisation of 75 per cent. This does not take in to account any level of productivity as one surgeon might perform two operations in 2.5 hours giving a utilisation of 62.5 per cent compared to another surgeon that does one of the same type of operation in three hours with a utilisation of 75 per cent.*Poor quality metrics - the focus of metrics tends to be on waiting times, complaints and finance rather than infections, outcomes and readmission rates.*Boards focus on finance not quality - quality is often perceived as being costly.Traditional data collection methodologies are unable to identify that a patient waited eight hours in a bed because pathology lost their blood sample. Having blood drawn for a second time, the patient waited a further three hours for the result which indicated that they were fine and could go home. The result is that a patient occupied a bed for 12 hours and consumed physician, nursing and other staff time for no material benefit.There are no measures in place that allow the organisation to identify this earlier. Traditional methods are unable to identify the level of duplication in information, by whom and when, such as completing a clinical assessment form in the emergency department and then completing a similar assessment form in the medical assessment unit containing much of the same information.Health care is incredibly complex and an enormous quantity of data is collected. Despite this it is still not possible to identify easily how an organisation is performing in terms of quality, cost and delivery.\n5. Observational data: A different approach to traditional data collection methods is that pioneered initially by Frederick Taylor and W.L. Gantt (Taylor, 2007) and then later by F.B. Gilbreth (Gilbreth and Carey, 2005) and Taichi Ohno. Rather than collecting the data that are currently available, the authors advocate physically going to the location of interest to directly observe what takes place and to record this in a variety of formats including video, documentation and photographs.This direct observational approach has been used with success in different areas of medicine including the counting of instruments into and out of theatres (Greenberg et al., 2007), the insertion of central lines in intensive care units (ICUs) (Pronovost et al., 2006) and the direct observation of activity within an ICU (Donchin et al., 2003).The next section details the results of a number of case studies of direct observational data collection.5.1 The application of video techniques in a sterile services department\n6. Results: Each of the case studies presented in the previous section resulted in fundamental changes in processes from observing the current process in a way that staff can relate to.Rather than management, analysts or management consultants reviewing the process and making recommendations back to the organisation, the data was collected by shop floor staff. As a result the need for change and urgency was more apparent.\n7. Limitations of methodology: There are limitations to the use of observational data and the methodology outlined in this paper including:*It is resource-intensive - it can require a significant number of people to collect sufficient quantities of data for it to be of use.*There may be a Hawthorne effect (Draper, 2008) - simply by observing people their performance may improve.*There may be a limited time period of analysis, i.e. 1 day, 1 shift - this might be a misrepresentation of what happens day-to-day as it is quite possible that the day upon which the data is collected is an anomaly.*People may feel as though they are being spied on so behave differently - this is a problem that arises in any type of work where a person, department or group is subjected to scrutiny in order to be able to improve.The authors have demonstrated through the case studies in this paper that the methodology is valid in a range of clinical settings for examining a process to identify opportunities for improvement.\n8. Accuracy and usefulness: The authors believe that if the limitations of the methodology are minimised the methodology can be extremely beneficial.The benefits of using observational data methodology are that:*front line staff see what actually happens with their own eyes;*there is an opportunity to stand back from daily routines to see things differently;*if 15 people observe the same thing the evidence is irrefutable;*it creates a burning platform and sense of urgency - one of Kotter's (Kotter, 1996) 8 steps to change; and*it helps people to move away from the acceptance of mediocre to the need for continuous improvement.In the authors' experience the usefulness of the methodology outweighs its limitations. The power of staff from a process witnessing the way the process works as an observer rather than a participant generates a completely new perspective on the system. This often results in staff moving from a position of denial of the need for change or acceptance of the current state, to the realisation of the need for change and innovation.\n9. Conclusions: Observational data have as much of a place today in improving the delivery of healthcare as it did when Taylor and Gantt first used it in industry. From its application by Gilbreth, later to be documented in the film Cheaper by the Dozen, to more recent applications in healthcare by people such as Atul Gawande, Peter Pronovost and Donchin et al.; the methodology is as valid today as it was a century ago. The case studies described in this paper support the reintroduction of observational techniques used by the early pioneers of productivity.The power of observational data has increased with time. In an era in healthcare where information technology bombards the sector with performance data, outcomes and every other metric conceivable; the use of basic data collection methodologies can provide a level of information that computers are simply unable to. The methodology allows instant root cause analysis, allows those participating to feel involved in the change and provides rapid feedback. These qualities, despite the limitations, make it as valuable today as when it was first used.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to compare and contrast traditional data collection methodologies employed in health care with more practical observational methods which are closely aligned with Lean thinking. When combined with problem solving, observational approaches achieve demonstrable improvements in clinical outcomes, productivity and efficiency. The paper aims to describe the changes in mindset and behaviour that are required to adopt the observational methods.\nDesign/methodology/approach: - The approach is to describe and evaluate case study examples on the use of observational data in the National Health Service in the UK. This is then used to derive generic principles about the wider application of observational data in health care.\nFindings: - Traditional data collection methodologies are often insufficient to expose the root cause of a problem and therefore may result in little or no action. The observational methods identify the root cause and as such offer a much more practical and real-time way of solving process-related problems.\nPractical implications: - The observational methods of collecting data described here offer staff at all levels of the organisation practical approaches to preventing mistakes and errors in health care processes.\nOriginality/value: - The case studies described here support the reintroduction of observational techniques used by the early pioneers of productivity. The originality of the paper is in the use of these observational methods in a wide range of clinical settings to provoke changes in working practices.\n",
"role": "assistant"
}
] |
Trickle-down effect of benevolent leadership on unethical employee behavior: a cross-level moderated mediation model
|
[
"Benevolent leadership",
"Employee unethical behavior",
"Trickle-down effect",
"LMX",
"Ethical climate"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Unethical employee behavior exists universally in various organizations, including business organizations, government agencies and academia (Kish-Gephart et al., 2010). In recent years, unethical behaviors have occurred frequently in domestic and international organizations. Research implemented by KPMG, which contains 5,065 American managers and employees, shows that in the past year, 74% of people observed unethical behaviors in their organizations. Research by the Compliance and Ethics Leadership Council on 1,752 managers and employees from large companies indicates that in the past year, 16% of people have observed harassment in their organizations, 15% have observed discrimination, 11% have stolen and 7% have forged their spending. As a result, unethical employee behaviors have attracted widespread attention in business circles and academia. Unethical employee behavior is defined as behavior that violates the moral code widely accepted by people (Trevino et al., 2015) and has strong concealment and dangers, breeds corruption easily, leads to inefficiencies and increases the losses of organization (Pierce and Aguinis, 2016). Many studies have indicated that unethical employee behaviors cause great harm to organizations (Sguera et al., 2018; Bonner et al., 2017). Therefore, research aimed at finding the causes of unethical employee behavior, reducing such behavior, and helping enterprises to develop healthily and in a sustainable way is especially urgent. In the past few years, research has shown that there are factors that influence employees' unethical behavior, such as time pressure (Koh et al., 2018), organizational climate (Zaal et al., 2019), challenging performance goals (Welsh et al., 2019). Leadership is a key contextual factor and has been proved to play an important role in employee unethical behavior (Paterson and Huang, 2019). However, the existing researches on the influence of unethical employee behavior mainly focus on the leadership style proposed in the western context. Relatively little is known about the leadership style in non-Western context, such as the leadership style in Chinese context (benevolent leadership).Benevolent leadership is a leadership style rooted in traditional Chinese culture, which refers to the leader's individual, comprehensive and long-term care for his subordinates and their families, including individual care, consideration and tolerance (Zheng et al., 2000). There is a clear difference between benevolent leadership and the leadership style proposed in the western context, for example, servant leadership includes guiding, supporting democracy and delegation (Chughtai and Aamir, 2019). Transformational leadership includes moral, vision to inspire, personalized care and leadership charm (Parveen and Adeinat, 2019). Ethical leadership mainly refers to that leaders require themselves and their subordinates to behave in an ethical way (Haar et al., 2019).It is of great theoretical and practical significance to study the relationship between benevolent leadership and unethical employee behavior. Theoretically, existing studies on the consequences of benevolent leadership mainly focus on voice behavior (Chou and Long, 2014), the creativity of employees and employee performance, and the research results all show the positive effects of benevolent leadership. This paper extends the research on the impact of benevolent leadership on unethical employee behavior. At present, antecedents of unethical employee behavior studied from the perspective of leadership mainly focus on ethical leadership (Kalshoven et al., 2016; Joosten et al., 2014) and responsibility leadership (Graham et al., 2018), which are mainly leadership styles in Western-oriented situations. This paper is among the first to study the negative influence of Chinese traditional benevolent leadership on unethical employee behavior, which enriches the theoretical research on unethical employee behavior in the Chinese context. In practice, on the one hand, benevolent leadership is a representative dimension of paternalistic leadership, and its research is more helpful to Chinese organizations in Chinese society and other eastern countries; on the other hand, unethical employee behaviors often occur in various organizations, and it is meaningful to explore their roots to reduce such behaviors. Therefore, conducting research on the relationship between benevolent leadership and unethical employee behavior can help us find solutions to reduce unethical employee behavior. From a review of relevant literature, we found that there may be a relationship between the two behaviors.The existing literature studies the influence of direct supervisor leadership behaviors on employees, including the benevolent leadership of supervisors. However, the formation of the employee's direct supervisor leadership style is likely to be influenced by the leadership style of senior leaders (managers); thus, it is extremely necessary to study how the supervisor leadership style is affected by the leadership style of senior leaders and how it ultimately affects the employee's behavior. This study attempts to make some contributions in this respect. Therefore, this paper discusses how the benevolent manager leadership can be transmitted to the employee's unethical behavior through the benevolent supervisor leadership. On this basis, this paper studies the situational factors that affect the trickle-down effect, that is, the moderating effect of LMX, and argues that benevolent manager leadership is more likely to affect benevolent supervisor leadership under high LMX than under low LMX, because under high LMX, supervisors are more likely to imitate managers' leadership style. Finally, we believe that the influence of benevolent leadership on unethical employee behavior is moderated by the organization's ethical climate, which can help enhance the influence of benevolent leadership on unethical employee behavior, for under positive ethical climate employees are influenced by climate to reduce unethical behaviors.\nTheory and hypotheses: Benevolent supervisor leadership and unethical employee behavior\nSample and measures: Sample and procedure\nResults: Data aggregation test\nConclusion and discussion: Through empirical analysis we draw several conclusions: (1) Benevolent leadership is negatively related to unethical employee behavior. This conclusion is consistent with the research results of other scholars. For example, studies by Li et al. (2015a) and Tian and Huang (2014) show that benevolent leadership is an important factor affecting the psychology and behavior of employees. Combined with this finding, we believe that positive leadership styles tend to generate satisfaction and reward in employees and ultimately lead to positive results. This result enriches the research on the concepts of benevolent leadership and unethical employee behavior. In contrast to previous research, this research focuses on leadership style in the Chinese context and unethical behaviors frequently occurring in organizations. The conclusion drawn from the research on the relationship between the two can be used for reference in both theory and practice to improve the leadership style of managers in Chinese organizations and reduce unethical employee behavior. (2) Benevolent supervisor leadership mediates the relationship between benevolent manager leadership and unethical employee behavior. Results of this study tell us that the benevolent leadership behaviors of employees' supervisors do not occur in many situations but are traced back to the benevolent leadership behaviors of managers. Meanwhile, the benevolent leadership behaviors of the managers do not directly affect the employees but rather influence the behaviors of the employees through the leadership style of their supervisors. This research helps us to better understand the root causes of unethical employee behavior and the indirect process of leadership's influence on employees' behaviors. (3) LMX moderates the relationship between benevolent manager leadership and benevolent supervisor leadership and mediating effect of benevolent supervisor leadership. LMX is sometimes used as a moderating variable to regulate the psychological or behavioral relationship between supervisors and subordinates; for example, Seo et al. (2018), Lindsey et al. (2016), and Schwarz et al., (2017) used LMX as a moderating variable in their research. This outcome tells us that it is necessary for managers to establish a high-quality LMX with supervisors to strengthen their influence on them. (4) Ethical climate moderates the relationship between benevolent supervisor leadership and unethical employee behavior and mediating effect of benevolent supervisor leadership. This result is consistent with the research conclusions of Shin (2012) and May et al. (2014), who also believe that no matter what type it is, ethical climate is conducive to some positive effects, especially on ethical behaviors. Creating a good organizational climate can help guide employees' positive behaviors and curb negative behaviors.Theoretical implications\n",
"role": "user"
},
{
"content": "Purpose: Conducting research on the relationship between benevolent leadership and unethical employee behavior can help us find solutions to reduce unethical employee behavior. This paper also discusses how the benevolent manager leadership can be transmitted to the employee's unethical behavior through the benevolent supervisor leadership and the moderating effect of LMX and ethical climate.\nDesign/methodology/approach: Through a questionnaire survey of 406 pairs of leaders, supervisor and employees, the authors use data aggregation test, confirmatory factor analysis, descriptive statistics and multilevel model (HLM) to test our hypothesis.\nFindings: (1) Manager supervisor leadership is negatively related to unethical employee behavior, (2) benevolent supervisor leadership mediates the relationship between benevolent manager leadership and unethical employee behavior, (3) LMX positively moderates the relationship between benevolent manager leadership and benevolent supervisor leadership and moderates the mediating effect of benevolent supervisor leadership, and (4) ethical climate positively moderates the relationship between benevolent supervisor leadership and unethical employee behavior and moderates the mediating effect of benevolent supervisor leadership.\nOriginality/value: First, based on previous studies, this study further proves that benevolent leadership is a popular positive leadership among the three dimensions of paternalistic leadership that extends its influence to unethical employee behavior. Second, the study traces the source of the benevolent leadership of employees' supervisors and reveals the action mechanism of how benevolent manager leadership affects unethical employee behavior (trickle-down effect). LMX and the ethical climate provide the organizational context of the trickle-down effect and the occurrence of unethical employee behavior.\n",
"role": "assistant"
}
] |
The new neck of femur fracture target: experience in a district general hospital
|
[
"Organisational performance",
"Clinical guidelines",
"Capacity management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Timely, efficient and co-ordinated multi-disciplinary neck of femur (NOF) fracture patient management can reduce morbidity and mortality (Fox et al., 1994; Holt et al., 1994; Rogers et al., 1995; Sexson et al., 1987; Villar et al., 1986; Zuckerman et al., 1995). Several guidelines advocating this approach have been published (Hip fracture - NICE, 2011; Blue book - BOA, 2007). However, their implementation in the UK has been patchy. Recently, the issue was addressed by new government incentives, the \"best practice tariff\" (BPT) for hip fractures that came into effect in April 2010, which meets the commitment to High Quality Care for All in Lord Darzi's (2008) NHS Next Stage Review. To achieve BPT, Lord Darzi advocated two key improvements:\nMethod: Data covering August 2009 to July 2010 were collected retrospectively from the hospital theatre information system and repeated for four-months from November 2010 to February 2011, when a designated Sunday trauma list was implemented. Data were an anonymous summary, removing the need for ethical approval. The theatre information system requires the dataset to be complete, including:\nResults: Total trauma cases undergoing surgery from August 9 to 10 was 1,903. Total NOF fractures requiring surgery from August 9 to July 10 was 442. The mean cases operated on per trauma list were 3.16. Half the hospital's (operative) trauma workload is dealt with on designated trauma theatre lists, with 32 per cent completed on the emergency lists and the remaining 17 per cent on elective lists where there is space or where space is created through cancellations (Appendix). Focusing on fracture NOF patients shows a greater proportion being dealt with on designated trauma (64 per cent) and elective lists (24 per cent) compared to general trauma. If we exclude weekends and bank holidays then the figures change - only 2 per cent of NOF cases were on weekday emergency lists, leaving 71 per cent on designated trauma and 27 per cent on elective lists.\nDiscussion: Delivering good care for patients with hip fracture is challenging and involves many health professionals including surgeons, anaesthetists, ortho-geriatricians, nurses and rehabilitation staff. The British Orthopaedic Association/British Geriatric Society BOA-BGS Blue Book care standards for patients with fragility fractures and the National Hip Fracture Database audit (NHFD) encourages fracture unit staff to develop interdisciplinary services for rehabilitation and secondary prevention to improve matters for these patients. The BOA-BGS Blue Book second edition confirms that delivering care to rising fragility fracture patients (BGS, 2010) is paramount. While efficient and effective fracture NOF pathways will no doubt help to improve patient care, a major hurdle will be incorporating these patients into an acute orthopaedic trauma service that covers several cases requiring operative intervention with varying urgency. Rising trauma cases, European Working Time Directive restrictions, NCEPOD recommendations and elective commitments makes it a challenging task to achieve the correct balance.\nConclusions: Our hospital trauma service has insufficient capacity to accommodate trauma demand and fluctuations, despite introducing a Sunday trauma service in November 2010. Managing the patient overspill has a significant knock-on effect on the elective service, which has been reduced but not eliminated by adding a Sunday trauma list. Using the emergency list staff when free, particularly when the morning session starts is one way to ensure trauma list efficiency; however, as fracture NOF workload increases, trauma operating time will have to grow to accommodate the extra workload especially if hospital staff are to meet BPT targets. While it is clear from the literature that achieving BPT is good for the NOF patients, it is, however, difficult to assess whether the changes needed to achieve BPT is cost effective and enough of an incentive to push hospital managers into making the changes.\n",
"role": "user"
},
{
"content": "Purpose: The hip fracture \"best practice tariff\" (BPT) came into effect in April 2010. It advocated two key improvements: surgery within 36hrs of arrival in the emergency department; and multi-disciplinary care directed by ortho-geriatrician from admission to discharge. The aim of this paper is to look at the 36 hours to operation target and its implications for orthopaedic department trauma service staff in a busy district general hospital, and to evaluate the measures implemented to meet the target.\nDesign/methodology/approach: Trauma-list data, collected from a theatre management system, was compared with trauma patients placed on elective and emergency lists, before and after designated daily trauma lists were implemented.\nFindings: After a designated daily trauma list was introduced, a significant rise (from 56 per cent to 85 per cent) became evident in the proportion of patients operated on within 36hrs, between November 2010 to February 2011, while hip fracture cases managed on the elective list fell from 24 per cent to 17 per cent.\nPractical implications: Despite adding a half-day trauma list, the trauma service has insufficient capacity to achieve the new BPT for all hip fracture patients in the hospital. Therefore, there is a significant knock-on effect for managing patient overspill on elective services. Will the significant changes in service provision designed to achieve this BPT be cost effective?\nOriginality/value: This paper aims to answer how busy department staff address an issue that professionals in every English hospital are facing.\n",
"role": "assistant"
}
] |
Shopping from a child's perspective: an anxiety-generating experience?
|
[
"Child",
"Retail",
"Shopscapes",
"Drawing",
"Anxiety",
"Experience"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Because of their increasing spending power, children have become explicit targets for retailers who view the market segment they represent as a new opportunity to enlarge their existing clientele. Children are often depicted wandering around stores from a very young age, with their parents, with friends ... or even sometimes alone. At around six-year old, marketing literature acknowledges that children are skilled enough to go shopping independently (McNeal et al., 1998). For decades, the academic research conducted in the field of child consumption has focussed mainly on the skills children need to master to become competent consumers, on the categories of products they like buying, on their preferred brands and stores, on their active influence on purchasing practices within their family group, etc. (McNeal, 1964, 1987, 1992; Wells and La Sciuto, 1966; Fady, 1979; Reece, 1986; Bree, 1993; Acuff, 1997; McNeal et al., 1998; Roedder-John, 1999; Valkenburg and Cantor, 2001; Ozgen, 2003; Chan, 2005). This academic research perspective relies on empirical data obtained through enquiries or surveys mainly aimed at parents and caretakers (Ekstrom, 2006). As Cook (2008) underlines, children hardly appear in the abundant literature about consumption practices, possibly because investigating children's perspectives raises unusual methodological issues. As works based on children's interviews or in-store direct observation of children are scarce, there is a need for innovative research putting children at the very heart of the investigative perspective and attempting to hear children's own views about their consumption experiences.\nChildren in stores: Previous research has allowed the academic community to gain insights into how children behave in stores. Children accompany their relatives to stores and begin to take part in shopping practices from a very young age. Dely-Gray et al. (2016) observe that young children, aged between three and six, are very active and involved in their parents' shopping practices and that they really enjoy this shared activity. During their visits to retail outlets, children gain information about the store and its features, the products and the brands, the store's layout and the role of packaging to differentiate products. Through vicarious learning based on direct observation and imitation of role models (Bandura, 1977) throughout the shopping process, children increase their knowledge of various retail environments and progressively develop their own consumer skills (McNeal, 1992; Roedder-John, 1999). Effectively the stores are physical places and social spaces that contribute significantly to the socialisation of children as consumers (Cook, 2008).\nMethodology of the research: In order to understand children's remembered experience of shopping, we first elaborated a conceptual framework based on the notion of \"scapes\" defined by Appadurai (1996) that were adapted later into specific research topics, such as \"workspaces\" (Whalen and Whalen, 2004), \"experiencescapes\" (O'Dell and Billing, 2005) or \"foodscapes\" (Brembeck and Johansson, 2010; Wenzer, 2010). Drawing also on the French Theory (Deleuze and Guattari, 1980; Deleuze, 1993), we use the definition of \"shopscapes\" (Nicol, 2014), i.e. the imaginary geographies each person or group of people builds based on their daily experiences and practices with reference to retail environments and activities. By definition, \"shopscapes\" are built unconsciously by each person through participation in everyday activities, and are culturally grounded and dynamic as they evolve constantly.\nFindings: an encompassing vision and understanding of the retail store: The general \"shopscape\" revealed throughout the 15 joint drawing-oriented interviews of siblings aged from 4 to 12 years showed that children have internalised an extensive body of knowledge about retailers and stores. The 15 \"shopscapes\" they produced included commercial brand names, a specific and sometimes quite accurate layout of the shelves within the store space, etc.\nDiscussion: Negotiations between siblings about the production of a joint shopscape\nManagerial implications: Proposing retail-tainment activities (Feenstra et al., 2015) for children, such as dedicated playgrounds or creative activities, in order to keep them occupied whilst their parents are shopping... does not appear to be a satisfactory answer to deal with children's eagerness to be involved in the actual shopping processes performed by their parents (Dely-Gray et al., 2014). In order to address this challenge, how could retailers improve their welcome of children to their stores and lessen the anxiety they might feel in such a stressful environment and offer them an educative space in which they could easily acquire consumption skills with the help of their parents?\nConclusion: Our exploratory research reveals some constitutive elements of children's \"shopscapes\" (Nicol, 2014). Our data collection is based on a remembering exercise performed at home and does not offer information about what children actually do in retail environments. Moreover, as we gave the children the specific scenario of buying a present for a friend's birthday, the information produced was principally about toy stores. Although children demonstrated an extensive body of knowledge about retail environments, surprisingly the recall of their shopping practices shed light on various anxiety-generating dimensions.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to offer some insight into how siblings aged between 4 and 12, engaged in a collaborative drawing activity at home, recall the shopping trips they have experienced.\nDesign/methodology/approach: Using a Vygotskian perspective, the data collection consisted of engaging 15 pairs of siblings in the production of a joint drawing of a shop of their choice. Drawing in pairs opens a Zone of Proximal Development (Vygotsky, 1978) where the younger child benefits from verbal guidance by the older one to achieve the common task. This situation enables the researcher to gain close access to children's knowledge about stores and to the words they use to describe their personal shopping experiences.\nFindings: This exploratory research reveals some constitutive elements of children's \"shopscapes\" (Nicol, 2014), i.e. the imaginary geographies they actively elaborate through their daily practices and experiences with regard to retail environments. In their communicative interactions when elaborating a joint drawing of the shop they have chosen, children demonstrate that they master a considerable body of knowledge about retail environments. Surprisingly, recalling their shopping practices sheds light on various anxiety-generating dimensions.\nResearch limitations/implications: The data collection is based on a remembering exercise performed at home and does not bring information about what children actually do in retail environments. Moreover, the children were asked to focus on buying a present for a friend's birthday, therefore the information gathered essentially relates to toy stores.\nPractical implications: This research underlines the necessity for retailers to endeavour to reduce some of the anxious feelings depicted and verbalized by children, by improving the welcome for children into their stores.\nSocial implications: There are also opportunities for retailers to invest in the consumption education area by guiding young visitors so that they learn how to behave as apprentice consumers in retail outlets.\nOriginality/value: The child-centric perspective of the study reveals new and surprising insights about the way children report their memorised shopping experiences.\n",
"role": "assistant"
}
] |
Patterns of graduate students' information seeking behavior: a meta-synthesis of the literature
|
[
"Graduate students",
"Information behaviour",
"Doctoral students",
"Students",
"Information",
"Information management",
"Behaviour"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Graduate students have information needs that are often very different than those of undergraduates. Their needs can be more sophisticated and complex. Understanding the information seeking behaviors of graduate students may better equip librarians, faculty, doctoral supervisors and administration to help shape those behaviors by offering appropriate and needed services and instruction.Wilson (1999) defined information behavior as \"Those activities a person may engage in when identifying their own need for information, searching for information and using or transferring that information.\" Graduate students have described themselves as being more confident and vested in the materials that they produce as a result of their research (e.g. Vezzosi, 2009). They should know their disciplines and how to engage in the research process far more adeptly than undergraduate students. Conversely, Kuhlthau (2004; cited in Barrett, 2005) states that undergraduate students cope more than seek information.This paper focuses in on information seeking as opposed to information behavior as a whole (e.g. data sharing, information handling skills). Behaviors that are specific to particular disciplines are treated separately within this paper. Studies from academic disciplines where students would typically be engaged in a dissertation, thesis or other capstone research project are included in this synthesis.This synthesis is aimed at identifying common patterns in the selection and use of resources by different disciplines and students at different stages in their graduate work. A synthesis is necessary as some results of information behavior studies, reported by individual institutions, are often a product of the culture or administration of the institution. For example, library service may be ineffectual, instruction may be weak or not discipline-specific, or outreach to the teaching faculty may have failed for any number of reasons not related to typical information behaviors. Faculty advisors play a dominant role in disseminating library research information to their students. Additionally, the literature on the information behavior of international graduate students presents a variety of issues that must be considered by those providing library services; limited language skills, and cultural differences, are just two examples of factors to consider.\nMethods: Meta-synthesis is a process of rigorously searching, evaluating, synthesizing and interpreting both qualitative and quantitative research on a specific topic. Meta-synthesis typically include highly structured search strategies that serve as a model for researchers looking for articles on the topic so that the search may be replicated at a future date. Inclusion and exclusion criteria are also identified (e.g. data range, topic focus, population). Critical appraisal based upon articulated criteria, data extraction and synthesis using a justified approach are also essential components of any type of review (Ankem, 2008; Brettle, 2009).Systematic reviews, in which there is a synthesis of quantitative findings, have been widely used by clinical librarians in order to evaluate and synthesize the findings of medical and healthcare research (e.g. Brettle and Long, 2001). Systematic reviews/meta-synthesis in academic librarianship have been conducted on information literacy instruction (Koufogiannakis and Wiebe, 2006) and computer-assisted versus face-to-face instruction (Zhang et al., 2007), for example. Only a few meta-synthesis using both qualitative and quantitative data have been conducted in library science. Most relevant to the study undertaken here, are those on information behavior research (Urquhart, 2011) and the information behaviors of women (Urquhart and Yeoman, 2010).Searching\nResults: In all 48 studies were identified as being about graduate students and information use or behavior; 25 of those studies used surveys as their method of data collection, and five of those interviewed a small sample of their participants. A total of 19 studies employed interviews and their primary methodology. One of those combined a photo diary with the interviews. The remaining studies used pre and post-tests, focus groups, evaluations of papers and citation lists, participant observations and think-alouds, along with interviews (see Table I for more information on research instruments employed in each study). Most of the surveys were deployed online, while two were conducted in person. Nearly all qualitative studies that used questionnaires in addition to interviews, administered the survey before the interview in order to collect demographic information and to capture discrete answers such as resource use and how often a student used a particular strategy or visited the physical library. Interviews were employed to glean more fine-grained detail about the information search process. Additionally, some qualitative studies quantified their results where possible. For example George et al. (2006) interviewed 100 students, but coded answers in such a way that allowed for frequency counts of statements that appeared to recur.Differences between masters and doctoral students\nPatterns of information behavior: Initial starting point for research\nConclusions and implications: The result of this meta-synthesis revealed that students at different levels of study, and from different disciplines, access different types of sources. Most students will consult the internet first when beginning their research, although doctoral students are more inclined to also consult their faculty advisors. In fact, people play a large role in helping students begin their research. After faculty, students consulted librarians and peers, although students acknowledge that they need to venture into the library themselves to find the resources that they need. Other researchers have pointed out that it is often the faculty themselves that are not adept at library research, and thus not equipped to help their students (Beile and Boote, 2005). Thus, the results of this synthesis suggest that it is faculty, particularly those who serve as advisors for graduate students, who should also receive advanced bibliographic instruction (Barry, 1997). RIN (2011) recommends that institutions make it easier for supervisors to keep up-to-date on training and services available. Additionally, the report encourages supervisors to discuss the information skills of their doctoral students so that they may be used to develop training opportunities.There are a wide variety of resources used by students at different levels and different disciplines. Formal resources range from bioinformatics databases to interdisciplinary research databases, particularly for doctoral students. Remote (electronic) access to research materials is of increasing importance to students; however, humanities scholars acknowledge that it is difficult to find all archives online. The literature also revealed different aspects of search strategies that have implications for library instruction and interface design. Prior knowledge, familiarity with search strategies and placement of search limiters (e.g. data limiters and proximity indicators) affects how students use databases.Lack of knowledge about the existence or use of a source or service, such as inter-library loan or a database, can inhibit the productive information behavior of a student. Additionally, computer/internet experience can contribute to a student's ability to effectively retrieve information (Korobili et al., 2011). Although other studies have pointed out students often have an inflated perception of their research abilities because they are adept at computer/internet use (Perrett, 2004). Lippincott (2005, p. 1.2) remarked that those \"born in the 1980s and later ... can generally multi-task ... [and] surf the web ... [however] they lack technology and information skills appropriate for academic work\".To some degree information literacy training can address some of the previously mentioned issues related to information behaviors. Kazlauskas (cited in Gratch and York, 1991), concluded that individualized instruction is more effective with graduate students than other methods, such as bibliographic instruction integrated within courses. Predominantly, this is because as graduate students evolve further into their field of study their library research needs change. Some institutions have instituted consultation services for graduate students to assist in completing a capstone project (e.g. Gratch and York, 1991). Offering training in a variety of formats (e.g. one-on-one consultations, one shot classes, workshops, online learning) at different events during a graduate student's coursework can address the needs of different learners.Adult learners evolve throughout the course of their studies, not only in the knowledge of their discipline but their approach to research and the way they think about themselves as learners. The needs of doctoral students are often as different from those of masters students as graduate students' are from undergraduates. Thus, it is recommended that, when researchers collect data on graduate students that masters and doctoral students be treated separately.This synthesis took a service provider (e.g. reference; information literacy instruction) approach to examining information behavior. Thus, it is important for information behavior researchers to become familiar with research on information literacy and vice versa, as one can inform the outcome of the other (Urquhart, 2010). While examining the cognitive process of information behavior lays the foundation for this type of research, it is important to consider the influence of cultural and environmental factors as well. Thus, there needs to be an expansion of the research on cultural differences in information behaviors.It can be difficult to draw conclusions from a synthesis of research that answers similar yet slightly different research questions. In order to gain a better consensus on information behaviors there is a need for researchers to use validated questionnaires. Urquhart (2010) recommends that new post-graduate researchers replicate studies using standard questionnaires as a more than acceptable approach to publication.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this review is to draw out patterns of information seeking behavior of graduate students as described in the empirical research published from 1997 to the present.\nDesign/methodology/approach: - A systematic search of databases for studies on information behavior and graduate students was employed in order to retrieve studies for a systematic review. Common themes that emerged from the literature were synthesized into a discussion of behavior patterns. Additionally a study quality analysis was conducted for all retrieved studies using a critical appraisal checklist for library and information research.\nFindings: - This review revealed that graduate students begin their research on the internet much like any other information seeker, consult their faculty advisors before other people, and use libraries in diverse ways depending on the discipline studied. Additionally differences were noted between international and home students, and doctoral and master's students.\nPractical implications: - The findings of this review indicate that information behavior research conducted on graduate students should delineate between masters' and doctoral students. Further, the findings may inform both academic librarian and faculty practice as to how to assist students with their research by helping them to understand how students typically approach research and how other institutions address common issues with special populations, such as non-native speakers and distance learners.\nOriginality/value: - No comprehensive review of information behavior studies, encompassing only the behaviors of graduate students has been conducted to date.\n",
"role": "assistant"
}
] |
Online consumers' reactions to price decreases: Amazon's Kindle 2 case
|
[
"Innovation diffusion theory",
"Online product review",
"Electronic word-of-mouth",
"Consumers’ reaction",
"Positive and negative emotion",
"Price decrease"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: As online shopping has become a regular part of our lives, online product reviews (OPRs), which form a collaborative social recommendation system for potential consumers, play a critical role during the decision-making process of online consumers (Cheung et al., 2008). Compared with traditional store shopping, consumers who plan to shop online rely on OPRs as a major information source for evaluating and purchasing products or services (Ghose and Ipeirotis, 2011; Hsu et al., 2013; Jalilvand and Samiei, 2012). Besides, OPRs can also be a communication channel among (potential) consumers, so that consumers share knowledge about the products they purchased online (Chen and Xie, 2008; Lee et al., 2006). Therefore, managing OPRs has become of utmost importance for online merchants as more and more online consumers depend on them (Bigne-Alcaniz et al., 2008). Prior research on OPRs has provided valuable insights by identifying electronic word-of-mouth (e-WOM) as a potential driver of consumers' behavior (e.g. Harrison-Walker, 2001), as well as by examining the diverse features of OPRs (i.e. star rating, review depth, positive or negative emotion, etc.) that influence review helpfulness (e.g. Mudambi and Schuff, 2010), which in turn, leads not only to helping consumers make smarter decisions, but also to influencing the sales of products or services sold online (Dellarocas et al., 2007).\n2. Theoretical background: 2.1. The comparison of online and offline consumers' reactions and behaviors\n3. Research methodology: 3.1. Research design\n4. Data analysis and results: 4.1. Study 1: the long-term reactions to price decreases on the four OPR features\n5. Discussion: The goal of this study is to observe and interpret how online consumers react to continual price changes of a product over its entire product life cycle expressed in the features of OPRs, and eventually to propose several future research topics on the role of price changes in consumers' reactions embedded in OPRs. Following the exploratory case study approach, we chose the OPRs of Amazon's Kindle 2 and designed three separate studies in order to investigate the reactions to price decreases embedded in the observable (star rating and review depth) and implicit (positive emotion and negative emotion) OPR features, as well as the number of OPRs per day. Then, based on theoretical perspectives of IDT and the literature on the price impact on consumers' reactions, we provide interpretation of the Kindle 2 case, followed by several future research topics.\n6. Conclusion: implications and limitations: 6.1. Theoretical implications\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate how consumers respond to price changes by analyzing online product reviews (OPRs) posted on a product (Amazon's Kindle 2), and to suggest several future research topics on online consumers' reactions embedded in OPRs.\nDesign/methodology/approach: An exploratory case study is conducted using OPRs added to the Kindle 2. By analyzing 6,714 OPRs, the authors examine how online consumers respond to two continual price decreases embedded in the observable (star rating and review depth) and implicit (positive and negative emotions) features of OPRs as well as how the number of OPRs per day has changed after two price drops.\nFindings: The authors found that all four features of OPRs (star rating, review depth, positive emotion, and negative emotion) and the number of OPRs per day had significantly changed after two price decreases for both long-term and short-term periods. In addition, online consumers' reactions to price decreases in terms of these four features and the change in the number of OPRs per day were different between the first and the second price drops.\nResearch limitations/implications: This study investigates online consumers' reactions to price decreases only. Future research should investigate other cases where price changes under the dynamic pricing strategy in order to find the relationship between price increases/decreases and consumers' reactions.\nPractical implications: This study implies that online merchants should consider consumer groups' innovation adoption stages and make strategic decisions for price decreases to improve the sales of their products.\nOriginality/value: While prior research involving the effects of price changes on consumers' reactions has focussed on offline consumers, this is among the first attempts to address the long- and short-term reactions to price changes in terms of both the observable and implicit features of OPRs, and suggests that consumers' reactions to price changes in OPRs are more complex.\n",
"role": "assistant"
}
] |
Reliability improvement of electronics standby display system of modern aircraft
|
[
"Aircraft",
"Failure modes and effects analysis",
"Navigation",
"Product reliability"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The electronic stand by display system (ESDS) is presently under development for the usage in Military Aircraft. The ESDS displays the important parameters like, calibrated airspeed and vertical speed, Mach number, angle of attack, rotor speed, turbine discharge temperature, variable nozzle area, fuel flow rate, etc. Further as ESDS being airborne equipment, its reliability should be very high. All measures have been taken during design and development of the unit by means of selecting right materials, processes, components, planning and various tests for conforming to the requirements, involving various agencies like quality assurance/quality control to increase confidence level of the equipment function. The equipment (ESDS) shares the load of main display units, i.e. multi function display (MFDs). In case of the failure of main display units, ESDS will provide vital flight information to the pilot to navigate the aircraft safely to the base. In normal mode, the flight performance information, which requires to be continuously monitored for the safety of the flying, can be also displayed by this unit and MFDs can be programmed to display other mission related parameters. The ESDS is stand alone color display unit of size 3 x 3 inches. ESDS generates and displays page information as predefined pages based on the parameters received through its various interfaces. Two units of ESDS are envisaged for one cockpit. One ESDS is to be located on left main instrumental panel and other on the right hand main instrumental panel.\nMethodology: The process used for carrying out FMEA is specified by MIL-STD-1629A procedure for carrying out failure mode, effects and criticality analysis.Analysis approach\nFailure mode and effects of ESDS (based on Ishikawa diagram): The FMEA was initiated at the lowest indenture level and was preceded through increasing indenture level (bottom-up approach), up to the main unit level until the FMEA for the entire system was complete. Figure 1 shows the various indenture levels of the system in an order. At the lowest indenture levels, the failure modes were based on the functions of the complete unit. Similarly at the higher indenture levels the failure modes were based on the functions of that module or section or subsection. In case if the highest indenture level fails, it will be actual failure mode of the component, with the exception to the integrated circuits are bought out finished modules for which design/constructional details were not available.Further for the FMEA only the functional modules and its components are considered. There are non-functional component like enclosures, gaskets, fixing screw, etc. are not considered for FMEA analysis. The failure due to workmanship defects, handling during transportation and maintenance was also not considered. Only the potential failures which were the outcome of the design and component selection were only used for the analysis.Levels of ESDS\nFailure effects of ESDS and other modules: As shown in the Figure 3 reliability diagram of ESDS, the failure of any of its module will cause the system failure, i.e. absence of display parameter. As for failure effects concern, it is a series system. In the diagram IO 1 and IO 2 are shown in the single block, but practically they are parallel. Only in case of both the card failure, the system will fail; else some of the parameter will not be displayed.\nFailure effects of power supply module: Power supply module provides the voltage required by the GIU, IO module, LCD module, Bezel and other external services like magnetometer, etc. Failure of this module will cause the failure of complete system. The reliability diagram of power supply module is shown in Figure 4.EMI Filter\nFMEA work sheet: The functions and failure were analyzed at the component level and the work sheet was prepared as per the military standard. Each module of the ESDS was divided into section and subsection (Individual levels) as done in previous sections and the component of the particular indenture were grouped, listed, analyzed and presented in the work sheet.The indication number (ID no.) of the component with the same part number, remains same with in the section/subsection. In nomenclature column of the worksheet, the circuit reference number was given along with the nomenclature of the component. This was done to facilitate, tracing the component in the master drawing index and to identify its function in the circuit. The function of the individual component in that circuit was reflected in the \"Function\" column of the work sheet. In few cases of micro circuits, where it was not possible to ascertain the failure mode, the mode of the failure was reflected as functional failure.The ESDS, being an avionics (aviation electronics) module have only few electromechanical components like connectors and flexi card. The remaining electronic components fail at random due to electrical over stress (EOS). From the analysis it was seen that supervisory circuits, which maintain the voltage to the threshold limit, is likely to deteriorate the other component when it fails. The deteriorated component may fail at later date, where the cause was EOS again due to failure of supervisory circuit. Therefore, the column 'probable cause of failure' in the work sheet was filled mostly with EOS.The failure of the system means there is no display by ESDS, which can be seen visually. But to identify the failure at subsystem level (module level/section level/subsection level), the power on self-test (POST) of continuous built in test (CBIT) was used. When the failure at the section/subsection level is identified, it becomes easy to identify the failed components. The column \"Failure detection method/warning\" is filled with POST and CBIT mostly. FMEA of power supply module is summarized in Tables I and II for DC-DC converter.\nSummary of critical failures: FMEA has been carried out on ESDS, which was used in modern civil and military aircraft, it may be seen from the worksheet, most of the failures categorized as critical \"severity class II\" are pertaining to power supply related components and functional IC's. A summary of the critical failure are tabulated and given in the Table III as per Mil-Std-1629A.\nConclusion: From the table it may be seen that there are 22 cases of capacitor failure and 15 cases of resistor failure which are categories as critical. But from the selection criteria of these components and the data sheets, the failure rate is below 0.001 per cent per 1,000 h. Similarly 13 cases of electromechanical components (connectors) are categorized under critical where they fail due to mechanical stress. The reliability of these components under mechanical stress is proved by subjecting the unit to vibration (sine and random) and shock as a part of qualification testing. The remaining failure of 24 cases of IC's, Transistors MOSFETs, optocoupler and regulators which are consider as critical, is a causes for concern for reliability of the units.Improving the reliability of a system by adding compensating mechanism or redundancy adds weights. But the weight is critical factor for any fighter aircraft. In case of ESDS, which is cockpit instrument, apart from weight there is space constraint also. At the same time one should not loose the reliability of a system, which is very important for any airborne equipment. Therefore, compromise between reliability of a system and weight at some point is necessary.In the design feature, it has been brought out that the unit is subjected to environmental stress screening at various level like, component level, card (PCB) level and unit level. Further at unit level thermo shock is also being planned. These measures are likely to bring out the inherent failure in the unit and weed out the infant mortality. Burn-in and endurance testing of the unit is also proposed to carryout on the unit. The reliability measure is included as design feature. If the development of the unit is considered to be satisfactory then no further compensating mechanism to cater for the critical failure is suggested at this stage.If necessary the selection of the components, which are likely to cause the critical failures may be reviewed. Further any failure or the discrepancies noted during the qualification testing and life/endurance testing should be considered in view of the outcome of FMEA for improving the reliability.\n",
"role": "user"
},
{
"content": "Purpose: - Through this analysis an attempt has been made to improve the reliability of an ESDS used in the cockpit.\nDesign/methodology/approach: - The process used for carrying out FMEA is specified by MIL-STD-1629A procedure for carrying out failure mode, effects and criticality analysis.\nFindings: - In an increasing demand in avionics sector, particularly in modern defence and civil aircraft, safety and reliability are the prime concern to complete the mission successfully. This made technocrats to rethink over the safety of complete system by adding redundancy to the critical activities or what else not? The electronic stand by display system (ESDS), being an avionics unit and fitted in the cockpit for displaying very important navigation and engine parameters to pilot whenever main display unit fail to perform it primary functions. So even standby system need to have high reliability to serve the purpose by providing important flight parameters for the safe landing of pilot and crew. Failure modes and effect analysis (FMEA) is one of the effective reliability assessment tool, which evaluate systematically and document the potential failure modes of a system or equipment and their causes. It helps in grading the severity of all potential failure modes and useful in carrying out the changes in early phase of design. The analysis starts with the potential failure of a smallest component at the final indenture and goes up to the initial indenture level.\nOriginality/value: - The paper adds insight into the reliability improvement of electronics standby display systems of modern aircraft.\n",
"role": "assistant"
}
] |
Estimation of building heat transfer coefficients from in-use data: Impacts of unmonitored energy flows
|
[
"Building thermal performance",
"Heat transfer coefficient",
"In-use measurement",
"Occupied homes",
"Unmonitored energy flows"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Discrepancies between as-designed and as-built thermal performance of dwellings are widely reported (Kelly et al., 2012; Stafford et al., 2012; Johnston et al., 2016), and in situ measurement is recognised as essential for identifying and addressing the \"performance gap\" (Gori et al., 2018). Space heating accounts for 63 per cent of UK domestic energy demand (BEIS, 2018), and is a key target for reducing greenhouse gas emissions in the 2011 UK Carbon Plan (HM Government, 2011); however, no compulsory whole-dwelling thermal performance tests have been adopted in the UK to date. The coheating test method described by Johnston et al. (2013) allows estimation of the in situ heat transfer coefficient (HTC) - which describes the net rate of heat losses (W) from a building per unit indoor-outdoor temperature difference (K) - with an uncertainty of +-10 per cent (Jack et al., 2017); however, the test is both intrusive and energy-intensive, requiring heating of a vacated dwelling to an elevated indoor temperature for one to three weeks (Jack et al., 2017). Development of alternative methods for inferring the HTC using data gathered from occupied, in-use dwellings is ongoing, with many efforts being focused through IEA EBC Annex 71 (Roels, 2017).\n2. Definition and measurement of the heat transfer coefficient: 2.1 Theoretical definition of the HTC\n3. Case study dwellings: Two UK case study dwellings are considered:\n4. Calculation of energy flows: The SAP model of building energy use is based on the BRE Domestic Energy Model (BREDEM), which applies a methodology consistent with BS EN ISO 13790 for calculating monthly whole-dwelling energy use (BRE, 2014). The SAP calculation of energy flows centres around the use of a monthly heat balance to calculate the load Eh on the dwelling's heating system. For each month:\n5. HTC measurement: impact of unmonitored energy flows: The impact of omitting or including traditionally unmonitored energy flows on in situ HTC measurements from in-use dwellings is now explored using a quasi-steady state heat balance approach. Here, it is assumed that the behaviour of the case study dwellings and their occupants are exactly as predicted by SAP modelling, and the predicted energy flows are treated as idealised monitored data streams which may or may not be included in a heat balance calculation. The results of the subsequent analysis therefore reflect the thermal behaviour of the as-modelled dwellings, rather than their real-world counterparts, and as such are only illustrative of what may be expected from real-world in situ HTC measurement.\n6. Discussion and conclusions: This paper has demonstrated the potential impact of unmonitored energy flows on the estimation of whole-building HTCs for two in-use dwellings at opposite ends of the energy performance scale. In the 1930s-built LBRO dwelling (HTC 507.3 W K-1), the comparatively high proportion of heating demand satisfied by metered gas and electricity means that the omission of unmetered heat sources (such as metabolic and solar gains, which contribute 11 per cent of heating) and sinks (such as hot water drainage and efficiency losses) have only minor impact on the HTC that may be estimated from in-use monitored data. In contrast, for the high-performing GBRO dwelling (HTC 67.7 W K-1), the increased proportion (35 per cent) of heating demand met by unmetered metabolic and solar gains informs a greater sensitivity of the heat balance estimation of the HTC to their omission. Furthermore, the greater proportion of internal and solar gains not contributing to maintaining the heat balance in the GBRO dwelling (\"internal losses\" and \"non-useful gains\" in Figure 2) contribute to a 23 per cent overestimation of the HTC even when solar and metabolic sources, along with hot water and efficiency sinks, are accounted for. It is therefore suggested that quasi-steady state heat balance methods for in-use HTC estimation, while perhaps robust to unmetered energy flows for older dwellings such as the LBRO case study, may be poorly suited for use in newer, higher energy performant dwellings.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to identify the impact of traditionally unmonitored energy sources and sinks on assessment of the as-built thermal performance of occupied homes. The analysis aims to demonstrate the potential scale of uncertainties introduced in a heat balance estimation of the heat transfer coefficient (HTC) when using in-use monitored data.\nDesign/methodology/approach: Energy flows for two UK homes - one a 1930s dwelling with high heat loss, the second a higher-performing 2014-built home - are predicted using the UK Government's standard assessment procedure (SAP) and visualised using Sankey diagrams. Selected modelled energy flows are used as inputs in a quasi-steady state heat balance to calculate in-use HTCs as if from measured data sets gathered in occupied homes. The estimated in-use HTCs are compared against SAP-calculated values to illustrate the impact of including or omitting various heat sources and sinks.\nFindings: The results demonstrate that for dwellings with low heat loss, the increased proportion of heating demand met by unmetered internal and solar gains informs a greater sensitivity of a heat balance estimation of the HTC to their omission. While simple quasi-steady state heat balance methods may be appropriate for dwellings with very high heat loss, alternative approaches are likely to be required for those with lower heat loss.\nOriginality/value: A need to understand the impacts of unmetered heat flows on the accuracy with which a building's thermal performance may be inferred from in-use monitored data is identified: this paper illustrates the scale of these impacts for two homes at opposite ends of the energy performance scale.\n",
"role": "assistant"
}
] |
Capability sequencing: strategies by township and village enterprises in China
|
[
"Management strategy",
"Small enterprises",
"Questionnaire",
"China"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In the last ten years, the success of township and village enterprises (TVEs) in China has contributed substantially to the country's fast economic growth. The total industrial output of TVEs has surpassed that of state-owned firms to become the dominant output contributor in China (Luo et al., 2000). TVEs are rural non-state enterprises that are owned and operated collectively or privately. TVEs in China possess few traditional competitive advantages. First, they are relatively new: the Chinese government has only permitted private enterprises since the 1980s. Start-up businesses usually suffer severe difficulties getting access to capital, technologies, information, scarce materials, and distribution channels (Porter, 1985). Second, TVEs are located away from cities, which means that it is difficult to attract skilled management teams and labor. Third, TVEs were born at a time when the government began to encourage western investment in the country. In other words, even in their cradle TVEs have had to compete against well-established multinational corporations. How can Chinese TVEs manage to survive and prosper with these fatal disadvantages?This paper aims to answer that question by shedding light on the strategies used by Chinese TVEs. We first review the extant literature on competitive advantages. Next, based on the qualitative data and the literature review we explain the components of strategies used by Chinese TVEs. We then use quantitative data to test our hypotheses. The paper concludes with an explanation of the contributions of our study to the literature.\nLiterature review: Why are firms different? A related and even more important question is, how can firms develop and sustain their competitive advantages? These are the fundamental questions in strategic management (Rumelt et al., 1994). During the 1980s, the dominant paradigm in the field was the competitive advantage approach developed by Porter (1985). This approach suggests that a firm should adopt one of three generic strategies to achieve above-average performance: cost leadership, differentiation, and focus. Having a low-cost position yields the firm above-average returns in its industry despite the presence of strong competitive forces. Its cost position gives the firm a defense against rivalry from competitors, because its lower costs mean that it can still earn returns after its competitors have competed away their profits through rivalry (Porter, 1980). Differentiation, on the other hand, provides defense because of brand loyalty by customers and resulting lower sensitivity to price. Focus rests on the premise that concentration can increase efficiency and effectiveness, which in turn lead to cost leadership or differentiation in a particular market segment. A firm that fails to adopt any of these strategies will be \"stuck in the middle\" and possess no competitive advantages.Another distinct class of approaches emphasizes the building of competitive advantages through organizational learning (Grant, 1996). Firm is a repository of social knowledge, where a competitive set of capabilities is replicated over time while subject to imitation (Zander and Kogut, 1995). The knowledge, in general, can be categorized into \"information\" and \"know-how\". Learning occurs when firms seek to increase their stock of knowledge or when such knowledge is transferred between individuals or firms through intentional teaching or imitation (Zander and Kogut, 1995). Competitive pressures create a value in developing a capacity of replicating knowledge between firms (Cockburn et al., 2000). Sometimes, the competitive pressures may also lead to the intentional knowledge transfer between partners in order to maximize their common interests against their common competitors (Helfat, 1997). Through collective and individual learning, an organization develops and exploits a core competence that differentiates it from other firms (Prahalad and Hamel, 1990). Learning is a dynamic process through which firms accumulate knowledge (Teece and Piscano, 1994). Innovations become more likely when a firm has accumulated reasonable product knowledge or R&D assets (Nelson and Winter, 1982). In today's business, innovations include not only new technologies and new products but also new positioning concepts and new distribution methods (Cockburn et al., 2000).The third strand of literature suggests that resources form the basis of differential firm performance (Barney, 1991; Wernerfelt, 1984). This perspective suggests that firms can develop and sustain their advantages only when they have resources that are valuable, rare, inimitable, and non-substitutable. A recent component of this approach emphasizes the dynamics of the combinations of a firm's competencies (Eisenhardt and Martin, 2000; Teece et al., 1997). It argues that the important source of sustained competitive advantages rests on managers' dynamic capabilities to integrate, build, and reconfigure internal and external competencies to address changing environments. These capabilities are complicated routines that emerge from path-dependent learning processes (Zott, 2003).\nDevelopment of framework: Our initial conceptual framework was based on field observation and a thorough review of the related literature. We conducted detailed personal interviews at eight Chinese TVEs in the Jinjian and Putian areas of Fujian Province. A basic model of TVE strategies emerged from the interviews. Further refinements were based on two additional steps. First, we showed the model to interviewees at seven other Chinese TVEs in the Huian area and asked them whether the model matched their experiences. Second, personal interviews were conducted at three more TVEs in Putian.Figure 1 shows the conceptual framework that resulted from the refinement process. The TVEs under observation took three steps to develop and sustain their capabilities, which are defined in this study as the organizational abilities to deploy resources and to develop new resources (Henderson and Cockburn, 1994). The sequencing of the three steps implies not only a reconfiguration that involves the development, retention, and addition of capabilities, but also the causal linkages across time between those capabilities. In other words, the capabilities developed at earlier stages serve as bases to create new capabilities at later stages. More importantly, different types of capabilities, developed at sequenced stages, co-exist in the end and contribute to a TVE's pool of competitive advantages. The causal linkages across time between those capabilities and the co-existence of those capabilities result in their inimitability.\nThe conceptual framework: Founding stage (step 1): capabilities to minimize costs\nResearch method: Data collection method\nResults: The data were entered, verified, and analyzed, and the anonymity of the respondents was preserved. Items were grouped according to the categories of strategic intents. For each of these categories, the mean value of all the items was calculated and used as the scale score of the specific strategic intent at each stage of business development. Correlation among the utilization of competitive advantages was estimated for each stage of business development, and the results are presented in Table I. Repeated measure analysis of variance was used to study the strategic intents at each of the stages. The results are summarized in Table II and presented graphically in Figure 2. All of the analyses were conducted using the statistical analysis system.There was statistical evidence to indicate that the strategic intents at the different stages were different (p-value=0.0013), and that the type of capability used depended on the stage of business development (p-value<0.0001). In the founding stage, TVEs aimed to build and utilize capabilities to minimize costs with an average score of 3.76, which is statistically significantly higher than that of systematic learning (2.15) and low-priced innovator positioning (3.1). Thus, H1 is supported. In the growth stage, the TVEs established partnerships with western multinational corporations and aimed to learn systematically. This is evidenced by the highest mean score in the realization of this learning (3.77) relative to the mean score of cost minimization (2.87) and positioning (3.32). Hence, H2 is supported. When the companies matured, they aimed to capture wealth with low-priced innovator positioning. The third hypothesis is supported by the significantly higher mean score of the utilization of low-priced innovator positioning at this stage (4.33) relative to that of cost minimization and systematic learning (2.51 and 2.42 respectively). Figure 2 displays these results graphically.\nDiscussion and conclusion: The capability sequencing strategies used by TVEs in China provide a dynamic framework that enables firms to systematically build and combine different capabilities into a complex bundle of competitive advantages. This study offers major contributions to the literature. First, it identifies the causal linkages across time between firms' different capabilities. Labor-intensive industries and rural locations offer TVEs opportunities to create capabilities to minimize costs. Such capabilities then make possible partnerships or systematic learning. The cost minimization and systematic learning capabilities, in turn, lead to low-priced innovator positioning. Second, the co-evolution and co-existence of different capabilities contribute to capability inimitability. The causal linkages across time between different capabilities make it difficult for competitors to duplicate the capabilities, as they have path-dependent histories.The framework has several features that differ from those of existing models. Porter (1985) suggests that a firm should choose either cost leadership or differentiation, but not both, which will cause the firm to be \"stuck in the middle\". The findings from this study indicate that firms in developing countries can achieve cost leadership and differentiation, but the route to the destination has a path-dependent history. Firms in developing countries can start with cost leadership strategies in labor-intensive industries and then develop differentiation through systematic learning and improvement. Systemic learning through partnerships may not incur high costs in developing countries, but it also may not create high short-term returns.The findings of this study indicate that a teacher-student or master-apprentice model of organizational learning is possible between firms, and it is an effective way for weak firms to swiftly duplicate knowledge. This model does not generate core competencies, but provides systematic leaning capabilities.Dynamic capability models suggest that firms should integrate, build, and reconfigure internal and external competencies to address rapidly changing environments (Eisenhardt and Martin, 2000; Teece et al., 1997). The findings of this study indicate that dynamic capabilities are important not only for firms in rapidly changing environments, but also for those in relatively stable industries, such as labor-intensive industries. Firms should develop different capabilities over time and combine these capabilities into complex bundles. Without unique capabilities in a particular period, a firm cannot survive competition. The development of different capabilities has a path-dependent history. The capabilities that are developed at earlier stages serve as the platform to develop different capabilities at later stages. More importantly, the capabilities that are developed at earlier stages do not disappear when firms develop new capabilities. The capabilities that are developed at different stages co-exist and comprise the firm's bundle of competitive advantages.It should be pointed out that our study, like any other, suffers from limitations. First, the collection of historic data can be biased because the respondents may not know the history of their firms very well. Some respondents may have joined their firms after the founding stage or growth stage. Second, the data were based on the TVEs in Fujian province only. Fijian province has been one of those coastal areas where most successful TVEs are located. As such, the findings may not apply to TVEs in the Northwest areas of China where TVEs suffer location-specific disadvantages, such as underdeveloped infrastructures, including transportation, banking, and communication systems, as well as much less foreign investments. Therefore, our future studies on Chinese TVEs should expand the empirical database and include those TVEs in underdeveloped areas. The competitive advantages and disadvantages of those Chinese TVEs in underdeveloped areas can enrich our understanding on how firms survive with severe resource limitations.\n",
"role": "user"
},
{
"content": "Purpose: - This study aims to identify and assess the strategies of township and village enterprises in China to capture competitive advantages.\nDesign/methodology/approach: - The paper employs a self-administered questionnaire survey approach, involving a sample of managing directors of township and village enterprises in the Fujian Province, China.\nFindings: - The analysis identifies the causal linkages across time between firms' different capabilities. Labor-intensive industries and rural locations offer township and village enterprises (TVEs) opportunities to create capabilities to minimize costs. The cost minimization and systematic learning capabilities, in turn, lead to low-priced innovator positioning. It also suggests that the co-evolution and co-existence of different capabilities contribute to capability inimitability.\nResearch limitations/implications: - Future studies on Chinese TVEs should expand the empirical database and include TVEs in underdeveloped areas and to investigate how firms survive within severe resource limitations.\nPractical implications: - The findings of this study indicate that dynamic capabilities are important not only for firms in rapidly changing environments, but also for those in relatively stable industries, such as labor-intensive industries. Firms should develop different capabilities over time and combine these into complex capabilities bundles.\nOriginality/value: - The findings from this study indicate that firms in developing countries can achieve cost leadership and differentiation, but the route to the destination has a path-dependent history.\n",
"role": "assistant"
}
] |
Implications of dividend tax reforms on M-REITs performance
|
[
"M-REITs",
"Dividend tax reforms",
"Jensen α",
"REITs performance",
"Sharpe index",
"Treynor index"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The development of real estate investment trusts (REITs) started in the USA in 1960 under the REIT Act of 1960. Following the success of the US-REITs, other countries started to establish their own REITs markets such as in the Netherlands (late 1960s), Australia (late 1970s), Japan (2001), South Korea (2001), Singapore (2002), France (2003), Hong Kong (2003), Taiwan (2003) and the UK (2007) (Brueggeman and Fisher, 2011; Newell and Wen Peng, 2012).\n2. Literature review: Performance is the return or the escalation in wealth over time of an investment relative to the amount of risk faced by investors (Christopherson et al., 2009) with the main goal of utility maximization (Levy and Sarnat, 1984). Markowitz (1952, 1959) modern portfolio theory provides a basic model in obtaining the expected rate of return and risk measure of a portfolio of assets under a few assumptions of investor's behavior (Reilly and Brown, 2012): investor looks at expected returns of investment as probability of distribution over a period; he/she faces diminishing marginal utility curve; his/her risk estimation is based on variability of expected returns; investment decision is based on risks and returns only; investor prefers the highest rate of return for a specified level of risk or the lowest risk for a specified level of return.\n3. Research design: The sample comprises all 19 M-REITs that are listed at BM. Monthly return of the M-REITs, FTSE BM KLPI (KLPI), FTSE BM KLCI (KLCI) and three month Malaysia T-Bills were taken from datastream from January 1999 to December 2014. In total, 16 years period are tested because longer sample period can portray a better picture of REITs performance as they are considered to be in a volatile industry going through the peak and sluggish period as stressed by Han and Liang (1995). KLCI was used as the market benchmark. A self-constructed tax-adjusted value-weighted M-REITs index was constructed from the summation of each M-REITs total return index taking into account the tax reforms in the 2007, 2009 and 2012 budgets in which the Malaysian government improved the tax incentives. Dividend tax rate for individuals and domestic investors have been reduced from 15 (2007) to 10 percent (2009, 2012) and for foreign institutional investors, from 20 (2007) to 10 percent (2009, 2012) (KPMG, 2006, 2011) following the implementation of tax transparent status where REITs were exempted from paying tax if they disbursed 90 percent of their income as dividends to unitholders. Furthermore, to ensure an accurate performance of M-REITs, a tax-adjusted return of individual REITs are calculated as follows:(1)\n4. Analysis of results: Table I exhibits the RAP of the M-REITs in comparison to the tax-adjusted value-weighted REITs index, KLCI, KLPI and T-Bills for the period from January 1999 to December 2014. The average monthly returns for 16 out of 19 REITs were higher than the KLCI which is consistent to the study by Newell and Peng (2009) who found that the A-REITs outperformed the other major asset classes from 1996 to 2007. The average monthly return of the KLCI amounted to 0.7100 percent. Among the 19 M-REITs, First-Malaysia-Property-Trust (FMPT) provides the highest average monthly return of 2.8440 percent. This concurs to the result reported by Kok and Khoo (1995). All M-REITs except for IGB (0.2444 percent) outperformed the KLPI and T-Bills average return of 0.5167 and 0.2405 percent, respectively. The highest total risk is exhibited by FMPT, with a monthly standard deviation of 19.0055 percent. Six M-REITs exceeded the KLCI standard deviation that shows 5.1154 percent. The M-REITs' total risk ranged from 2.5498 to 19.0055 percent. In all, 12 out of 19 M-REITs have a higher total risk as compared to the tax-adjusted value-weighted M-REITs Index; whereas 18 M-REITs reveal a lower volatility against the KLPI except for FMPT. All M-REITs are found to be more volatile than the T-Bills.\n5. Conclusion: Performance of M-REITs is re-visited in this study to take into account of the effect of the tax reforms in 2007, 2009 and 2012. The study improves upon the existing literature on REITs by utilizing tax-adjusted returns of individual REITs and tax-adjusted value-weighted REITs index. Performance of 19 M-REITs was measured by using the Sharpe (1966), Treynor (1965) and Jensen (1968) risk-adjusted measures from January 1999 to December 2014. Most of the M-REITs outperformed the KLCI, KLPI, tax-adjusted value-weighted REITs index and the Malaysia three month T-Bills during the period of study. FMPT generated the highest mean return and total risk. Based on the Treynor and Jensen performance measures, it was ranked first. All the RAP measures showed that the average return of M-REITs and the tax-adjusted value-weighted REITs index outperformed the KLCI and KLPI, implying that during this period, investing in REITs is better than investing in the equity market, property sector and the government three month T-Bills. The T-Bills underperformed the other financial assets with a monthly mean return of 0.2405 percent but had the lowest total risk of 0.0396 percent.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the performance of 19 Malaysian Real Estate Investment Trusts (M-REITs) over the period 1999 to 2014, following the implementation of dividend tax reforms announced in the 2007, 2009 and 2012 budgets.\nDesign/methodology/approach: Sharpe index, Treynor index and Jensen a are utilized to compare the performance of M-REITs against a newly developed tax-adjusted value-weighted M-REITs index, equity market, property sector and three month Malaysia Treasury Bills (T-Bills). The calculation of M-REITs returns has been adjusted to take into account the dividend tax reforms which have never been considered in previous studies.\nFindings: Most M-REITs outperform the tax-adjusted value-weighted REITs index, equity market, property sector and three month T-Bills. Property sector performs worst during those periods. Some of the M-REITs have a higher standard deviation than the equity market and the tax-adjusted value-weighted M-REITs index. Most M-REITs have a lower total risk than the property sector. Further analysis shows that before (after) the tax reforms, most M-REITs underperform (outperform) the other sectors. The introduction of the tax reforms benefits both REITs and investors. A significant positive Jensen a for some M-REITs indicates that fund managers are able to time the market or to select undervalued assets.\nPractical implications: Findings of the study would enable investors to evaluate the performance of all REITs in comparison to other financial assets during the period of study for better investment decision making. A more accurate assessment on REITs performance that take into account the tax reforms, is available for investors and fund managers to decide on the investment mix to be included in their portfolio. Moreover, fund managers' performance can be assessed whether they perform better or worse than the equity market, property sector and three month T-Bills.\nOriginality/value: This study contributes to the scant literature on dividend tax reforms and their implication toward REITs performance. It is the first study to thoroughly assess the returns of REITs by taking into account the changes on dividend tax rates announced in the 2007, 2009 and 2012 budgets.\n",
"role": "assistant"
}
] |
Shock Therapy and Entrepreneurial Flare #Brexit
|
[
"Entrepreneurs",
"Economic growth",
"Financing",
"International entrepreneurship"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Many claim that Brexit benefits will lead to a reinvigorated UK economy, spurred on by competing in newly opened global markets through signing new and innovative trade deals (Minford et al., 2017). Underpinning competitive performance in the new trade arrangements will be a supply side policy driven by especially deregulation[1] (Halligan and Lyons, 2018) and tax cuts (Singham et al., 2017; Johnson, 2018a, b). This message is reinforced by arguments that \"protectionism [can no longer be] perceived as just a problem of border barriers. It now lurk[s] in national regulations, performance requirements, buy-local provisions, investment benchmarks, regulatory standards, intellectual property laws, and other domestic laws, regulations, and rules\" (Ikenson et al., 2018, p. 6). In this argument, trade deals are therefore as much about deregulation as they are about reducing tariffs. In this paradigm, trade deals, deregulation and tax cuts are seen together as being synonymous with unleashing a new wave of entrepreneurial creativity. In this viewpoint piece and working from the assumption that the UK can sign trade deals that are at least as good as the EU, we ask how robust are the supply side de-regulation and tax cutting legs as drivers of innovation, growth and entrepreneurship?\nEconomic Shock Therapy: Friedman (1962, p. xiv), the Intellectual Architect of \"Economic Shock Therapy\", wrote:Only a crisis - actual or perceived - produces real change. When that crisis occurs, the actions that are taken depend on the ideas that are lying around. That, I believe, is our basic function: to develop alternatives to existing policies, to keep them alive and available until the politically impossible becomes the politically inevitable.\nBrexit: a crisis to force productive change?: Shock Therapy suggests a crisis - a rupture with the past forces productive change. In a UK Brexit context, the suggestion has been made that European Union membership has distorted growth and innovation through burdensome regulation and restricted access to non-European Union markets (Ormerod, 2017). In this nexus leaving the EU is the shock the UK needs to enable the market to adjust presumed distorted UK business structures that have evolved as being part of the European Union into structures that can more effectively compete and beat competitors globally:In fact, Brexit, as described above, constitutes a major economic reform, similar to the many market reforms introduced by UK governments since the early 1980s and recently. Such a reform would benefit industries where we perform best and would remove subsidies from some others: jobs and investment (including foreign direct investment) will expand in the first at the expense of the latter. Overall, as our pre- and post-Brexit forecasts show, jobs and investment will expand as UK output rises, helped by the boost to spending power from lower prices(Minford et al., 2017, p. 6).\nEconomic Shock Therapy in action: History can often shed helpful light and wisdom on specific policy alternatives. In the recent past a number of states left a protective community and then went through a series of market shocks to adapt them to the global economy and the results varied according to country. The fall of the USSR and Comecon in the period 1989-1991, plus the liberalisation procedures adopted under the \"Washington Consensus\" provide a real-life laboratory to see how Shock Therapy works in practice, when applied to economies that have been subject to restrictive regulation and protected markets. Could this be a reasonable simile through which to understand the assertions by Minford et al. (2017)? We take two examples from the breakdown of the Soviet Union, following the fall of the Berlin Wall in 1989 - Poland and Russia.\nA necessary and sufficient condition?: As children, we may have been taught the fallacy of the following statement \"a bus has four wheels, therefore everything with four wheels is a bus\". The relative GDP performance from our brief review suggests that Shock Therapy in itself is not a \"necessary and sufficient\" condition for economic growth (Kierzenkowski et al., 2016). Rather it is \"context\" because it worked in one context does not mean it will work in another.\nWhen is a shock positive? - innovation and Shock Therapy: How then is this cycle of innovation enhanced by Economic Shocks? When is a shock positive and when is it negative?\nFirst-order entrepreneurship issues: If a Brexit shock is to improve entrepreneurial success chances arguably there needs to be an identifiable positive impact on the key enablers for entrepreneurial growth. We cite these as: access to finance; skills; appropriate legal and regulatory frameworks; and community and clustering. A study of youth entrepreneurship in Serbia (Bobic, 2017) found the first three of these to be key and Thompson (2010) gave a vivid account of the importance of community to entrepreneurial activity. Issues underlying these enablers implicitly raise the question how Brexit Shock Therapy or cutting taxes will raise entrepreneurial growth if these measures fail to address core market failures.\nConclusions: In this viewpoint piece, we have argued that Shock Therapy does not work every time and does not automatically raise entrepreneurship and innovation levels. Shock Therapy success depends upon context and upon having financial markets or inward investment streams that will allocate funding flows to innovative entrepreneurs. We suspect that in most cases substantial Economic Shocks are unlikely to help raise lending flows to entrepreneurial businesses.\n",
"role": "user"
},
{
"content": "Purpose: Much has been written about trade deal opportunities after Brexit (e.g. Minford et al., 2017; Singham and Tylecote, 2018) but much less about envisaged \"supply side mechanisms\" that would translate a Brexit shock into improved UK competitive performance. Indications as to the supply side mechanisms involved can be found in some pro-Brexit writings and speeches and revolve around cutting regulation and reducing taxation, to spur innovation entrepreneurship. The authors contend that these measures align to a broad set of policy measures associated with Economic Shock Therapy, the Laffer Curve and the associated \"Washington Consensus\" (Williamson, 2005). The authors are looking to stimulate a conversation around whether these measures are most likely to stimulate entrepreneurial innovation and growth. The authors open by contrasting these concepts to growth equilibrium dynamics drawn from Wicksell, Keynes and Schumpeter - and by implication dynamic Walrasian General Equilibrium - to pose the question, is entrepreneur-led growth best led via slashing regulations and taxes or by focussing on correcting existing market failures? The purpose of this paper is to promote controversy and debate as to which \"supply side measures\" are most effective in enabling entrepreneurial growth.\nDesign/methodology/approach: The authors briefly review the pro-Brexiteer economic framework and relate this to broader Economic Shock Therapy and Laffer Curve concepts; how these have been applied and how some argue they can become \"supply side\" enablers in a positive Brexit innovation and entrepreneurship transformation. By drawing upon fundamental economic relationships such as Wicksell's (1898) \"Natural Rate of Interest\", the authors highlight the importance of information asymmetry and regulatory distortion in financial markets, resulting in some entrepreneurs (and associated innovations) failing to receive the capital their project merit. The authors pose the question, whether Shock Therapy, Laffer Curve type tax cuts and any Brexit \"bonfire of regulation\" will raise entrepreneurial growth and success.\nFindings: Both Shock Therapy and Laffer Curve inspired tax cuts have a patchy record of success, despite notable achievements in post-1991 Poland. The authors stress entrepreneurs drive innovation and growth, and a key support to them requires correcting \"access to finance\" market failures. It is questionable if Economic Shocks contribute anything to resolving this fundamental problem.\nOriginality/value: The authors open the supply side debate on anticipated \"Brexit Transformation\" in the context of long standing (some maybe long forgotten) theoretical understandings, thereby posing the question as to whether potential Brexit-related deregulation, tax cuts and \"Economic Shock\" therapy are likely to raise entrepreneurial competitive advantage and success rates. Market failure in financial market support for small firm growth and innovation needs are highlighted. Arguably, economic growth and innovation would be better sustained by addressing these failures, than introducing the \"unknowns\" and risks associated with a substantial Economic Shock.\n",
"role": "assistant"
}
] |
Self-employed and online: (re)negotiating work-learning practices
|
[
"Self‐employed workers",
"Workplace learning",
"Virtual organizations",
"Communities",
"Online operations"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: While many studies acknowledge the influence of web-based technologies, few probe how they redefine everyday learning activities and require working adults to re-negotiate the social, material, and political aspects of learning spaces online. There is plenty of buzz about online communities whether they are part of new social media configurations, such as Facebook, LinkedIn, or Twitter, or commercialized online product-related spaces carefully cultivated by enterprises. Online communities appear to offer socializing, networking, support, business opportunities, or social activism. But what practices in and around these spaces come to be labelled as learning?This research project focuses on online communities as sites of learning; more specifically, the informal work-related learning practices of self-employed workers in these cyberspaces. Self-employed workers are perhaps more likely to turn to online communities given that they work outside the sphere of the conventional workplace and its organized learning resources. Moreover, the self-employed are a significant part of the labour force. Industry Canada (2009) states that, in 2008, they represented 15 percent of all employed workers in Canada: 2.6 million people. Yet, Edwards and Nicoll (2004) argue that accounts of the workplace often do not reflect the workplace of the self-employed.Regarding online communities as rich learning spaces seems to make sense. However, educators are still wrestling with how to best facilitate the collective learning of an online community (i.e. Riverin and Stacey, 2008); a challenge magnified in the informal communities outside the purview of formal education (Gray, 2004; Ross, 2007). Some online communities are structured, as in an online course, complete with tightly bounded membership and purposeful design and pedagogical strategies. However, beyond these formal learning spaces, McLoughlin and Lee (2007) write that there are compelling examples of self-managed, thriving, knowledge-creating communities. Community can therefore also describe a gathering of people online that is organic, driven by a shared interest, and highly social (i.e. Boyd, 2006). These collectives form because someone is interested in a topic and searches for like-minded others on the Internet. These kinds of spaces, may also be purposefully nurtured, by professional associations, workplaces, or businesses. This research project focuses on spaces such as these - spaces outside the auspices of formal online courses.It is timely to question how informal pedagogical moments are being renegotiated by the technology woven into our lives. A critical look at the tensions in these cyberspaces is important when they become, perhaps even unintentionally, important sites of learning for some workers. This research project creates an opportunity to bring work-related learning practices of the self-employed worker to the fore. This inquiry may help facilitate a more careful look at the nuances of online communities and offer insights for people engaging in these spaces.\n2. Positioning in the literature: This study asks: what are self-employed workers learning in online communities? Although it is not surprising that workers click their way into all sorts of cyberspaces, there is not yet much research on these nascent online work-learning practices. However, there a few studies, which inform this exploration of informal work learning in online communities, and these are presented. I begin by exploring the practice-based perspective, which guides the conceptualization of work learning in this paper.Situated learning: a practice-based theory\n3. Inquiry methods: Situated learning, and Wenger's (1998) CoP theory, is often used to frame studies of online community or informal work-related learning. Although researchers frequently draw on LPP indirectly, a few studies use this concept more purposefully to help conceptualize the research project. For example, Jubas and Butterworth (2008) evoke LPP to study women's alternative informal learning pathways in the IT sector; Guldberg (2008) examines how adult learners in a professional development course move from peripheral membership to build distinct identity in an online discussion; Allan and Lewis (2006) explore how membership in a virtual learning classroom can support lifelong learning and impact careers; and Peters (2010) researches where and how learning takes place within a blended online-offline program for small business owners. Consistent with these studies, I draw on qualitative methodology.Participants were own-account self-employed workers (contractors and consultants who do not have staff) and selected by first employing purposive sampling through my network of contacts to \"select unique cases that are especially informative\" and then snowball sampling as the initial participants suggested others that might be willing to talk with me (Neuman, 2000, p. 198). As I found people who were interested in participating, I set up 20-minute telephone calls which enabled me to find out more about them and their work to ensure fit. Semi-structured interviews, which varied in length from one to two hours, were conducted with 11 self-employed workers (all based in Canada except one in the UK); ten by telephone and one face-to-face (F2F). Follow up dialogue, either by e-mail and/or short conversations, provided additional data. Interviews were recorded, transcribed, and sent back to participants for verification. Pseudonyms were assigned to protect the anonymity of participants.Thematic analysis was a process of understanding my data, systematically identifying key concepts and links between them, and developing more abstract theorizing of the data (Coffey and Atkinson, 1996; Merriam, 2001; Wolcott, 1994). Category construction began with the first set of notes. Notations were made by potentially relevant bits of data in the transcripts as I jotted down questions, tentative interpretations, and possible connections. After working through the notes, \"like\" comments and notes were grouped together and a running list of these groups created. The next transcript and related data were treated similarly, and the list of notes and groupings then compared to the first set. Applying this process for each transcript led to a set of categories derived from the data. These four categories were named and then made \"more robust by searching through the data for more and better units of relevant information\" (Merriam, 2001, p. 185).Participants ranged in age from 35 to 51. They had been self-employed for six months to 21 years and worked in a variety of fields. They engaged in a range of online communities, turning to these spaces at various stages in their career, and reported an array of positive and negative experiences. Turning to online communities was one informal learning strategy and it seemed to fit with their approach to work-learning and view of themselves as a \"self-managing\" learner. While six participants had been involved for as long as five to ten years, others had engaged for two to four years. One participant had tried several online communities in spurts over seven years and recently decided it was not for her. Technologies used include listservs, discussion boards and forums, Yahoo groups, e-mail, blogs, and RSS feeds. Eight participants mentioned venturing into Facebook, LinkedIn, and/or Ning - popular social networking sites - although these online spaces were perceived as places for socializing and networking.\n4. Findings: Participants described tremendously fluid ways of engaging: dipping into communities to explore and experiment, moving in and out of online spaces in waves, and feeling a sense of freedom to participate as intensively as they chose. The data in this study suggest that within these fluid ways of engaging, these self-employed workers were learning:* work practices;* the viability of doing particular work;* how to participate in online communities for work-learning; and* how to draw on and participate in fluid knowledges.Work practices\n5. Discussion: practices of learning online: Social media extend connectivity in ways that both fulfil and transgress our expectations. In this discussion, LPP is used as an analytic tool to examine the work-learning practices described in this study. What is considered legitimate- peripheral-participation is still evolving within the context of relational networks that increasingly rely on web-technologies to facilitate, challenge, and change ways of knowing.The learning\n6. Conclusion: This research brings work-related learning practices of the self-employed worker to the forefront. Framed by practice-based theory, this study examined what these workers are learning in online communities, networks of enough of the right people, ideas, objects, sensibilities, and caring to be useful. What these workers learned was intertwined with how their learning is enacted. This study juxtaposed these learning practices with the turbulence created by the endless possibilities of cyberspace. Despite the way \"digital technologies have come to dominate the domestic and commercial spaces of our everyday lives\", Beer (2005, p. 7) argues that they have become so embedded in our everyday lives that they are often camouflaged and urges a more reflexive approach toward digitalization. Web-technologies and shifting configurations of online collectives shake up notions of expertise, beliefs about who is able to produce and consume information, and where one locates themselves, in order to build work-learning practices. When a \"community\" is not bound by space, co-location, or time, ambiguities emerge. Who names what is useful learning when there is not one CoP, but multiple communities and practices? This study made a few inroads into exploring such complexities of learning in online collectives. Better understanding multiple positionings across several online communities (which are not bound to cyberspace) and ways of participating that are peripheral, partial and part-time (but nevertheless still meaningful and productive) help to re-orient discussions of learning in communities.Although Situated Learning, and Wenger's (1998) subsequent work on CoP theory, frames many studies of online communities as well as informal work-learning, it is often used to operationalize the concept of \"communities\" of practice. Yet, LPP creates opportunities to explore how different possibilities for learning are shaped by locations and trajectories within the practice and larger community of practitioners. One contribution of this study is the purposeful use of LPP, including a critique of its utility in fluid networked technologically mediated cyberspaces. Other researchers might also continue to (re)shape this concept.One strength of this research study could also be a limitation. These participants did a wide variety of self-employed work, engaged in an immense array of online spaces, and had very different experiences in online communities. This diversity provided rich data but also made it challenging to identify patterns as I was forced to constantly attend to the many nuances. Fortunately, this prevented me from jumping to conclusions and neatly summing up experiences. Perhaps more focused studies on a particular group of self-employed workers in a more tightly defined type of online community (i.e. those set up by professional associations) might enable more contextualized findings. Another potential limitation is reliance on interview data. These are narratives that people thought to tell about their everyday nature of informal learning, which is often taken for granted. Despite these challenges, the interview data was rich. Interviews enabled me to talk with people about their experiences in an array of spaces and to see how what they were learning unfolded through a complex blend of media, online spaces, and online and offline encounters. When I think about designing future studies to explore online cyberspaces I am not convinced that \"camping out\" in online community spaces would provide better data. Baym (2007) concludes that while it has been the norm to go to online communities in order to study them, it is no longer clear this is an appropriate strategy given the increasingly distributed nature of community.There is growing interest in nurturing online communities for ongoing professional development and lifelong learning initiatives. This study highlights the practice of work learning online, a practice tightly meshed with ways of working and not yet well defined or understood. Although I studied a particular group of people, these findings may be more broadly applicable. Merriam (2001, p. 222) advises that by providing detail, practitioners can determine themselves \"whether the research setting sufficiently resembles their own situation to warrant adopting the same practices\". As one anonymous reviewer pointed out, beyond this study's usefulness to the self-employed, it has wider implications for understanding the use of online communities more broadly (both formal and informal) as they infiltrate almost all areas of practice. Better understanding the work-learning practices of these workers may enable adult educators to question what comes to be construed and labelled as learning and how best to support these organic practices.\n",
"role": "user"
},
{
"content": "Purpose: - In order to explore how informal pedagogical moments are being renegotiated by the technology woven into people's lives, this paper aims to focus on online communities as sites of learning; more specifically, the informal work-related learning practices of self-employed workers in these cyberspaces.\nDesign/methodology/approach: - This paper draws on the notion of legitimate peripheral participation (LPP) from situated learning theory in order to examine the development of work-learning practices online. Semi-structured interviews were conducted with own-account self-employed workers (contractors and consultants who do not have staff) about their engagement in online communities for work learning.\nFindings: - Findings indicate that these self-employed workers were learning work practices, the viability of doing particular work, how to participate in online communities for work learning, and how to participate in fluid knowledges. The significance of developing a work-learning practice is emphasized, as is the impact of multiple and peripheral positionings across on- and offline spaces.\nResearch limitations/implications: - Web technologies and shifting configurations of online collectives shake up notions of expertise, beliefs about who is able to produce, and consume information, and where one locates themselves, in order to build work-learning practices. Multiple positioning across several online communities, and ways of participating that are peripheral, partial and part-time warrant further examination.\nOriginality/value: - The value of this paper is its exploration of how self-employed workers develop an online work-learning practice and the tensions that these practices bring. The paper also attempts to discuss the utility of LPP for contemporary learning practices.\n",
"role": "assistant"
}
] |
An analysis of risk for defaultable bond portfolios
|
[
"Risk analysis",
"Bonds",
"Autogressive processes"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: A corporate institution faces many types of risks. These risks affect its financial well being on a regular basis. In the extreme case, as a result of its exposure to such risks, the institution may not be able to survive as a going concern. In addition to the business risks that are specific to its market environment, a firm faces market risk and credit risk. Market risk arises from adverse movements in prices of financial assets such as equity, and market rates such as interest rates and exchange rates. Credit risk is the risk of loss arising from the failure of a counter-party to make a promised payments. Market risk and credit risk are not only intrinsically related to each other, but also nonseparable (Jarrow and Turnbull, 2000). In this paper we provide a regression quantile method to estimate the value-at-risk (VaR), and apply the model to returns data on portfolios of treasury securities and rated, defaultable corporate bonds, thus capturing the dynamics of market risk and credit risk[1].VaR has become a popular tool in the measurement and management of market risks (see Beder, 1995; Duffie and Pan, 1997; Dowd, 1998; Saunders, 1999) for reviews of literature on VaRs). It is viewed as the best measurement for market risk (Group of Thirty, 1993). Specifically, VaR is the loss in market value that is exceeded with a certain probability over a given time horizon, such a probability is often set at 1 or 5 percent. In requiring all US publicly traded corporations to report their quantitative market risk exposures, the Securities and Exchange Commission (SEC) (1997) lists VaR as a disclosure method \"expressing the potential loss in future earnings, fair values, or cash flows from market movements over a selected period of time and with a selected likelihood of occurrence.\" VaR disclosure communicates a single dollar amount for a public company's aggregate risk exposures, allowing for leverage, diversification of a variety of risk factors that affect the company's trading portfolios. Reporting VaR also forces companies to develop a systematic approach for risk measurement.Given the importance of VaR in reporting market risk and its prominence in risk measurement and risk management, it is not surprising that its estimation has attracted much attention from researchers. One popular approach to estimate VaR assumes a conditionally normal stock return distribution. The estimation of VaR is equivalent to estimating conditional volatility of returns[2]. Another popular method is to compute the empirical quantile nonparametrically, for example, rolling historical quantiles or Monte Carlo simulations based on an estimated model[3].However, these models are based on some restricted assumptions, such as normal, about the distributions of stock returns. There has been accumulated evidence that portfolio returns (or log returns) are usually not normally distributed. In particular, it is frequently found that market returns display structural shifts, negative skewness and excess kurtosis in the distribution of the time series. These market return characteristics suggest that more robust method in estimating VaR is needed. In this paper, we estimate VaR using a robust method based on quantile regressions. The quantile regression method is an extension of the empirical quantile methods. While classical linear regression methods based on minimizing sums of squared residuals enable one to estimate models for conditional mean functions, quantile regression methods offer a mechanism for estimating models for the conditional quantile functions, thus quantile regression is capable of providing a complete statistical analysis of the stochastic relationships among random variables (see Koenker and Bassett, 1978; Koenker, 1999).In this paper, we estimate VaR via a quantile regression model that allows for ARCH effect. VaRs estimated by this quantile regression approach display certain nice properties: they track VaRs estimated from GARCH volatility models well during normal market conditions. However, during a market turmoil when market drops are followed by further drops or rebounds, GARCH volatility models tend to predict implausibly high VaRs. This is due to that GARCH models treat both large positive and large negative return shocks as indicators of higher volatility, while only large negative return shocks indicate higher value at risk. Therefore, volatility and VaRs are not synonymous. VaRs estimated by the ARCH quantile regression model, while predicting higher volatility in the ARCH component, assigns a much larger weight to a big negative return shock than to a big positive return shock. The resulting estimated VaRs are therefore closer to reality: a large drop in market return is indicative of a high probability in both a further market drop and a market rebound.Other methods that allow for more general distributional assumptions include Engle and Manganelli (1999). Engle and Manganelli (1999) consider a different quantile regression based method. In particular, they consider an autoregression on the estimated VaRs. Comparing with these methods, the ARCH quantile regression model that we use in the current paper has the advantage of well-developed distributional theory that facilitates statistical inference and construction of confidence interval, as well as efficient computational advantages[4].While most existing VaR estimation methods were developed for stock portfolios and foreign exchange positions, financial institutions such as pension funds, mutual funds, insurance companies and hedge funds generally hold large positions of treasury securities and defaultable bonds, especially in recent years[5]. It is conceivable that fixed income securities display different risk characteristics from that of stocks and foreign exchanges. For example, defaultable bond returns are closely linked to the term structure of interest rates and the dynamics of the default premiums. Therefore, understanding the unique features of bond VaRs not only recognizes the growing popularity of bond market in the current financial world, but also has important practical implication for managing risks of fixed income securities[6].When applying our method to estimate bond portfolios, we add information in short-term interest rates and term spreads. Most existing methods of calculating VaR use univariate time series, ignoring information in related time series. However, information contained in other time series can be quite helpful and ignoring this information reduces the estimation efficiency, thus we propose a general model of calculating VaR that takes into accounts of useful covariates.In addition to measuring bond VaRs using quantile regressions, we also provide a sense of accuracy to our individual VaR estimates. While current VaR estimation methods generally do not provide a systematic approach to examine how accurate the estimated VaRs are, a number of analysts and investors have expressed concerns to the SEC about the accuracy of the quantitative, probabilistic information contained in firms' VaR disclosures (Hodder et al., 2001). Therefore, evaluating the accuracy of VaR estimates is important in VaR market risk reporting. In this study, we calculate confidence bands of the estimated VaRs under fairly general distribution assumptions. The results indicate that for the portfolios of treasury bonds, the estimated VaRs are generally small and the confidence intervals are narrow. However, large estimated VaRs and sometimes wide confidence intervals are associated with defaultable bond portfolios.This paper makes the following contributions to research in VaR and risk management in general. First, we propose a quantile regression model using useful covariates in estimating VaR for defaultable bond portfolios. The model allows for conditional heteroskedasticity. This procedure is easy to implement and the estimation programs are available in standard statistical packages such as S-Plus, and can also be easily written in other programming languages. Given that one criticism of using VaR as a reporting alternative of companies' quantitative market risk exposure is its difficulty in estimation (see Hodder et al., 2001), the approach we propose in this paper provides a convenient solution to estimating VaR. Second, we apply the proposed model to defaultable bond portfolios after considering information in related time series. In doing so, this paper illustrates the importance of using short-term interest rates and term spreads as covariates in defaultable bond VaR estimations. These covariates should be used in the ARCH component of the model. Lastly, we show that since a complete mean equation is hard to estimate with high statistical significance, keeping it as simple as an AR(1) model in returns is often sufficient. Therefore, covariates should not be used in the mean equation.The remainder of the paper is organized as follows. Section 2 presents a quantile regression model to estimate VaR. It derives confidence bands and model specification tests. Section 3 provides data descriptions and conducts the estimation. Empirical results regarding estimated VaR, confidence bands and model performance are reported and discussed in section 4. Section 5 contains concluding remarks.\n2. The model: 2.1 Analyzing risk by VaR\n3. Estimation procedure: 3.1 The data\n4. Empirical results: 4.1 Estimated VaRs\n5. Conclusions: The SEC has now listed VaR as one reporting alternative for quantifying companies' market risk exposures. Although the Securities and Exchange Commission (1998) stated that quantitative market risk disclosures provide \"new and useful information\" to the market, the quality of these disclosures, particularly the quality of VaR disclosures needs to be improved (Jorion, 2002). To improve the quality of VaR disclosure, a more robust estimation method is needed. In this paper we estimate value at risk using the quantile regression approach pioneered by Koenker and Bassett (1978). This method does not assume a particular conditional distribution for the returns. The model is applied to defaultable bond portfolios. We find that for defaultable bonds the use of information variables such as short term interest rates and term spreads as covariates improves the model performance significantly. We also find that confidence intervals constructed around the estimated VaRs can be very wide during volatile markets, making them less reliable when their accurate measurement is most needed.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this article is to estimate value at risk (VaR) using quantile regression and provide a risk analysis for defaultable bond portfolios.\nDesign/methodology/approach: - The method proposed is based on quantile regression pioneered by Koenker and Bassett. The quantile regression approach allows for a general treatment on the error distribution and is robust to distributions with heavy tails.\nFindings: - This article provides a risk analysis for defaultable bond portfolios using quantile regression method. In the proposed model we use information variables such as short-term interest rates and term spreads as covariates to improve the estimation accuracy. The study also finds that confidence intervals constructed around the estimated VaRs can be very wide under volatile market conditions, making the estimated VaRs less reliable when their accurate measurement is most needed.\nOriginality/value: - Provides a risk analysis for defaultable bond using quantile regression approach.\n",
"role": "assistant"
}
] |
HR holds back economic development in Kazakhstan: ... and may thwart ambitions to be a top 50 nation by 2020
|
[
"Human resource management",
"Economic development",
"Developing countries",
"Kazakhstan"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: President Nursultan Nazarbayev of Kazakhstan announced in January 2010 that the country would focus on becoming one of the world's top 50 economies by 2020. For a country only 20 years old and emerging from more than seven decades of communism, this is an ambitious target. Human-resource management has been highlighted as critical to realizing it.\nHRM and the 2020 Vision: The Strategic Plan of the Kazakh Ministry of Industry and New Technologies for 2011-2015 and The Strategic Plan of the Kazakh Ministry of Labor and Social Protection 2011-2015 both mention the central importance of human-resource management to achieving national development and innovation targets.The mission statement of the Ministry of Industry and New Technologies is \"to create conditions for the development of advanced, high-technology, competitive productions.\" Creating these conditions extends both a challenge and an opportunity to HR specialists involved in areas such as job design, workforce planning and learning and development.More specifically, the Kazakhstan prime minister, Karim Masimov, has said that by 2020 Kazakhstan will \"possess human resources sufficient for the development of a diversified economy\" (Kazakhstan 2007-2011 in Figures, 2011a, p. 1).Economic diversification is at the center of Kazakhstan's strategic development plans. The country depends heavily on oil and gas exports, which account for 90 percent of all outbound goods. The target is to reduce this to 55 percent by 2020 (Kazakhstan 2007-2011 in Figures, 2011b). Economic diversification, the government observes, will be achieved by \"increasing the potential and efficiency of human capital [which] will be facilitated by an integrated modernization of [the] national education and health-care systems, as well as more effective human-resource management\" (Kazakhstan 2007-2011 in Figures, 2011b). Policymakers evidently realize the role human-resource management must play in achieving the country's development goals, but do they know exactly how?More efficient human capital and more effective human-resource management, as the plans call for, will require HR specialists in Kazakhstan to drive innovation in companies; lead in the building of learning organizations and design strategies to attract, engage, develop, promote and retain the best people.Even improving Kazakhstan's performance in key areas of social and economic well-being will depend significantly on the work of HR. For example, reducing corruption in organizations requires, in part, the development and enforcement of codes of practice, leadership training, ethics workshops for all employees, embedding of corporate values in work practices and insightful recruitment and on-boarding of new hires. Similarly, increasing life expectancy includes making the workplace safer and healthier through job redesign, designing and enforcing occupational health and safety rules and procedures, introducing work-life balance initiatives to reduce employee stress and burn-out and introducing employee-wellness programs that educate and encourage healthy eating, exercise and relaxation.But are human-resource managers in Kazakhstan up to the challenge?\nHRM in Kazakhstan: At first glance the development of Kazakhstan's human capital seems to be progressing well. The workforce is increasingly well educated and skilled. Last year, for example, a large new university opened in Astana, the capital. Academics from UK universities have run start-up courses and now the university is recruiting full-time faculty internationally. Furthermore, in 2009 the United Nations named Kazakhstan as a country with high human potential, along with countries such as Malaysia and Brazil. Foreign investment, too, has been steadily increasing, especially in the oil and gas industries. Meanwhile, Almaty, Kazakhstan's commercial and business hub, has become the central Asian headquarters of international corporations including Microsoft, Procter and Gamble, Ernst and Young, Kimberly Clark, Coca-Cola, Samsung, BASF, British American Tobacco and Pricewaterhouse Coopers.However, ambition and investment have outpaced the capacity of human-resource specialists to provide frameworks to develop the human-capital competencies required to propel Kazakhstan forward. Innovation cannot be sustained with the current approach to human-resource management in Kazakhstan. The county is economically and commercially embracing the twenty-first century while its human-resource management - a critical enabler of this future state - has no discernible voice.One illustration of the challenge that lies ahead for human-resource specialists in Kazakhstan is the government's agenda to significantly reduce the number of foreign workers in the country in order to protect local jobs and reduce unemployment. In 2008 there were 54,204 foreign workers in Kazakhstan. Within 12 months the foreign labor quota had whittled this number down to 30,988 (Kazakhstan 2007-2011 in Figures, 2011b). While such policies play well on campaign trails, Kazakhstan is a long way from achieving fully-developed-nation status without these people.Many expatriate workers are knowledge workers - all are essential workers - employed mostly in oil and gas, health-care, education and financial services. Their role should be to transfer knowledge and oversee the skilling of local labor - much as they have been doing successfully in the Middle East for many years. Kazakhstan's economic rise is in its infancy and the country is far from being self-sufficient in the numbers of knowledge workers needed to meet its economic growth and diversification plans for Vision 2020. Without this infusion of foreign expertise and seeding of advanced work practices in the local workforce, Kazakhstani human-resource specialists will have it all to do.There is also evidence that indigenous companies are likely to be largely unaware of modern human-resource management practices. This poses a significant barrier to maximizing local human capital for the achieving the 2020 Vision. For example; Kazakhstan has no national human-resource institute or association while many other emerging, industrializing economies such as India, Philippines, China and Russia have well-established associations. HR conferences of international repute are unknown in Kazakhstan and it is expensive for local HR managers to travel overseas, even if it were encouraged.Currently, the organization of HR specialists exists largely at the grass-roots level. For example; there is a Kazakhstan HR community group on LinkedIn, a social-networking site for professionals. This group was formed four years ago yet still has only 513 members (LinkedIn, 2012) illustrating perhaps that the usually exponential growth of such virtual communities has failed to coalesce the country's HR community. Kazakhstan's premier university, KIMEP at Almaty, provides another example of the organic, localized nature of HR organization in Kazakhstan. The university has an HR club where a couple of dozen human-resource specialists from local companies gather periodically to network and hear guest speakers.Recent research (Davis and Yugay, 2011, 2012) has also highlighted that Kazakh-owned companies tend towards a dogmatic, prescriptive and functional approach to employee management, reminiscent of Soviet times, which significantly inhibits the ability and opportunity of employees to contribute in any creative way to their organization. The research suggests that strategic human-resource management and modern employee-management techniques are unknown to indigenous firms. Local managers focus on controlling and dictating to employees.Although literature on human-resource management in Kazakhstan is sparse, what there is concurs: human-resource management thinking and practice are under-developed in Kazakhstan and the specialism, such as it is, is disorganized. HR is also isolated both within and outside the country (International Business Wiki, n.d.).\nConclusion: Kazakhstan needs quickly to develop a labor force with the skills and knowledge capable of delivering Vision 2020 while at the same time radically accelerating the competence, connectedness and co-ordination of human-resource management specialists across the nation so that they can develop a shared practice. These specialists must implement the modern policies and practices that organizations rely on to maximize and strategically deploy their human capital.This currently appears to be a problem that people are ignorant of; there is no evidence anybody is concerned. However, without a monumental change in direction, the strategic plans and proclamations will likely remain rhetoric and much promise will go unfulfilled.\nNote: Paul J. Davis is Assistant Professor of Management at the Kazakhstan Institute of Management, Economics and Strategic Research in Almaty, Republic of Kazakhstan. He is an international management consultant specializing in strategic HR and has recently provided consultancy in India, Tanzania, Brunei and Malaysia. He earned a PhD in leadership and organizational development from Deakin University, Australia. He can be e-mailed at: pdavis@kimep.kz\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to describe how the under-developed state of human resource management (HRM) seems likely to thwart the economic and social development behind Kazakhstan's ambition to be a top 50 nation by 2020.\nDesign/methodology/approach: - The paper explains the background to Kazakhstan's Vision 2020, the current state of HRM in the country and the progress HRM specialists need to make if they are to be able to support the nation's rapid economic and social development.\nFindings: - The paper reveals that Kazakhstan needs quickly to develop a labor force with the skills and knowledge capable of delivering Vision 2020 while at the same time radically accelerating the competence, connectedness and co-ordination of HRM specialists across the nation so that they can develop a shared practice.\nPractical implications: - The paper urges the need for these specialists to implement the modern policies and practices that organizations rely on to maximize and strategically deploy their human capital.\nSocial implications: - The paper highlights how HRM can support a nation's entire economic and social development.\nOriginality/value: - The paper describes the state of HRM in a little-discussed country.\n",
"role": "assistant"
}
] |
Ownership structure, business growth and financial performance amongst SMEs: From Australia's business longitudinal survey
|
[
"Corporate ownership",
"Structures",
"Business development",
"Financial performance",
"Small to medium‐sized enterprises",
"Australia"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: A theoretical perspective that has contributed significantly to modern financial theory is that of agency theory, which considers a business enterprise from the viewpoint of the various stakeholders it might have, and explores how their financial interests are furthered and protected in their dealings with each other. The stakeholder relationships that receive most attention in the small and medium-sized enterprise (SME) literature are those between:* managers and owners;* owner-managers and other owners; and* insiders (primarily owner-managers, other owners and managers) and outsiders (mainly creditors and lenders).The central dilemma in agency theory is that day-to-day control of an SME's activities and financial fortunes very often rests in the hands of only some stakeholders who are usually managers or owner-managers. Yet all stakeholders have a legitimate expectation that their interests will be well served. By analogy with legal notions of agent and principal, those who exercise control are seen as being agents for the other stakeholders who are considered to be principals.The most significant problems that may arise from agency relationships in SMEs are:* information asymmetry;* moral hazard;* adverse selection; and* residual loss.One rational response to the inherent risk posed by self-interest on the part of agents, and other behaviours detrimental to principals in agency relationships, is for each stakeholder to increase the reward expected in return for participation in the business. Other possible responses are of two broad types:1. monitoring; and2. bonding.An important point about these responses to agency-related risk is that they are not costless. The out-of-pocket costs of maintaining an agency relationship may therefore include higher costs of financing, costs entailed in overcoming information asymmetries to achieve better monitoring of agents' activities, transaction costs incurred in setting up formal contractual arrangements with agents, and costs of incentive payments to agents. There are, in addition, less evident agency costs such as excessive perquisite consumption and shirking by agents, and also the opportunity costs arising from non-optimal financial decisions made by agents. Agency costs in total are the incremental costs explicitly and implicitly incurred over those that would be experienced in a perfect capital market situation, and typically they must ultimately be borne by the owners of the businesses concerned. Against these costs the owners must weigh the perceived benefits of agency relationships such as access to finance and management expertise they do not themselves possess.As a remedy to agency problems, bonding involves devising means by which congruence between the goals of agents and those of principals may be promoted. One approach is to use contracts which formally bind the parties to agreed types of behaviour, and which provide for sanctions should actual behaviour deviate from that specified in the contract. Another means of bonding agents and principals is through incentive schemes which encourage mutually acceptable behaviour without requiring it and without explicit sanctions. The possibility of failing to realise the benefits specified in the incentive scheme provides the motivation for the desired behaviour. Goal congruence may also be encouraged by adopting certain management practices that must be employed such as strategic and operational planning. Finally, goal congruence between agents and principals may be achieved by merging ownership and control so that managers are also at least part owners of the business. The idea is that if managers are also owners they are unlikely to knowingly undertake actions that are financially detrimental to owners. If this is the case, it is anticipated that the performance of a business with a high coincidence of ownership and control will exceed that of a business with some degree of separation of ownership and control. As will be seen, whether or not this expectation is realised has been the subject of extensive research (Dalton et al., 2003; Sundaramurthy et al., 2005). In the main, this research effort has been concerned with larger businesses and very little has focused on SMEs.The principal objective in this paper is to examine in some depth the relationships between ownership structure, business growth and financial performance amongst SMEs with different degrees of overlap between management and ownership. This is made possible by the recent availability of data from the Australian federal government's Business Longitudinal Survey (BLS). The paper proceeds as follows. After reviewing prior research on the relationship between ownership structure and performance, the current research method is outlined. Thereafter, the findings of the research are presented, followed by conclusions arising from this investigation.\nPrior research: Research on large businesses\nResearch method: The panel data employed in this research are drawn from the Business Longitudinal Survey (BLS) conducted by the Australian Bureau of Statistics (ABS) on behalf of the federal government over the four financial years 1994-1995 to 199719-98. Costing in excess of $4 million, the BLS was designed to provide information on the growth and performance of Australian employing businesses, and to identify selected economic and structural characteristics of these businesses.The ABS Business Register was used as the population frame for the survey, with approximately 13,000 business units being selected for inclusion in the 1994-1995 mailing of questionnaires. For the 1995-1996 survey, a sub-sample of the original selections for 1994-1995 was chosen, and this was supplemented with a sample of new business units added to the Business Register during 1995-1996. The sample for the 1996-1997 survey was again in two parts. The first formed the longitudinal or continuing part of the sample, comprising all those remaining live businesses from the 1995-1996 survey. The second part comprised a sample of new business units added to the Business Register during 1996-1997. A similar procedure was followed for the 1997-1998 survey. Approximately 6,400 business units were surveyed in each of 1995-1996, 1996-1997 and 1997-1998. The BLS did not employ completely random samples. The original population (for 1994-1995) was stratified by industry and business size, with equal probability sampling methods being employed within strata. Further stratification by innovation status, exporting status and growth status took place for the 1995-1996 survey.Data collection in the BLS was achieved through self-administered, structured questionnaires containing essentially closed questions. Copies of the questionnaires used in each of the four annual collections can be obtained from the ABS. The questionnaires were piloted prior to their first use, and were then progressively refined after each collection in the light of experience. Various imputation techniques, including matching with other data files available to the ABS, were employed to deal with any missing data. Because information collected in the BLS was sought under the authority of the Census and Statistics Act 1905, and thus provision of appropriate responses to the mailed questionnaires could be legally enforced by the Australian Statistician, response rates were very high by conventional research standards - typically exceeding 90 per cent.The specific BLS data used in this study are included in a Confidentialised Unit Record File (CURF) released by the ABS on CD-ROM in December, 1999. This CURF contains data on 9,731 business units employing fewer than 200 persons - broadly representing SMEs in the Australian context. Restricted industrial classification detail, no geographical indicators, presentation of enterprise age in ranges, and omission of certain data items obtained in the BLS all help to maintain the confidentiality of unit records. Furthermore, all financial variables have been subject to perturbation - a process in which values are slightly varied to provide further confidentiality protection.This research focuses on a longitudinal panel of 1,872 SMEs legally organised as proprietary companies. The main reason for limiting consideration to incorporated businesses is that this study necessarily involves comparing SMEs in terms of certain financial performance measures. This becomes problematic if the study sample contains both incorporated and unincorporated businesses because of the customary procedural difference in accounting for owners' wages, which are not separately reported in the BLS data. Furthermore, because a key question requesting information on the proportion of an SME's equity that is held by owner-managers was not asked in the 1994-95 survey, the analysis presented in this paper is confined to panel data for the 1995-96, 1996-97 and 1997-98 financial years only.Variables used in this research are either categorical in nature or, if metric, have irregular distributional properties (that is, they are non-normally distributed). Transformation of metric variables to produce normal distributions is avoided because of difficulties of interpretation often created by such procedures. Thus, non-parametric/distribution free techniques of statistical analysis are employed exclusively.\nResearch findings: Ownership structure\nSummary and conclusions: The key findings from this research into ownership structure, business growth and financial performance amongst Australian SMEs included in the BLS CURF panel can be summarised as follows:* On the basis of this study, it is not possible the reject the null hypothesis H0A that there is no statistically significant relationship between the proportion of equity held by SME managers and achieved business growth in the businesses examined.* On the basis of this study, it is not possible to reject the null hypothesis H0B that there is no statistically significant relationship between the proportion of equity held by SME managers and achieved financial performance in the businesses examined as measured by return on owners equity, return on total assets and net return on sales. While it is possible to reject H0B for asset turnover ratio, the effect is not strong enough to flow through to return on total assets and then to return on owners equity.It is important to recognise that seemingly negative findings like those above have their role to play in the advancement of knowledge. Aldrich (1992, p. 208) points out that:Entrepreneurship research resembles many other social sciences in the way that it systematically disdains two processes central to normal scientific endeavour: 1) attempted replication and confirmation of previous findings; and 2) publication of negative findings. Instead, we are treated to \"new\" concepts and \"positive\" findings, as authors try to differentiate their products from their potential competitors. Presently, we have persons calling themselves theorists who need not take seriously the stream of contradictory research findings pouring out of various research programs.The present study has attempted to replicate prior research conducted on large businesses with a sample of smaller concerns. It is a more thorough examination of the relationships between ownership structure, business growth and financial performance than can be found in the extant SME research literature. Ultimately, the findings do not differ from those emanating from larger business research. This is a worthwhile outcome to have, especially because (inter alia) it seriously challenges the relevance of agency theory in its application to smaller businesses.Beyond the inevitable bounds introduced to this research study by the broad data and methodological choices made, an important limitation to the analysis presented is that only linear relationships between ownership structure, business growth and financial performance have been sought. Furthermore, the presumed direction of causality has been from ownership structure (as the independent variable) to business growth or financial performance (as dependent variables). The underlying assumption has been made that ownership structure is exogenously rather than endogenously determined.After commenting on the relatively few studies that have examined non-monotonic relationships, Dalton et al. (2003, p. 21) express the opinion that \"Consideration of the potential for nonlinear relationships in the ownership-performance area may be...fruitful\". Of course, in their meta-analysis Sundaramurthy et al. (2005) anticipated this development by examining an hypothesis embodying an alignment effect followed by an entrenchment effect at higher levels of managerial ownership. More recent studies not covered by the Dalton et al. (2003) and Sundaramurthy et al. (2005) meta-analyses have pursued this line of inquiry. For example, Ng (2005) examines large family controlled listed corporations in Hong Kong and finds a cubic relationship between ownership structure and return on owners equity. This posits an entrenchment effect followed by an alignment effect followed by another entrenchment effect. Ng (2005) also finds evidence that ownership structure is exogenously determined and that the direction of causation is from ownership structure to financial performance. Clearly, a useful extension to the present study would be to employ non-linear statistical methods to see if Ng's (2005) findings can be replicated in an SME sample.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this research is to examine in some depth the relationships between ownership structure, business growth and financial performance amongst small and medium-sized enterprises (SMEs) with different degrees of overlap between management and ownership.\nDesign/methodology/approach: - The research employs panel data for 1,872 SMEs legally organised as proprietary companies, taken from the Australian federal government's Business Longitudinal Survey for three financial years from 1995-1996 to 1997-1998.\nFindings: - The study finds that there is no statistically significant relationship between the proportion of equity held by SME managers and achieved business growth in the businesses examined. Furthermore, for most financial performance measures examined, it would appear that there is no statistically significant relationship between the proportion of equity held by SME managers and achieved financial performance in the businesses examined.\nOriginality/value: - This is the first such investigation undertaken with SMEs and the findings are consistent with prior research on large business concerns. The findings can be seen as seriously challenging the relevance of agency theory in its application to smaller businesses.\n",
"role": "assistant"
}
] |
Influence of total quality management on human resource management practices: An exploratory study
|
[
"Sri Lanka",
"Human resource management",
"Total quality management",
"Process management",
"Human resource function",
"Human resource practices",
"Quality results"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The implementation of quality management techniques enables organisations to improve internal efficiencies, which is considered as a prerequisite to become competitive in global marketplace (Lambert and Ouedraogo, 2008; Stading and Vokurka, 2003). Total quality management (TQM) is an organisation-wide process-oriented philosophy that requires changes not only in production, but also in decision-making processes, employee development, and employee participation and involvement (Abdullah et al., 2009; Mehra et al., 2001; Power and Sohal, 2000). The research evidence shows that when organisations aim towards TQM they adopt more constructive approach to the management of human resources (HR) by upgrading the role of HR function, and redesigning existing human resource management (HRM) practices to fit into quality plans and objectives (Blackburn and Rosen, 1993; Vouzas, 2004). In other words, organisations introduce process improvement initiatives within the HR department in such a way to support strategic aims of quality and to bring it more in line with TQM principles (Greasley, 2004; Hur, 2009; Santos-Vijande and Alvarez-Gonzalez, 2009).However, existing research has not yet paid the due attention to changes occurred in the HR function and HRM practices as a result of process improvement initiatives under went by the HR department due to the implementation of TQM. The main shortcomings of previous empirical studies are that they include a limited number of HRM practices (e.g. recruitment and selection, performance management) in a single study (Ahmad and Schroeder, 2002; Soltani, 2005; Soltani et al., 2004, 2006). Therefore, it is required to analyze changes occurred in the HR function and HRM practices as a result of the implementation of TQM within a broad framework. Therefore, the current study extends the previous studies by analyzing the influence of TQM on several HRM practices, which until recently had been explored separately, in a single study.In the above context, this article presents results of the study conducted in Sri Lanka that investigated:* changes occurred in the HR function due to the implementation of TQM; and* changes occurred in HRM practices due to the implementation of TQM.The study focuses to the extent that TQM practices have been adopted by export-driven organisations in developing economies, and the related changes occurred in the HR function and HRM practices.An understanding of the above aspects is important to the development of quality management theory in the international context (Gosen et al., 2005; Mandal et al., 2000; Maheshwari and Zhao, 1994; Youssef and Zairi, 1995). The findings of this paper will add knowledge to current TQM practices since the international quality management literature provides little empirical evidence on the influence of TQM practices on HRM. Moreover, findings will potentially serve as a baseline, and become a source of general guidance in stimulating future studies in this area.Consistent with the objectives, the rest of the paper briefly describes the Sri Lankan context related to the study and reviews the theoretical background of the study. This is followed by the methodology adopted. Subsequently, the main findings are presented and discussed. The paper concludes with a discussion on the managerial implications of the findings.Background of the study: Sri Lankan context\nTheoretical background: Dimensions of TQM\nMethodology: Population, sample selection criteria and sample characteristics\nResults: Correlations of the TQM variables along with means and standard deviations are shown in Table II.Tables III and IV shows changes occurred in the HR function due to the implementation of TQM. As shown in Table III, 70 percent of the sample responded that HR activities become more formalised due to the implementation of TQM while 13 percent responded that HR department/unit has been newly established. As shown in Table IV, majority identified that firms are more focused on HRM issues (X-=4.25), HR function is more closely integrated with business needs (X-=4.12), and more involved in strategic decision making (X-=3.96) due to the implementation of TQM.Table V presents the results of factor analysis showing the changes occurred in HRM practices due to the implementation of TQM. Factor analysis yielded six factors that explained 84 percent of the variance. Factor 1 was labelled as \"performance management\", Factor 2 as \"competence development and career planning\", Factor 3 as \"rewards and recognition\", Factor 4 as \"recruitment and selection\", Factor 5 as \"HR planning\", and Factor 6 as \"satisfaction and well-being\". Therefore, results suggest that changes occurred in the way HRM practices were performed due to the implementation of TQM.\nDiscussion and implications: Whilst it is recognised that organisations introduce process improvement initiatives within the HR department to fit into TQM principles, no single study had attempted to investigate changes occurred in the HR function and several HRM practices as a result of process improvement initiatives under went by the HR department due to the TQM implementation. Therefore, the findings presented in this article make some contribution to the international quality literature on the influence of TQM practices on HRM.Being focused on a single industry, which could be identified as the leader in implementing quality management systems in Sri Lanka, ensured a high level of internal validity (Ahire et al., 1996). All the firms in the sample hold ISO 9000 quality certification which visibly demonstrates an organisation's commitment to worldwide quality. SLNQA programme that was started in the year 2001 has been formulated in line with the full model of MBNQA. Considering all these circumstances, the firms that implemented a formal TQM programme and running it for at least three years were taken into the sample. We consider all these as strengths of the study. We have used widely accepted multiple parameters drawn from literature to evaluate TQM dimensions. The TQM constructs used in this study were empirically based and shown to be valid and reliable in the Sri Lankan context. The 24-item measure was developed to explore changes occurred in HRM practices as a result of process improvement initiatives. Factor analysis yielded six components (performance management, competence development and career planning, rewards and recognition, recruitment and selection, HR planning, and satisfaction and well-being) and these were shown to be valid and reliable in the Sri Lankan context.The study investigated the level of TQM implementation based on the seven dimensions proposed by Raghunathan et al. (1997). These seven dimensions are leadership, information and analysis, strategic quality planning, HR development, quality assurance, supplier relationship, and customer orientation. Leadership, which is described in terms of personal involvement and visibility in creating goals, values and systems that guide in developing and maintaining an environment of quality excellence is vital in pursuit of continuous performance improvement (Raghunathan et al., 1997; Sun et al., 2006; Zhang et al., 2000). Continuous quality improvement relies on a steady flow of accurate information about processes and other constituencies such as employees, suppliers and customers. The importance of effective information usage is well highlighted in the quality literature (Ahire et al., 1996; Raghunathan et al., 1997; Sun et al., 2006). To achieve world class quality, firms should develop and realise the full potential of the HR and maintain an environment conducive to participation, quality leadership, and continual updating and improvement of individual and organisational growth. The findings highlight the importance of quality assurance by monitoring processes and systems to maintain the designed quality and make continuous improvements in quality. In the export-led industries like apparel production, quality assurance and customer orientation are key areas for success. When producing apparel targeting developed countries in the West, local manufactures have to strictly adhere to the product design specifications provided by the foreign buyer, who defines product design and quality. It should be noted that the industry's customers are the foreign buyers. Hence, customer satisfaction is the key for the survival and growth. The quality management literature also highlights the importance of customer satisfaction identifying it as the ultimate measure of firm performance, which may very well predict the future success or failure of an organisation (Kanji and Asher, 1993). However, mean values for supplier relationship is also important because when the apparel manufactures receive orders from foreign buyers with product design specifications, suppliers of the clothing materials and accessories and required quality are also specified (mainly materials and accessories are 100 percent imported as per product design specifications).One of the main objectives of the study was to explore changes occurred in the HR function and HRM practices as results of the quality improvement efforts. Majority of the firms identified HRM activities has become more formalised due to the implementation of TQM while some responded that HR department/unit has been newly established. Firms have become more focused on HRM issues, more closely integrated with business needs, and more involved in strategic decision making due to TQM initiatives. Some firms identified a decentralisation of HRM activities to line management and the emergence of a consultancy role for the HR department. This situation is also highlighted in previous quality literature (Psychogios et al., 2009). Overall, firms perceive several changes occurred due to the introduction of quality initiatives and firms have more business-oriented approach to HRM. The literature on quality system implementation also identifies that business process improvement activities due to quality system implementation leads to greater process understanding, ownership and redesign (Bendell, 2005).Considerable changes have taken place in the way HRM practices are performed as a result of quality initiatives. The findings led to identify such changes under six broad categories, namely, performance management, competence development and career planning, rewards and recognition, recruitment and selection, HR planning, and satisfaction and well-being. Overall, the findings imply that the implementation of quality management systems play and important role by altering the HR function, role of HR professionals, and HRM practices, and ultimately, move towards more strategic approach to HRM.\nLimitations and future research: There are some limitations in this study. First, data was collected from main persons involved in quality and HR departments as their awareness of quality and HR areas were considered as an advantage. However, the data collection from employees working in quality and HR departments, as well as other employees, would strength the results of the study. Due to scarcity of available previous empirical research investigating the HRM practices explored in this study, we developed item-measures for this study. Therefore, the prevalence of survey instrument related problems have to be acknowledged. However, the validity of a measure cannot be truly established on the basis of a single cross-sectional exploratory study. Further, the study dealt with seven TQM dimensions and six HRM practices. As the sample size is small (77 firms), structural modelling was not conducted to explore relationships between TQM dimensions and HRM practices. Though the study inquired into changes occurred in the HR function and practices due to quality system implementation the study has not inquired into tools and techniques used for process improvement. Future studies can investigate those. The nature of organisational change due to the implementation of quality management systems is both continuous and episodic as initial changes are often followed by waves of organisational change that would be political, social, cultural, economic or environmental. As an initial attempt, the study has not explored such changes. Future research could be designed to address such changes.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to explore changes occurring in the human resource (HR) function and human resource management (HRM) practices due to the implementation of total quality management (TQM) in Sri Lanka.\nDesign/methodology/approach: - Quality managers and HR managers from 77 export-based firms with ISO 9001 certification that have implemented a formal TQM programme, running for at least three years, participated in the survey. Factor analysis, correlation and regression were used for the data analysis.\nFindings: - It was found that firms introduce process improvement initiatives within the HR department by upgrading the role of the HR function, and by redesigning HRM practices of performance management, competence development and career planning, rewards and recognition, recruitment and selection, HR planning, and satisfaction and well-being to bring those in line with TQM requirements.\nOriginality/value: - Although QM literature is extensive, it is difficult to find studies that investigated to what extent TQM practices have been adopted by export-driven organisations in developing economies and what are the changes occurring in the HR function and HRM practices due to TQM initiatives. Understanding these is important to the development of QM theory in the international context. It is expected that the paper's findings will be able to establish baseline data to stimulate further research in this area.\n",
"role": "assistant"
}
] |
Finding that creative spark: Employee suggestion schemes and what makes them succeed
|
[
"Employees",
"Innovation",
"Suggestion schemes",
"Workplace creativity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Innovation is critical for any company hoping to successfully meet challenges and remain competitive. Finding ways to enhance creativity among the workforce should therefore be high on the agenda.\nEmployee suggestion schemes: A proven mechanism for achieving these goals is an employee suggestion scheme (ESS). This long-established tool is widely deployed to encourage idea generation among workers. Businesses have been using ESSs for more than a century now and their popularity continues to grow in several regions across the world. Nevertheless, many organizations still find it highly difficult to sustain them.\nWhy systems matter?: The latter also demands that an effective system is in place. This is acknowledged as one of the factors most critical to success. Having an appropriate mechanism that employees find simple to use can positively impact on both the amount and quality of ideas generated. And an infrastructure which enables ideas to be processed within a specified time frame will encourage a wider range of employees to participate. Increase the suggestions and increase the savings too.\nKey organizational attributes: A dedicated mechanism can reasonably be categorized as a factor pertaining to the organization or work environment. Numerous others can be similarly classified and include:\nThe individual touch: Of course, individual traits and job-related factors are bound to have considerable influence on the tendency to be creative. Personality, attitude and motivation are among the factors likely to be significant. In addition, creativity is more probable when the employee:\nComment: The review is based on \"Critical success factors for ESSs: a literature review\" by Lasrado et al. (2016). The authors suggest that workplace innovation can benefit from employee suggestion initiatives. Numerous key factors likely to impact on scheme success are identified in a piece that will be of considerable interest to practitioners because of its comprehensive and original approach.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to review the latest management developments across the globe and pinpoint practical implications from cutting-edge research and case studies.\nDesign/methodology/approach: This briefing is prepared by an independent writer who adds their own impartial comments and places the articles in context.\nFindings: Innovation is critical for any company hoping to successfully meet challenges and remain competitive. Finding ways to enhance creativity among the workforce should therefore be high on the agenda. A proven mechanism for achieving these goals is an employee suggestion scheme (ESS). This long-established tool is widely deployed to encourage idea generation among workers. Businesses have been using ESSs for more than a century now and their popularity continues to grow in several regions across the world. Nevertheless, many organizations still find it highly difficult to sustain them. The payoff for implementing suggestion schemes can be considerable.\nPractical implications: The paper provides strategic insights and practical thinking that have influenced some of the world's leading organizations.\nOriginality/value: The briefing saves busy executives and researchers hours of reading time by selecting only the very best, most pertinent information and presenting it in a condensed and easy-to-digest format.\n",
"role": "assistant"
}
] |
Managing change, growth and transformation: Case studies of organizations in an emerging economy
|
[
"Emerging economies",
"Organizational reputation",
"Model of organizational change",
"Value delivery capabilities",
"Managing change"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: It is a paradoxical but profoundly true and important principle of life that the most likely way to reach a goal is to be aiming not at that goal itself but at some more ambitious goal beyond it.(Arnold J. Toynbee)\n1. Introduction: Nations converge for shared purpose, stay interdependent for common challenges and diverge into their chosen paths (Dervis, 2012; Drabble et al., 2015; Jamali and Neville, 2011; Zhao and Liu, 2010). The changing roles of emerging economies in the global marketplace are worthy of careful examination. Emerging economies generally go through rapid transitions with fast economic growth. Two crucial growth factors - demographic and economic - make a huge impact on the speed of change in the lives of individuals, organizations and government sectors. One of the predominant emerging economies is India. Despite increasing research attention to India's dynamic economic growth, what has not been clearly explored are the processes of growth and transformation of various organizations in this emerging economy (Magala, 2016; Nairm and Vohra, 2011; Stephan et al., 2016). At the same time that India is experiencing rapid economic growth, the country also represents a culture with strong traditions and norms. The juxtaposition of these two factors - fast economic growth and robust cultural conventions - presents an intriguing opportunity to examine organizational change.\n2. Literature review: organizational change: Organizational change is defined as the \"difference in form, quality, or state over time in an organizational entity\" (Van de Ven et al., 1995, p. 512). The entity may be the overall organization, a subgroup within the organization or an individual job function. Research in the field of organizational change can be traced to the work of Kurt Lewin in the middle of the twentieth century. His work launched a generation of studies on the role of human behavior in organizational change, exploring how and why people respond to change, as well as the processes, methods and practices for effective change leadership.\n3. Case study methods: The literature on the subject of organizational changes in India is still in the early stage of development (Bhatnagar et al., 2010; Nairm and Vohra, 2011; Nandan and Verma, 2013; Rees and Hassard, 2010). For sound theory development in this area, it is essential to observe, examine and report the actual findings. For this purpose, a field study was conducted. India in particular provides rich soil to examine organizational change practices. First and foremost India is experiencing rapid change at both the macro-level (i.e. national and regional economies) and the micro-level (i.e. individuals and firms). With increasing business demands of globalization and economic growth imperative for a large population, India's annual economic growth rates and shifting demographic factors (e.g. rural regions to urban centers) signify huge societal changes. Yet, India is also a very traditional country where age-long customs may not likely change soon. In this context, it is worthy to note how organizations manage change processes while respecting timeless traditional norms and developing constructive citizenship behaviors in terms of individual work ethics and vibrant organizational culture (Bhatnagar et al., 2010; Miroshnik, 2012; Rees and Hassard, 2010; Wiener, 1988).\n4. Research model and case study findings: In this section, we summarize the case study findings and present our model of organizational change in an emerging economy. The organizations indicate different growth and change patterns. Growth and changes are not necessarily in linear fashion. Like individuals, organizations also go through ups-downs in each stage. However, change processes also suggest certain patterns of growth - from birth to competition for survival, growth, distinction and prominence. Both organizations have not experienced all these steps. Realistically, Vindhya is going through the third stage of growth, whereas CHRIST (Deemed to be University) is passing through the fourth stage, respectively, for excellence and distinction. CHRIST (Deemed to be University) has demonstrated instructional excellence at national accreditation by National Assessment and Accreditation Council with the highest grade of A and is in the course of developing excellent research capabilities. These organizations are common in that their organizations successfully implemented change management with proper steps of transition and transformation.\n5. Discussion: The extant organizational change literature predominantly focuses on content, context, process and criterion issues in the context of advanced economies. We have examined how organizations in an emerging economy, India, have managed change requirements for rapid quantitative growth goals while maintaining incremental qualitative cultural norms. Organizations within this emerging economy present an interesting case study given their unique challenges of managing growth along with respecting long-time cultural traditions.\n",
"role": "user"
},
{
"content": "Purpose: In view of dynamic and widespread economic transformation in emerging economies, managing organizational change and growth in this context deserves more research attention. The purpose of this paper is to examine how three organizations in different industries manage change, growth and transformation in their organizational ecosystem.\nDesign/methodology/approach: The authors conducted in-depth interviews with the leadership of three organizations in different economic sectors in India, a country representing an emerging economy. The authors also reviewed historical data from these organizations. Three case studies illustrating the evolution of these organizations were developed from the data collected.\nFindings: Lessons and implications from the three case studies suggest the following key elements of effective organizational change mechanisms in an emerging economy: visionary entrepreneurial leadership; program quality excellence; scale growth and scope expansion; network capabilities; and sustainable stakeholders' engagement. At the same time, this study also shows how these organizations manage change, growth and transformation in the context of a society with strong traditions and cultural norms.\nResearch limitations/implications: Results and conclusions may be limited by the fact that the study is based on three case studies. Additional studies from a variety of industries with large numbers of participants will be helpful in more fully understanding the ways in which change, growth and transformation can best be developed and deployed in different organizational settings.\nPractical implications: The proposed model of organizational change in an emerging economy may assist organizational leadership in designing and sustaining their change efforts.\nSocial implications: This study highlights the role of visionary entrepreneurial leadership and the impact of organizational growth mechanisms on organizational value delivery capabilities and organizational reputation.\nOriginality/value: Lessons and implications of five growth steps of outstanding organizations in an emerging economy context provide valuable insight for organizational change, growth and transformation in other emerging contexts.\n",
"role": "assistant"
}
] |
Eliciting consumer preferences for credence attributes in a fine-dining restaurant
|
[
"Organic",
"Consumer preferences",
"Animal welfare",
"Fine-dining restaurant",
"Natural field experiment"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The restaurant sector is a large and growing food sector that experiences tough competition. The knowledge about consumer preferences is needed for restaurants to survive. To date, most consumer studies on commercial full-service restaurants have been survey based (Harrington et al., 2011; Heung, 2002; Hwang et al., 2012; Jin et al., 2016; Longart, 2010; Ryu and Jang, 2007; Upadhyay et al., 2007; Wu and Liang, 2009) or focussed on calorie labelling (see Long et al., 2015 for a review and meta-analysis). This paper presents a menu-labelling experiment conducted in a fine-dining restaurant in Norway. In manipulating the menu description and price of an organic veal entree, this paper investigates how information about credence attributes and price affects customers' choice in a commercial full-service restaurant.\n2. Literature review of restaurant experiments: Most of the recent studies that used some kind of restaurant experiment focussed on nutritional labelling (e.g. Howlett et al., 2009; Long et al., 2015; Roberto et al., 2010). Few researchers have studied the effect of other types of credence attributes, such as origin, organic, animal welfare, sustainability, and fair trade. Furthermore, no study has performed a natural field experiment in a fine-dining restaurant. Most restaurant experiments were conducted in restaurants associated with universities. In most of these studies, the customers were fully aware that they are part of a study even though they usually did not know the purpose of the study. This review includes experimental papers that manipulated their menus in other ways than calorie labelling.\n3. Natural field experiment in restaurants: illustrative case: 3.1 Description of the menu-labelling experiment\n4. Discussion: 4.1 Why conduct a natural field restaurant experiment?\n5. Concluding remarks: The restaurant market is a very important market for agricultural goods; however, only a very small portion of the very extensive literature on consumer preference for food products attributes focus on this market. The exception is the relatively new literature on nutritional labelling in restaurants. This paper described a menu-labelling experiment in a commercial full-service restaurant where the participants were ordinary customers walking in from the street and did not know that they were participating in an experiment. The standard choice methodology was used in both design (fractional factorial design for attributes) and analysis (logit models). The pros and cons of natural field experiments in restaurants and some design challenges were also discussed.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to illustrate and discuss methodology issues in menu-labelling experiments in commercial full-service restaurants, and to investigate how the menu description and price affects customers' choice of an organic veal entree in a Norwegian restaurant.\nDesign/methodology/approach: A menu-labelling experiment was conducted in a fine-dining restaurant during ordinary opening hours over a period of two weeks. The menu description of an organic veal entree was altered repeatedly and the effect of these changes on the sales of this entree was investigated.\nFindings: Adding words to the menu description, such as \"organic\", or describing animal welfare had a very limited effect on customers' choices in the restaurant.\nResearch limitations/implications: The research illustrates the use of a natural field experiment in a commercial full-service restaurant and discusses strengths and weaknesses of the methodology.\nOriginality/value: Few experiments have been performed on the effect on credence attributes in commercial full-service restaurants and there is little knowledge about research challenges in menu-labelling experiments. This paper contributes to the knowledge on both issues by conducting a natural field experiment in a fine-dining restaurant.\n",
"role": "assistant"
}
] |
Statistical and measurement pitfalls in the use of meta-regression in meta-analysis
|
[
"Meta-analysis",
"Statistics",
"Pitfalls"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An important aspect of almost all meta-analyses today is the search for moderators - study characteristics that correlate with study outcomes (which are usually r values or d values) and might have explanatory power. Moderator variables are often referred to as \"boundary conditions\" and can be of practical and/or theoretical importance. The two most frequently used methods of attempting to identify moderators are meta-regression and subgrouping of studies, with hierarchical subgrouping being required to avoid the confounding of correlated moderators (Schmidt and Hunter, 2015, pp. 381-384). This paper discusses the relative strengths and weaknesses of these two methods. The resulting conclusion is that hierarchical subgrouping of studies is usually preferable to meta-regression.\nInflated multiple rs: The first problem is the capitalization on chance (sampling error) resulting in inflated multiple rs (Raudenbush, 2009), especially when more than a few moderator candidates are included in the equation. The square of the multiple r is then falsely interpreted as the proportion of variance in the r or d values \"explained by the moderators.\" In meta-regression applications in the literature, the appropriate shrinkage formula to adjust for the inflation in the multiple r (Cattin, 1980) is almost never applied. What makes this problem especially serious in meta-regression is the fact that the sample size (k) is usually small in relation to the number of moderator candidates in the regression equation. In addition, as pointed out by Viechtbauer (2007), there is a serious risk of Type I errors when significance tests are applied to multiple individual regression coefficients, as is usually done. Statistical significance tests are almost always used with meta-regression. Such tests are not used when the subgrouping method is employed. It is well known that the use of significance tests in data analysis creates serious problems in the attainment of accurate scientific conclusions (e.g. see Harlow et al., 1997; Schmidt, 1996; Schmidt and Hunter, 1997).\nEx post facto predictor selection: data dredging: The second problem is referred to by Thompson and Higgins (2002) as \"the principal pitfall in meta-regression.\" This problem stems from the fact that the potential moderators examined in the meta-regression are almost never specified in advance. Instead, only those with the highest ex post facto observed correlations with the effect sizes are included in the meta-regression analysis (\"the most promising moderators\"). This ex post facto selection of potential moderators results in highly inflated multiple rs. (See Schmidt and Hunter, 2015, pp. 80-83, for some dramatic examples of this.) Shrinkage formulas do not and cannot correct for this sort of capitalization on sampling error. Some who follow this practice focus not on the multiple r but on the statistically significant regression weights. But these are also distorted by capitalization on sampling error (and by misspecification of the model resulting from ex post facto predictor selection). The practice of ex post facto selection of predictors is so common as to be nearly universal in published meta-analysis using meta-regression. In a conversation I had with Larry Hedges, he stated that this problem is so widespread that one can almost not trust any published meta-regression analyses. Thompson and Higgins (2002) state that this form of data dredging is the most serious problem in published meta-regressions and the most difficult one for a reader to detect. It essentially is a form of harking (hypothesizing after the results are known; Kerr, 1998) at the level of meta-analysis.\nLow statistical power: The third major problem is that statistical power is typically low, because the number of studies (k is the relevant n) is almost never large (Hedges and Pigott, 2004; Schmidt and Hunter, 2015; Thompson and Higgins, 2002; Viechtbauer, 2007). (Recall that significance tests are almost always used in meta-regression.) Because of low power, the regression weights for most real moderators will be non-significant. (At least they should be, given known statistical principles; yet most moderators in the literature are reported to be statistically significant, which raises suspicions about capitalization on chance and/or selective reporting.) Hedges and Pigott (2004) recommend that the statistical power of any planned meta-regression should be estimated prior to application; they further recommend that if power is found to be low (usually the case), then meta-regression should not be applied because of the danger that failure to detect moderators will be taken as evidence of the absence of moderators, rather than correctly being attributed to low statistical power. Hedges and Pigott (2004) did not address the problems created by capitalization on sampling error, discussed in the first two points above. So the picture is even direr than presented by them.\nOutsized effect of outlier data points: The fourth disadvantage is the susceptibility to distortion by outlier data points. This consideration exists in all applications of regression but it is much more serious when sample sizes are small in relation to the number of predictors (Stevens, 1984). In meta-regression, the sample size (the number of studies, k) is often as small as 15, 20, or 30, and the number of potential moderators (predictors) may be as large as 5 or 10 or more. (Outliers can also affect the subgrouping method of moderator detection but the effect is not as distorting and an outlier is likely to affect only one of the subgroups rather than the whole analysis).\nUninterpretable unstandardized regression weights: The fifth disadvantage stems from the fact that the obtained regression weights are unstandardized (raw score) regression weights. Because of this, they are difficult or impossible to interpret in any substantive way and their magnitude cannot be compared with that of other regression weights in the meta-regression equation. The size of the regression weight on a hypothesized moderator depends on how that moderator is scaled or measured. Because different moderators are measured on different scales with different standard deviation (SDs), the magnitudes of the regression weights are not comparable and their magnitudes cannot be compared to each other. This fact leads many users of meta-regression to focus almost entirely on p-values in comparing moderators, which is an undesirable emphasis. Those with the smallest p-values are then deemed to be the largest moderators, which is typically a false conclusion.\nDifferential measurement biases in the dependent variable: The sixth disadvantage is the fact that meta-regression results are inaccurate when the d or r values have not been corrected for the measurement error (and for range restriction, where applicable) (Ones et al., 2012; Schmidt and Hunter, 2015, Chapter 9, pp. 384-390; Schmidt et al., 2009). While all the values will be biased downward by the measurement error, some will be biased more than others, undercutting the construct validity of the observed ds or rs as measures of the real effects. The rank order of the effect sizes can be changed from the true rank order by the effects of differential measurement error levels, thus destroying even the ordinal construct validity of the dependent variable. This effect artifactually reduces the apparent strength of all true moderators. It is rare in published meta-regressions for these corrections to be made. Thompson and Higgins (2002) and Viechtbauer (2007) acknowledge this problem but do not present any means of correcting for the measurement error.\nMeasurement error in moderator measures: The seventh problem is measurement error in the measures of the hypothesized moderator variables. Orwin and Cordray (1985) showed that failure to correct for the measurement error in the measures of the moderator variables leads to serious errors in the meta-regression results. This finding is important because almost no meta-analyses in the literature using meta-regression make this correction, meaning that their moderator results are suspect. (For further discussion of this problem and others, see Cordray and Morphy, 2009.) Thompson and Higgins (2002) acknowledge this problem but do not discuss the possibility of correcting for the measurement error in the moderator measures.\nUnreliability in the dependent variable: The eighth problem is that, even if the d or r values are corrected for the measurement error and other artifacts, there is still the problem created by the fact that much (often most and sometimes all) of the variance in the d values or rs (i.e. the dependent variable) is due to the sampling error, creating low or zero reliability for the dependent variable and hence low or zero statistical power to detect moderator effects (as discussed in Cook et al., 1992, Chapter 8, pp. 325-326). Aloe et al. (2010) have proposed an adjustment for the sampling error in the effect sizes (which functions as the measurement error in meta-regression).This adjustment removes the downward biasing effects of sampling error from the multiple correlation and is similar to the Schmidt-Hunter correction for effect size unreliability due to sampling error variance. (See pp. 107-110 in Schmidt and Hunter, 2015, for an illustration of this correction.) However, the Aloe et al. method does not address the resulting problems of low statistical power in the moderator analysis.\nProblems of missing data: The ninth disadvantage stems from problematic data requirements. If the potential moderator measures are to be standardized, the use of meta-regression requires estimates of the correlations among the potential moderators (predictors). If unstandardized (raw score) regression is to be used, there is the additional requirement of knowledge of the SDs and means of the moderator measures. Often, estimates of some of these correlations are not available and must be guessed at or somehow imputed. The same is true for the SDs and means. This potentially adds additional error to the meta-regression results. Thompson and Higgins (2002) discuss this missing data problem (along with the fact that some studies fail to include some of the potential moderators of interest that the meta-analyst would like to investigate).\nConclusion on meta-regression: In articles and textbooks, meta-regression procedures are often presented and described without any mention of these problems (e.g. Lipsey and Wilson, 2001), and virtually none of the published meta-analyses using meta-regression discuss or even mention these problems. In light of the serious limitations discussed in this paper, it can be seen that it is only under rare and unusual circumstances that meta-regression produces reliable and valid results. Meta-regression is used with great frequency in the literature today, and it is highly likely that most of the results are not trustworthy.\nWhat should be done?: If moderators are dichotomous or categorical (e.g. sex or race), the subgrouping approach to moderator analysis is superior. However, it is important to bear in mind that moderators are often correlated and that it is important to apply hierarchical moderator analysis to avoid confounding of correlated moderators. Hierarchical meta-analysis is discussed in Schmidt and Hunter (2015, pp. 381-384). In hierarchical meta-analysis, moderator categories are sequentially subgrouped until the breakdown is complete. Examples of hierarchical meta-analysis include Ones et al. (1993) and Rodgers and Hunter (1986). For example, in the Rodgers and Hunter meta-analysis of the effectiveness of management by objectives, the potential moderator variables of upper management commitment and duration of the program were correlated, and when they were considered separately (in non-hierarchical subgroup analyses), both appeared to increase the effectiveness of the program. But hierarchical subgroup analysis showed that only one of these variables increased program effectiveness - upper management commitment to the program. When that moderator was absent, the program was ineffective, no matter how long it had been in place.\nOverall conclusion: Although there is no perfect method for attempting to identify and calibrate moderator variables in meta-analysis, on balance, the hierarchical subgrouping method is preferable to meta-regression because meta-regression suffers from more statistical and measurement pitfalls that render its results deceptive and untrustworthy.\n",
"role": "user"
},
{
"content": "Purpose: Meta-regression is widely used and misused today in meta-analyses in psychology, organizational behavior, marketing, management, and other social sciences, as an approach to the identification and calibration of moderators, with most users being unaware of serious problems in its use. The purpose of this paper is to describe nine serious methodological problems that plague applications of meta-regression.\nDesign/methodology/approach: This paper is methodological in nature and is based on well-established principles of measurement and statistics. These principles are used to illuminate the potential pitfalls in typical applications of meta-regression.\nFindings: The analysis in this paper demonstrates that many of the nine statistical and measurement pitfalls in the use of meta-regression are nearly universal in applications in the literature, leading to the conclusion that few meta-regressions in the literature today are trustworthy. A second conclusion is that in almost all cases, hierarchical subgrouping of studies is superior to meta-regression as a method of identifying and calibrating moderators. Finally, a third conclusion is that, contrary to popular belief among researchers, the process of accurately identifying and calibrating moderators, even with the best available methods, is complex, difficult, and data demanding.\nPractical implications: This paper provides useful guidance to meta-analytic researchers that will improve the practice of moderator identification and calibration in social science research literatures.\nSocial implications: Today, many important decisions are made on the basis of the results of meta-analyses. These include decisions in medicine, pharmacology, applied psychology, management, marketing, social policy, and other social sciences. The guidance provided in this paper will improve the quality of such decisions by improving the accuracy and trustworthiness of meta-analytic results.\nOriginality/value: This paper is original and valuable in that there is no similar listing and discussion of the pitfalls in the use of meta-regression in the literature, and there is currently a widespread lack of knowledge of these problems among meta-analytic researchers in all disciplines.\n",
"role": "assistant"
}
] |
Documenting acousmatic music interpretation: Profiles of discourse across multiple dimensions
|
[
"Grounded theory",
"Cultural heritage",
"Acousmatic music",
"Discourse profile",
"FRBRoo",
"Music information objects",
"Performance documentation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The development of electroacoustic music during the second part of the twentieth century raised new questions concerning live performance in concert. Could a musical piece which is fixed on a physical medium and diffused through loudspeakers be \"interpreted\"? In 1997, the Academie Internationale de Musique Electroacoustique de Bourges dedicated their proceedings to Maurice Leroux, the first interpreter of electroacoustic music, who performed the collaborative work entitled \"Symphonie pour un homme seul\" by Pierre Schaeffer and Pierre Henry at the Theatre de l'Empire in Paris on July 6, 1951 (Schaeffer et al., 1955). Leroux performed with a device called the \"Portique potentiometrique de Relief,\" which was conceptualized by composer Pierre Schaeffer and designed by engineer Jacques Poullin. The interpreter held a transmitter coil which he moved within a space defined by four receiver coils. His movements allowed for the creation of sound trajectories among multiple loudspeakers positioned in the concert hall (Poullin, 1955, p. 16). Sound systems used for diffusing electroacoustic music became more and more sophisticated until the development of the loudspeaker orchestra in the 1970s. It was here that the term acousmatic music first came into usage.\nMethodology: From October 2014 to November 2017, we conducted 12 semi-structured interviews with musicians with renowned expertise in the composition and interpretation of acousmatic music. The scope of the research was limited to French-speaking artists, for several reasons: the ability of the researchers to analyze the data; the presence of a large international francophone acousmatic community; as well as the historical relation of the domain with French-speaking countries. The authors conducted interviews in France, Belgium and Quebec across a wide range of approaches and schools, following the principles of Strauss and Corbin's (1994) grounded theory, aiming at theoretical saturation (Glaser and Strauss, 1967). The participants included eight men and four women; this discrepancy is noted as a second limitation. The distribution of participants' expertise is presented in Figures 1 and 2. Participants include pioneers from the 1960s to the 1970s, the next generation of the 1980s and more recent musicians who started in the 1990s and early 2000s and who have already acquired relevant expertise (see Figure 1). Six participants interpreted more than 100 works from other composers while three have mostly played their own works (see Figure 2). The former category (i.e. more than 100) is unrelated to the five categories of the distribution articulated by Figure 1, as it includes at least one participant from each one of them, and is also unrelated to the geographical distribution as it includes participants from France, Belgium and Quebec.\nThe multiples dimensions of acousmatic interpretation: Two main categories emerge from the analysis in relation to the scope of the research implemented in the semi-structured questionnaire. These categories are: definition of interpretation and organization of interpretation practice. The latter provides us with structured and detailed insights into the preparatory work, off-site and on-site, as well as socio-economic constraints for the production of acousmatic concerts. The hierarchical structure of the conceptualization resulting from the grounded theory analysis of the data emerged inductively from the more concrete concepts to the more abstract concepts of these two main categories.\nDiscussion: Theory development\nConclusion: \"Music for tape is not dead, it also needs to be 'performed,'\" argues Risset (1998). The holistic framework we propose aims at facilitating the documentation and the study of acousmatic interpretation. We have defined and discussed the relevance of eight dimensions pertaining to the conceptualization of this activity, including the discussion about the ambiguities it builds upon: ontological, linguistic, anthropological. These eight dimensions provide us with a theoretical and practical tool for the validation of documentation methodologies as well as the development of more specific frameworks targeted at some of them. We also propose to distinguish between two perspectives on the conceptualization of documentation for acousmatic interpretation.\n",
"role": "user"
},
{
"content": "Purpose: Extending documentation and analysis frameworks for acousmatic music to performance/interpretation, from an information science point of view, will benefit the transmission and preservation of a repertoire with an idiosyncratic relation to performance and technology. The purpose of this paper is to present the outcome of a qualitative research aiming at providing a conceptual model theorizing the intricate relationships between the multiple dimensions of acousmatic music interpretation.\nDesign/methodology/approach: The methodology relies on the grounded theory. A total of 12 Interviews were conducted over a period of three years in France, Quebec and Belgium, grounded in theoretical sampling.\nFindings: The analysis outcome describes eight dimensions in acousmatic performance, namely, musical, technical, anthropological, psychological, social, cultural, linguistic and ontological. Discourse profiles are provided in relation to each participant. Theory development led to the distinction between documentation of interpretation as an expertise and as a profession.\nResearch limitations/implications: Data collection is limited to French-speaking experts, for historical and methodological reasons.\nPractical implications: The model stemming from the analysis provides a framework for documentation which will benefit practitioners and organizations dedicated to the dissemination of acousmatic music. The model also provides this community with a tool for characterizing expert discourses about acousmatic performance and identifying content areas to further investigate. From a research point of view, the theorization leads to the specification of new directions and the identification of relevant epistemological frameworks.\nOriginality/value: This research brings a new vision of acousmatic interpretation, extending the literature on this repertoire's performance with a more holistic perspective.\n",
"role": "assistant"
}
] |
Corporate life and death: who succumbs, who survives during disruption?
|
[
"Six Sigma"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nThe economy disrupted, 1996-2018: The purpose of this paper is to understand why a company dies, survives or thrives during periods of major macro transformation and what, specifically, we can learn from such diverse outcomes. This is particularly relevant today as the US economy has faced and continues to face major disruptions, including the internet, China's rise as a manufacturing powerhouse and, most recently, a new wave of tariffs and related trade uncertainties, discussed in the following text. To understand the extent of change in the macro environment in the USA since 1996 and its impact on US companies, this study used the S&P 500 as a proxy for the broader US economy and compared the S&P 500 lists from 1996 and 2018.\nLessons from the US steel industry: The steel industry is a particularly rich area of exploration for this analysis because of the diversity of outcomes across companies and because the US steel industry has faced severe macro challenges during the past two decades[2]. Once the dominant world player and a major employer of USA workers, the USA steel industry has been under siege for decades and is now ranked below China, Japan and India in terms of production. More than 30 US steel companies went bankrupt between the 1990s and 2004, and, adding in the effects of automation, the industry has lost some 48,000 jobs since 2000 (Pete,2016). Emblematic here is Bethlehem Steel. Launched in 1904, the company that gave us such landmarks as the Golden Gate Bridge and the Chrysler Building died in 2003. This death is especially troubling given the company's once formidable role in the US economy, with 300,000 workers at its peak in 1941 (Loomis, 2004; Berg, 2001).\nLessons from Korea's POSCO: Looking farther afield, Korea's POSCO managed not just to survive but to thrive when faced with cheap steel from China, rising to third worldwide today in terms of production, with levels of efficiency, productivity and profitability that consistently exceed industry averages. Named the world's most competitive steelmaker for nine straight years by World Steel Dynamics, POSCO has also been listed on the Dow Jones Sustainability Index for 13 years in a row (SteelGuru, 2018; AutomotivePurchasing.com, 2017). POSCO has achieved these results by creating a culture of continual innovation among its employees and its customers, suppliers and other partners, all of which enables the company to compete on the basis of value-added differentiation.\nLessons from OMC: In the case of OMC, as with Bethlehem, poor financial management was the proximate cause of the company's death. But the deeper problems centered on a longer-term failure to adapt as new entrants disrupted the market. An outgrowth of Evinrude, founded in 1907, OMC made outboard motors as well as several lines of boats. Over the years, OMC also diversified, mainly through acquisitions, into lawn mowers, golf carts, drainage pumps, and other products. Some products, including golf carts, proved less profitable than OMC's outboard motor segment. Periodically saddled with high levels of debt and losses as a result of its diversification strategy, OMC nevertheless remained dominant for years in outboard motors, with about 50 per cent of the global market in the 1960s (BOATMAGS.com, 2016).\nA disruption survival guide: Disruption, whether from foreign or domestic competitors, should be seen as inevitable, requiring long-term planning and ongoing strategic responses. To survive, even thrive, in times of disruption, the key takeaways from these cases are as follows:\n",
"role": "user"
},
{
"content": "Purpose: This paper provides a disruption survival guide for companies during times of macro transformation.\nDesign/methodology/approach: The research compared the S&P 500 - as proxy for the US economy - from 1996 and 2018. Of companies no longer on the list, 12 died, including two from manufacturing: Bethlehem Steel and Outboard Marine Corporation (OMC). Still, some US manufacturers survived or even thrived during the period. To understand why, the paper compares Bethlehem, which died in 2003, to US Steel, which has survived but was removed from the S&P 500 list in 2014, and to Nucor, which has stayed on the list. POSCO is also used for comparison. The OMC case adds a different industry perspective.\nFindings: The main findings from these cases are as follows: stay fit financially and avoid overreaching in good times, use operation strategies such as Lean and Six Sigma to build a culture of continual innovation and stay close to customers to compete on the basis of differentiation, particularly if competing on price is not a realistic option. The good news is differentiation is possible even in seemingly commoditized sectors like steel.\nResearch limitations/implications: This paper contributes to the literature on differentiation as a strategy for competing with low-cost disruptors.\nPractical implications: This paper provides insights into the use of Lean, Six Sigma and other strategies for creating a culture of continual innovation among employees, customers, suppliers and other strategic partners. And, building on this culture, to compete on the basis of value-added differentiation, particularly if competing on price is not a realistic option.\nOriginality/value: The paper cuts through complex, fast-changing, transformative macro issues - e.g., Chinese competition and trade uncertainties related to new tariffs - and provides practical, timeless insights for navigating in such times. The focus here is on strategies for competing on the basis of value-added differentiation, particularly if competing on price is not a viable option. The good news is such competition is possible even in seemingly commoditized sectors like steel.\n",
"role": "assistant"
}
] |
Exploring key antecedents of purchase intentions within different services
|
[
"Service quality",
"Customer satisfaction",
"Purchase intentions",
"Physical environment",
"Advertising campaign familiarity",
"Service-oriented employee behavior"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this issue.\nIntroduction: The recent global economic crisis has affected most businesses, including those in South Korea. Subsequently, the Korean business milieu has become intensely competitive, particularly in the retail services sector (e.g. supermarkets and financial institutions). Managers are, thus, under increased pressure to maintain sustained competitiveness within an uncertain environment. To accomplish this, managers have to constantly align marketing stimuli in ways that help create, maintain and sustain competitive advantage. With the objective of establishing further understanding in this area, this research explores relationships between key antecedents and purchase intention in the South Korean retail services sector.\nResearch methodology: Measures\nResults: Measurement model testing and results\nDiscussion: We tested a model of marketing stimuli in relation to quality and satisfaction, which in turn influence purchase intentions in two different industries. This study further investigated the independent role of three key dimensions (ad campaign familiarity, service-oriented employee behaviors and physical environment). These findings add value to extant research in the following ways:\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this research is to investigate the direct and indirect effects of these determinants on purchase intentions. Competitive environments such as those in retail and banking industries impose increased pressure on managers to enhance customer satisfaction and purchase intentions. Even though satisfaction and purchase intentions are well studied in prior literature, their determinants such as service-oriented employee behavior, advertising campaign familiarity, physical environment and service quality have not been fully investigated together.\nDesign/methodology/approach: Data from a survey of 508 customers of Korean up-market retail supermarkets and banks are utilized to empirically evaluate the model.\nFindings: Results indicate that service-oriented employee behavior and the physical environment have no direct effect on purchase intentions. However, these constructs indirectly influence purchase intentions through the mediating role of service quality and customer satisfaction. Service-oriented employee behaviors play a major role in enhancing service quality and customer satisfaction, but do not directly impact purchase intentions. Interestingly, the strongest direct effect on purchase intentions is more likely to come from service quality, rather than satisfaction.\nPractical implications: Findings suggest that for service industries such as retail and banking, it may be strategic to invest more resources aimed at enhancing service-oriented employee behaviors and the physical retail environment compared to advertising campaigns.\nOriginality/value: Even though prior research has considered the concepts studied here, this study aims to empirically evaluate a variety of antecedent factors that potentially affect purchase intentions. Relationships are established utilizing data collected in South Korea (an increasingly important consumer market) which adds value to extant knowledge in this area.\n",
"role": "assistant"
}
] |
Interaction, engagement, and perceived interactivity in single-handed interaction
|
[
"Behaviour",
"Cognitive mapping",
"Mobile communications",
"Human-computer interaction"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Single-handed thumb interaction has widely been used for mobile devices with touch screens. This is because users generally prefer one- to two-handed interactions when operating mobile devices (Park and Han, 2010). Using smartphones in a mobile context imposes both physical and intellectual demands on users. Single-handed interaction can offer multiple benefits to the users by freeing one hand from such demands. Previous research (Karlson et al., 2006) has consistently shown that single-handed interaction should be seriously considered when exploring the usability of mobile devices.\n2. Literature review: 2.1 Interactivity as a technological feature\n3. Research questions: With the aforementioned hypotheses in place, this study explores the mediatory effects on the interaction technique and behavioral variables. As single-handed thumb interaction is widely used, a number of researchers have wondered what triggers user motivation and how such behaviors are formed and enhanced. A group of researchers (e.g. Karlson et al., 2006) argued that the usability of mobile devices is heavily dependent upon single-handed interaction. As mobile devices became smarter and more sophisticated, single-handed interaction and its effects have also become more complicated. Understanding the cognitive mechanisms of users in single-handed interaction requires further study - for example, a study encompassing the multiple mediation model. A mediation model explores the direct and indirect effects between an independent variable and a dependent variable with inclusion of mediators. Mediation analysis has recently been used to dissect the direct and indirect effects of interaction variables on the attitudes and behaviors toward new technologies (e.g. Shin, 2013). This method is particularly effective when seeking to identify and explicate the mechanisms or processes underlying an observed relationship between an independent and dependent variable via the inclusion of a mediating variable. The mediator serves to clarify the nature of the relationship between the independent and dependent variables. Understanding this relationship can provide insight into user experience in single-handed interaction. From there, better heuristics can be used for the design of user-centered mobile interactions in the future. In this light, three mediation models for engagement, attitude, and behavioral intention, according to the range of thumb movement, are proposed and tested herein. The following research questions guide this study:\n4. Research hypotheses: With the exception of a few studies, such as those by Ng et al. (2014) and Faizuddin et al. (2014), very few of the interaction studies have examined the interaction effects in a mobile context. Indeed, this lack of research is even more conspicuous in smartphones, where user interaction remains under-researched and thus unanswered. Furthermore, the single-handed studies carried out in computer sciences and ergonomics have not comprehensively examined the cognitive, attitudinal, and behavioral effects of mobile interactions. Lepicard and Vigouroux (2010) conducted experiments with a tabletop touch panel interface and found that the direct interaction metaphor was easy to understand, and had a pleasing effect that attracted and motivated the elderly participants in their study. Similarly, Blasko and Feiner (2004) examined the relation between single-handed interaction techniques and user acceptance. Despite the numerous empirical studies on interaction techniques (e.g. Blasko and Feiner, 2004; Caprani et al., 2010), previous studies have not specifically addressed the comparison of swiping and tapping. Only few studies have compared swiping and tapping (e.g. Rantala et al.'s (2013) study which explored touch gesture vs finger touch). Thus, it is worthwhile to compare the effect of swiping and tapping on attitude. As implied in Rantala's (2013) study, it can be hypothesized as follow:\n5. Method: A 2 (interaction techniques: swiping vs tapping) x2 (range of thumb movement: wide vs narrow) between-subject experiment was conducted in order to explore how the touch-based interface on screens contributes to shaping the perceptions of smartphone users (e.g. perceived interactivity, engagement, attitude, and behavioral intention).\n6. Results: RQ1 and the multiple mediating effects were measured with the PROCESS macro in SPSS. For testing of the main effects (RQ2 and RQ3), the independent t-test and two-way ANOVA functions of SPSS were used. Where necessary, SAS was used for double-checking and additional data tests were used for reliability and validity.\n7. Discussion: In an exploratory manner, this study examined the effects of interaction techniques and the range of thumb movement on interactivity, engagement, attitude, and behavioral intention in single-handed interaction with smartphones. Despite the rising popularity of smartphones and the interactions involved, this topic has not yet been thoroughly researched. Thus, the goal of this study was to contribute knowledge in the form of empirically backed design guidelines and interaction techniques for improving the one-handed usability and operation of mobile devices, with particular emphasis on smartphones. With the exception of a few previous studies (Faizuddin et al., 2014; Negulescu et al., 2012), this study was one of the few attempts to test the effects of interactivity for identification of a useful predictor when interacting with smartphones. The findings showed that the range of thumb movement had a critical impact on the cognitive, attitudinal, and behavioral experiences of users during single-handed interactions, whereas the interaction techniques tested (e.g. tapping and swiping) did not have a significant impact. This implies that range of thumb movement becomes more important than interaction techniques. It can be seen that thumb movement can be a key factor as it increase the feeling of users' control. While the significant effect of thumb movement was expected, the insignificant results regarding tapping and swiping contradict the results of previous studies (e.g. Lai and Zhang, 2014). Though this is a rather unexpected result, it can be interpreted in a different way. The insignificant impact in itself has heuristic implications. With the development of numerous interaction techniques, such techniques themselves have become commoditized to the users. Regardless of the advancement and diversity of interaction techniques, users may have similar feelings and experience with different kinds of interaction modalities. In other words, interaction modalities, like tapping and swiping, are of little difference to users, and they might take such techniques for granted as part of embodied interaction. Therefore, it seems that interaction modalities, as features, make little difference to users. What is more important than tapping or swiping is the users' own activity and experiences that involve their actual behavior. This interpretation is in line with the significant relationship observed between thumb movement and engagement. Interactivity is greatly dependent upon users' perceived interactivity, which is highly related to engagement. Actual interactivity can only offer the potential to allow interaction; however, if users are not using or do not fully appreciate the interactive features, then perceived interactivity is low. This study showed that perceived interactivity can be enhanced by user-engaged interaction. This is an insightful heuristic implication for HCI in general. Although previous studies have researched the relationship between interactive features and perception of interactivity, the relationships so far identified have not been conclusive. With the changing nature of interactivity and the advancement of features, examination of smartphones suggests some possible relationships. Despite the belief that perceived interactivity lies at the heart of various interactions between users and technologies, users and people, this topic has been given little attention in the context of smartphones. The present research discussed the positive impact of the level of feature-based interactivity and actual user-engaged interaction on the perceived interactivity of smartphones. Feature-based interactivity, actual user-engaged interaction, and perceived interactivity are all related in a sequence with independent and multivariate random composition. Once this sequence of logic is empirically clarified, we will be able to really see the ensemble of actual and perceived interactivity. Academic researchers should further explore this relationship by potentially looking at other aspects of interaction techniques, while smartphone designers should focus on making smartphone interactivity a user-based interaction influenced by engagement. The relationship between actual and perceived interactivity in smartphones should be further pursued.\n8. Implications: From the findings, implications can be drawn in terms of practice and theory. As to practical implication, users' actual experiences with smartphones of course are essential and, at the same time, users' perceptions are also critical as they form their attitude and intention. From this perceptual point of view, the mobile phone industry ought to focus on improving users' perception of smartphone interaction. Improved user perception on interaction itself increases behavioral intention. While users continue to use, users may become accustomed to interaction modality and features. This can be a virtuous cycle of perception and behavior. The smartphone industry ought to develop a new design practice to form this virtuous cycle that closely link user experience and cognitive perception.\n9. Limitations and future studies: Despite the meaningful findings of this study, it bears several weaknesses that should be addressed in future studies. First, there was a lack of explanation for the non-significant effect of interaction techniques in single-handed interactions with smartphones. Previous literature suggested that in evaluating usability, the relationships between the interface and the real world should be considered. This refers to natural mapping (Tamborini and Bowman, 2010) or intuitiveness (Turner, 2008). Future study should consider the effects of natural mapping with interaction techniques to explore the interaction effects of mobile devices.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the effects of interaction techniques (e.g. swiping and tapping) and the range of thumb movement on interactivity, engagement, attitude, and behavioral intention in single-handed interaction with smartphones.\nDesign/methodology/approach: A 2x2 between-participant experiment (technological features: swiping and tappingxrange of thumb movement: wide and narrow) was conducted to study the effects of interaction techniques and thumb movement ranges.\nFindings: The results showed that the range of thumb movement had significant effects on perceived interactivity, engagement, attitude, and behavioral intention, whereas no effects were observed for interaction techniques. A narrow range of thumb movement had more influence on the interactivity outcomes in comparison to a wide range of thumb movement.\nPractical implications: While the subject of actual and perceived interactivity has been discussed, the issue has not been applied to smartphone. Based on the research results, the mobile industry may come up with a design strategy that balances feature- and perception-based interactivity.\nOriginality/value: This study adopted the perspective of the hybrid definition of interactivity, which includes both actual and perceived interactivity. Interactivity effect outcomes mediated by perceived interactivity.\n",
"role": "assistant"
}
] |
A study using a graphical syntax for actor-network theory
|
[
"Conceptual model",
"Health information systems",
"Medical record systems",
"Care delivery",
"Research methodology",
"Actor-network theory"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. The adoption of actor-network theory (ANT) in information systems (IS) research: ANT, which originated in sociology and anthropology, denies any difference between human and non-human entities (actors) at an ontological level and explains that alliances between actors change, are bilateral and are unpredictable. Technology, the organisation, groups, individuals, culture, policy documents and other artefacts are all actors and are intertwined in constantly changing relationships (Mitev, 2009). This has as a logical outcome the rejection of all of the following: techno-deterministic views; beliefs that the human actors fully control non-human actors; and assumptions that non-human actors are neutral. Thus ANT proposes a fundamental change in thinking and a new way of looking at groups of actors and their relationships. Indeed, as Hanseth et al. (2004, p. 119) note, \"[a] central idea and motivation behind ANT is to study the construction of things normally taken for granted. The border between social and the technical is one such\".\n2. Literature review: This section identifies a number of criticisms and challenges the IS researchers face when using ANT. The first is a general criticism regarding insight and quality. Concerns have been raised regarding the quality of research using ANT, which on occasion is judged to be mechanistic and instrumentalist, to be selective regarding aspects used and even to misinterpret the theory (Stanforth, 2007; Mitev, 2009; Hanseth et al., 2004).\n3. A graphical syntax for ANT: Actor-networks change continuously; therefore two different types of diagrams are required. The first set of diagrams, which we will call the ANT models, all have the same form. Each of these depict the actor-network of a particular case at a particular time and there will usually be several such diagrams, allowing us to see how the actor-network changes over time. Tsohou et al. (2012) refer to their own rich pictures of an actor-network at a particular moment as freeze frames. In the subsection that follows the individual symbols making up the ANT-gs syntax are introduced and reasons are given for differentiating between these ANT concepts by giving them separate symbols. The second diagram, encounter-episode frameworks that we will use is a time line that simply locates the ANT models in a time sequence. The framework and lexicon of ANT-gs are intended for use at the analytical phases of a research project.\n4. Research goals, methodology and case: The purpose of this paper is to demonstrate how actor-networks can be presented in a visual way through a graphical syntax and how this assists analysis. Hence the case studied focuses on the use of ANT-gs by IS researchers to depict and analyse the example. Having done so we evaluate ANT-gs in terms of how successfully it addresses the analytical challenges faced by IS researchers using ANT as a research method.\n5. ANT-gs models of various actor-networks over the life of the EHRS: 5.1 Encounters and episodes\n6. Evaluation: 6.1 Strengths\n7. Conclusion: The graphical syntax is based on a full set of ANT concepts and can be seen as new knowledge as it expresses some concepts in a different way, but it is not an alternative for ANT, it is an alternative format for expressing ANT. The syntax highlights translation, black boxes and actors at a distance and is believed to be unique in this respect. This makes it particularly useful as an educational tool, as examples illustrating key ANT concepts and using the notation can be created, the provision of the notation highlights the importance of the concepts, and students' work can be evaluated by looking specifically at how the concepts are used.\n",
"role": "user"
},
{
"content": "Purpose: Actor-network theory (ANT) is considered to be both a theoretical position and a methodology. ANT has been the centre of vigorous debates regarding its ontological viewpoint but has also been found to have some drawbacks as a methodology. The purpose of this paper is to use a graphical syntax for ANT to describe the development and implementation of a health information system (IS) in order to assess whether the graphical syntax improves the use of ANT as a methodology.\nDesign/methodology/approach: An extensive example derived from a case study, describing the development and implementation of an electronic patients' records system, is illustrated using this graphical syntax. This serves to makes the actors, translations and black boxes in the case visible.\nFindings: The syntax is found to help the researcher to conceptualise the research, to highlight assumptions as black boxes and to follow the actor but above all it encourages the researcher to understand the translations being made between actors and to make them explicit. Hence the syntax is found to assist at the analytical phases of a research project. The graphic syntax found to address the criticisms identified for ANT as an IS research methodology.\nResearch limitations/implications: The proposed graphical syntax assists researchers who use ANT as a methodology. Use of the syntax in education is also foreseen to be practical. The evaluation of ANT-gs in the paper provides an argument for its use but a more reliable argument would be made by collating and analysing feedback from independent modellers who actually use the graphical syntax. This is an important part of future research.\nPractical implications: The visualisation of the actor-network provided by ANT-gs provides a relatively simple representation while at the same time it makes key ANT concepts explicit. This is expected to address the issue of superficial understanding of ANT and selective use of its constructs; it makes the actor-network boundaries immediately visible. Thus the paper proposes that ANT-gs will be useful both by IS researchers and as an educational tool but future research is required to verify both these practical implications. The development of a computer-based modelling tool based on this syntax is proposed to facilitate effective and efficient modelling.\nOriginality/value: This paper proposes a unique tool to support ANT as a methodology. Its use encourages the researcher to focus on constructs that are essential elements of ANT and, by making each of the translation instances explicit, it strengthens the analysis in a way that is true to the view of ANT as a sociology of translation.\n",
"role": "assistant"
}
] |
The adoption of soft skills in supply chain and understanding their current role in supply chain management skills agenda: A UK perspective
|
[
"United Kingdom",
"Supply chain management",
"Soft skills",
"Supply chain skills"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nBackground: The literature on supply chain skills indicates the need for a balance of both hard skills (functional) as well as soft skills (relational) in order to manage global supply chains (Cottrill, 2010; Ellinger and Ellinger, 2014; Murphy and Poist, 2007; Jordan and Bak, 2016; Gammelgaard and Larson, 2001). Dubey and Gunasekaran (2015, p. 89) also note that \"hard skills combined with soft skills, definitely offer a competitive edge to managers\". In the last 20 years, there has been a growing emphasis towards \"softer\" aspects of supply chain skills (Ellinger and Ellinger, 2014; Murphy and Poist, 2007), whereby soft skills have been defined as abilities and traits that pertain to personality, attitude and behaviour rather than to formal or technical knowledge (Moss and Tilly, 1996). However, the tendency in the supply chain research agenda was driven towards hard skills, as it has been regarded more attributable to employers' needs (Balcar, 2016; Bak and Boulocher-Passet, 2013; Tadesco et al., 2014). Hence, in the context of supply chain skills, hard skills have been interlinked with functional and management skills, such as inventory management, transport management and logistics service management (Derwik and Hellstrom, 2017; Sodhi et al., 2008), which has entailed a level of \"technical expertise and knowledge needed for a job\" (Robles, 2012, p. 453), which were far easier for firms to measure (Balcar, 2016). However, a recent study conducted by Groysberg (2014) found that skill requirements that were once primarily aimed at sourcing functional and hard skills have now become a standard requirement, whilst key decisions about hiring tend to focus on the interpersonal and relational skills. Similarly, a report by Langley and Capgemini Consulting (2015) supported the fact that future supply chain skills will need to include soft skills to supplement hard skills, and another survey found that 72 per cent of CEOs felt that soft skills are more important for their business in the current environment (The Economist, 2016). Although the importance of soft skills are on the rise, there is limited research, which may be due to the fact that \"the measurement of soft skills is difficult as there is no objective way to test the skill itself as opposed to hard skills\" and hence soft skills requirements and its measurement could show variance across companies' needs (Balcar, 2016, p. 454).\nDefining the role of soft skills: a supply chain perspective: Traditionally, the definition of hard and soft skills was based on the nature of skills with hard skills associated with technical and administrative categories, and soft skills corresponding to the skills in the human, conceptual, leadership and interpersonal categories (Weber et al., 2009). Within this context, soft skills are defined as the skills related to the acquisition of interpersonal, human, people or behavioural skills in the workplace (Rainsbury et al., 2002). Within the supply chain context, a combination of hard (functional and managerial) and soft (relational and behavioural) skills is required to manage supply chains effectively (Anon, 2014). According to Russell (2016), supply chains in 2020 will require interpersonal, communication, collaboration and sales skills, which will help integrate supply chains. Giunipero and Pearcy (2000) as well as Jordan and Bak (2016) grouped soft skills in supply chains under: decision-making skills, management skills, behavioural skills, but also included a new direction highlighting the role of negotiation skills as derived skill sets from the literature.\nMethodology: This study utilises a mixed-method research which include the\"[...] collection and analysis of both qualitative and quantitative data in a single study in which the data are collected concurrently or sequentially [which] involve the integration of data at one or more stages in the process of research\" (Creswell et al., 2003, p. 212). A mixed-method study is especially a good choice when the study requires \"the researcher to simultaneously answer confirmatory and exploratory questions, and therefore verify and generate theory in the same study\" (Teddlie and Tashakkori, 2003, p. 15).\nFindings and analysis: The research has provided an insight to supply chain soft skills based on four categories based on decision-making skills, behavioural skills, management skills and negotiation skills with a set of 15 soft skills that has been addressed in the supply chain literature. The results suggest that soft skills, especially behavioural skills such as communication, planning, initiative and negotiation were seen to be more important when compared to decision making, negotiation and management skills. The findings indicate that the changing supply chain scope encourages the requisition and development of different supply chain soft skills with varied levels of emphasis in relation to 15 soft skills identified in the literature.\nConclusion and research implications: This study began with a discussion of soft skills in the supply chain arena, identified in the academic literature. The discussion surrounding supply chain soft skills needs is not new, although the emphasis on specific soft skills areas has evolved over time based on industry needs (Bak and Boulocher-Passet, 2013). However, ongoing discussions have omitted the soft skills impact and development, whereas hard skills have been widely explored and acknowledged (Bowersox, 2002; Kirby, 2003; Mangan and Christopher, 2005). Supply chains are increasingly people-driven and are highly dependent on the specific skills, and the strength of these skills, for their overall success. When investigating the soft skills, our study indicated several findings. First, the changing competitive global environment indicated the increasing need for supply chain soft skills with an emphasis placed on behavioural skills, decision-making skills and management skills as critical in soft skill discussions in the UK. The results suggest that soft skills, especially behavioural skills such as communication, planning, initiative and negotiation, were seen to be more important when compared to decision making, negotiation and management skills. The findings indicate that the changing supply chain scope encourages the requisition and development of different supply chain soft skills with varied levels of emphasis in relation to 15 soft skills identified in the literature.\nLimitations and directions for future research: The current study has examined the supply chain soft skills, moreover, the four categories of soft skills adoption. Despite the interesting findings of the study, there are certain limitations that need to be addressed at this stage, which revolves around four main categories. The first limitation is the limitation of context whereby the results are derived from a UK-based sample, which may limit the generalisability of our findings for other contextual settings as the soft skills involving country-level variations due to variances driven by culture, values, politics, management styles, as stated by Al Mehrzi and Singh (2016). Such individual differences maybe enhance the understanding of soft skills and differences in supply chains across borders. Hence, the soft skills in current study may be enhanced by additional factors, which can be explored in further studies. Another limitation relates to the time limit, as a longitudinal study may have been also beneficial to examine the development of supply chain soft skills, in order to understand the impact of change over time within an organisation as increased demands on employees have been placed constantly to develop their workplace skills. For example, in organisation literature, Pradhan et al. (2017) and Almatrooshi et al. (2016) identified emotional intelligence as a component for adaptability to organisational change and can be a factor in driving positive behaviour; hence, it would be interesting to evaluate the emotional intelligence as an attribute to soft skills in the supply chain context. These limitations of the current study can be assumed as the bases for carrying out further research on supply chain soft skills, providing future research avenues which may expand and test our findings in new contextual settings.\n",
"role": "user"
},
{
"content": "Purpose: With supply chains expanding in scope and scale globally, the academic literature underlined the increasing role and importance of soft skills. Traditionally, the supply chain literature geared towards hard skills including functional and technical skill sets with limited discussion on soft skills. Therefore, the purpose of this paper is to assess and explore the soft skills demand in supply chain management arena.\nDesign/methodology/approach: This study has utilised a mixed methods study in two phases, with the first stage including a questionnaire distributed to 120 supply chain employees in the UK, followed by six interviews with supply chain experts in the UK.\nFindings: The results suggest that soft skills, especially behavioural skills such as communication, planning, initiative and negotiation, were seen to be more important when compared to decision making, negotiation and management skills. The findings indicate that the changing supply chain scope encourages the requisition and development of different supply chain soft skills with varied levels of emphasis in relation to 15 soft skills identified in the literature.\nResearch limitations/implications: This study employs a mixed-method approach to establish the perceived importance of soft skills in the UK supply chains. This limits the generalisability of the results to other contextual settings.\nPractical implications: This paper presents soft skills impact upon the supply chain. Specific soft skills are critical to supply chain employees compared to others (e.g. behavioural and people management skills), which may lead to articulation of supply chain soft skills training initiatives.\nOriginality/value: This paper contributes to the soft skills discussion in the supply chain context and discusses the role of soft skills. Topical gaps in the literature are identified as areas for future research. The findings have generated additional supply chain skills to the academic literature as well as provided an understanding of the weighting of soft skills in terms of their importance and application to industry needs.\n",
"role": "assistant"
}
] |
Subordinates' helping, voice, and supervisors' evaluation of job performance: The moderating effects of supervisor-attributed motives
|
[
"Helping",
"Job performance",
"Motives",
"Attribution",
"Voice"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Since Bateman and Organ (1983) introduced the concept of organizational citizenship behavior (OCB), a great deal of research has examined its antecedents, consequences, and various dimensions (Podsakoff et al., 2000). Organ (1988, p. 4) defined OCB as \"individual behavior that is discretionary, not directly or explicitly recognized by the formal reward system, and that in the aggregate promotes the effective functioning of the organization.\" Based on this definition, which emphasizes individual voluntary behaviors for organizational effectiveness, many researchers have examined whether OCB can really contribute to the overall success of an organization, such as improved sales (MacKenzie et al., 1991) and customer satisfaction (Walz and Niehoff, 2000).\nInfluences of helping and voice on job performance: It is well recognized that task performance - which refers to the completion of duties and tasks that contribute to the organization's technical needs or the fulfillment of prescribed job duties (Rotundo and Sackett, 2002; Williams and Anderson, 1991) - has a stronger influence on job performance than OCB (Rotundo and Sackett, 2002). However, it is also apparent that OCB has a positive influence on supervisors' evaluations of subordinates' overall job performance that extends beyond the influence of task performance. In a related study, Motowidlo and Van Scotter (1994) found that both task performance and OCB made unique contributions to supervisors' job performance ratings. Consistent with previous findings, we expected that subordinates' greater engagement in helping and voice behavior would be positively related to their supervisors' ratings of their job performance; the theoretical basis for this hypothesis can be explained in two ways.\nProsocial and impression management motives in OCB: It has long been assumed that employees engage in OCB with pure and/or altruistic motives (Bolino, 1999; Organ, 1988). That is, previous researchers have posited that employees exhibit OCB in exchange for good relationships with their organizations and/or job satisfaction (Bolino et al., 2004). In this respect, from a positive perspective, OCB has been considered analogue to altruistic, other-oriented, prosocial behavior (Grant and Mayer, 2009). For this reason, many studies have attempted to investigate OCB predictors such as individuals' affective states (e.g. job satisfaction or organizational commitment) and contextual factors (e.g. supportive leadership) (Organ and Ryan, 1995; Podsakoff et al., 2000).\nModerating effects of supervisor-attributed motives: Besides the importance of motivations to engage in helping and voice, how supervisors perceive subordinates' motives to help others or be a voice for the organization is a key factor in their performance evaluation process. According to social information searching theory (Feldman, 1981), supervisors tend to search for information that will confirm their expectations of their subordinates' performances when conducting evaluation. However, due to the difficulties involved in gathering relevant information, supervisors can evaluate subordinates' behaviors by inferring their motives for those behaviors (Green and Mitchell, 1979).\nMethod: Sample and procedures\nResults: Validity analysis\nDiscussion: Although there is a prevailing assumption that OCB has positive consequences for employees' job performance evaluation (e.g. Podsakoff et al., 2009), there have not been enough attempts to understand when it is instrumental in making supervisors' performance evaluations. To fill this research gap, we used attribution theory to examine the moderating effects of supervisor-attributed prosocial and impression management motives in the relationship between helping, voice, and job performance.\n",
"role": "user"
},
{
"content": "Purpose: It is recognized that employees' helping and voice behaviors are dimensions of organizational citizenship behavior used by supervisors to evaluate their job performance. However, existing empirical studies of these relationships have shown inconsistent findings. From the perspective of attributional theory, the purpose of this paper is to explain when subordinates' helping and voice behaviors are more positively related to job performance by considering supervisor-attributed prosocial and impression management motives.\nDesign/methodology/approach: Using a sample of 200 supervisors in South Korea, the authors tested the hypotheses with hierarchical multiple regression analyses.\nFindings: Results indicate that the positive effects of helping and voice behaviors on job performance were stronger when supervisors attributed such behaviors as driven less by impression management motives related to self-interest. However, contrary to the expectations, the positive influences of helping and voice behaviors on job performance were stronger when supervisors perceived low prosocial motives.\nPractical implications: Findings suggest that supervisors need to avoid making the wrong attributions with regard to their subordinates' helping and voice behaviors during the evaluation process. In addition, subordinates need to have clear motives and demonstrate consistent behavioral stances when engaging in such behaviors.\nOriginality/value: Using social information theory and attribution theory, this study contributes to explain when helping and voice behaviors improve evaluations of employees' job performances by considering supervisor-attributed motives.\n",
"role": "assistant"
}
] |
A systematic approach for public-private partnership projects CSFs evaluation
|
[
"Management",
"Case study",
"Engineering",
"Approach"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: A public-private partnership (PPP) is defined as a long-term contractual arrangement between a public sector agency and a private sector concern, whereby resources and risk are shared for the purpose of developing a public facility (Akintoye et al., 2003). Many researchers believe that PPP can release a governments' tight budgetary pressure by injecting the private sector's resources, encouraging innovation, enhancing productivity, allowing better risk allocation and improving cost effectiveness (Ng et al., 2012). Given the growing interest in PPP since the late 1990s, PPP projects have been widely used in many developed and developing counties to procure infrastructure assets for the provision of essential public services (Tang et al., 2010; Love et al., 2015). For instance, according to public data, at the end of year 2017, the National PPP Comprehensive Information Platform established by the Ministry of Finance of the People's Republic of China showed 14,424 projects planning to employ the PPP model to attract investments, with a total investment amount of RMB13.518.2tn (National PPP Integrated Information Platform Project Library, 2017). There is no doubt that PPP will continue to be of great interest to researchers in the future as the PPP market keeps growing and maturing around the world (Chan et al., 2010c; Osei-Kyei and Chan, 2015).\n2. Identification of CSFs: 2.1 Selection principles\n3. Methodology: 3.1 Preliminaries\n4. Case study: 4.1 Model application\n5. Conclusions: The CSFs of PPP projects are a growing concern. This paper developed an evaluation model to explore the interrelationships of various related dimensions and factors and, then, to determine the impact of CSFs by considering their interactions. Through an empirical study, the driven-dependency relation and internal impacts among the dimensions and the impacts of the CSFs were successfully investigated, indicating that this method can help project participants to better understand the CSFs and further improve project performance.\n6. Limitation: It is important to acknowledge the limitations of this research.\n",
"role": "user"
},
{
"content": "Purpose: Public-private partnership (PPP) projects are increasingly significant in many countries. The purpose of this paper is to assess the impact of critical success factors (CSFs) on PPP projects and comprehensively consider the interrelations and interaction among dimensions and factors to achieve a better understanding of PPP project management.\nDesign/methodology/approach: An evaluation index system for PPP projects such as the presented case study is proposed based on a literature review and a survey. Then, interpretative structural modeling is used to transform the CSFs dimension into a multi-level hierarchical model to reflect the driven-dependency relation of each dimension; the fuzzy analytic network process model optimized by moment estimation theory is used to investigate the impact of CSFs by considering their internal impact.\nFindings: Regarding the project used as the case study, the driving force and dependence for driving layer and dependent layer are determined. Moreover, in driving layer, efficient and well-structured payment mechanism is the most important CSF if considering the internship and interaction among CSFs, and efficient and well-structured payment mechanism and good governance provide most positive interaction; in dependent layer, population of beneficiaries is the most important CSF if considering the internship and interaction among CSFs, and public client's satisfaction provides most positive interaction.\nOriginality/value: This paper developed an evaluation model to explore the interrelationships of dimensions and factors and then determine the impact of CSFs. The model propose in this paper relaxes the independence assumptions of traditional methods and is more in line with reality; besides, weighting method is optimized to obtain more objective and reasonable evaluation results. Through an empirical study, the validity of the model has been verified; therefore, the study can help project stakeholders better understand the CSFs and further improve project performance.\n",
"role": "assistant"
}
] |
Frontline employees' high-performance work practices, trust in supervisor, job-embeddedness and turnover intentions in hospitality industry
|
[
"Frontline employees",
"Turnover intentions",
"High-performance work practices",
"Job-embeddedness",
"Trust in supervisor",
"Voluntary turnover"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The acquisition and retention of high-performing employees is critical for organizational success. The costs of employee turnover are high, both in terms of finances related to rehiring and training of new employees as well as those experienced employees lost to potential competitors. A study by the Hay Group revealed that replacement costs are about 50-60 per cent of an employee's annual salary in hospitality industry (Mitchell et al., 2001a). Therefore, every organization aims to retain their well-functioning frontline employees for as long as possible. Thus, the question of concern for the academicians and practitioners alike is \"Why do employees leave?\" and \"Why do they stay?\". This conundrum is well illustrated by the quote of Coleman (2001) that \"Employees don't leave companies, they leave managers and supervisors\".\nTheoretical background and hypotheses development: The current study selected three key indicators of HPWPs, namely, empowerment, training and reward. There are two justifications for this. First, according to Tang and Tang (2012), empowerment, training and reward are among the most effective HRM practices, especially for the hospitality industry (Kusluvan et al., 2010; Enz and Siguaw, 2000; Cho et al., 2006). Kotler et al. (2006) found these management practices to be critical for retaining employees at Ritz-Carlton. The simultaneous emphasis on empowerment, training and reward indicates that management is highly committed to service quality (Karatepe, 2013b; Daskin et al., 2015). Second, despite being critical management practices, the effects of empowerment, training and reward on OTJE have been minimally tested empirically (Bergiel et al., 2009). Therefore, this study would examine the most important indicators of HPWPs and their joint effects on OTJE.\nMethod: Sample and procedure\nResults: A partial least squares (PLS) path model was used to analyze the data. PLS is considered as the best method when both formative and reflective indicators are to be analyzed (Haenlein and Kaplan, 2004) as is the case in the present study. We employed SmartPLS 2.0 and 3.0 (Ringle et al., 2005) to test H1-H5. Logistic regression using SPSS 18 was used to test H8, because voluntary-turnover as the dependent variable was dichotomous (resign = 0, active = 1). The reliability of the measurement model was assessed in the first stage of a PLS model. HPWP is conceptualized as a second-order formative construct as Empowerment, Training and Rewards were treated as first-order reflective dimensions. Likewise, TIS is operationalized as a second-order formative construct with two first-order reflective dimensions' cognitive and affective trust. Finally, OTJE is also conceptualized as a first-order formative construct, and as a consequence, reliability and validity analysis are not relevant for OTJE, HPWPs and TIS (Hulland and Business, 1999).\nDiscussion and conclusions: Conclusions\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to investigate the mediating effect of job embeddedness on the relationships between high-performance work practices, trust in supervisor and turnover intentions of frontline employees in the hospitality industry.\nDesign/methodology/approach: Data were collected from 343 frontline employees working in four- and five-star hotels of Thailand. Partial least squares was used for analysis because it is considered as the best method to analyze the data containing both reflective and formative indicators.\nFindings: Results suggest that job embeddedness fully mediates the effects of high-performance work practices and trust in supervisor on turnover intentions and turnover intention positively affects the actual voluntary turnover.\nPractical implications: The study confirms that high-performance work practices (empowerment, training and rewards) and trust in supervisor affect turnover intentions through on-the-job embeddedness. Hence, high-performance work practices embed hotel employees in their jobs, and they are unlikely to display turnover intentions. Furthermore, low level of trust in supervisor must be addressed to maintain a healthy environment where employees are able to develop their job embeddedness.\nOriginality/value: This study contributes to the body of research on the theoretical explanation of the consequences of trust in supervisor in hospitality industry, as well as to the growing body of research on turnover intentions in frontline employees.\n",
"role": "assistant"
}
] |
The impact and interplay of national and global CSR discourses: insights from France and Brazil
|
[
"Social responsibility",
"France",
"Brazil"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Throughout the world, the relations between business and society are undergoing profound changes, leading to a redefinition of the role and responsibilities of companies. On the one hand, our societies are facing ever more complex economic, social and environmental challenges, such as climate change, threats on biodiversity, increasing unemployment, poverty, urbanization or demographic change. These challenges can hardly be solved without the active contribution of the business sector, which has acquired new economic and political powers over the last decades and which often develops interesting technical and organizational innovations. On the other hand, the role of public authorities is increasingly weakened. National governments are grappling with tight budgetary constraints, while international regulations have difficulties to emerge or even regress in a context of global competition between national systems (Berthoin Antal et al., 2009).\n1. The impact of the national contexts on the CSR discourses and practices: Compared with the USA, the discourses and practices in the area of CSR seem to be more recent in France and Brazil. France is often considered as an example of the countries in which CSR has for a long time been implicit (Matten and Moon, 2008), since the integration of economic, social and environmental responsibilities by companies was part of mandatory legislation. This explains that academic research has only recently started to pay an increased attention to the practices developed by French companies and other stakeholder groups in this area (Beaujolin and Capron, 2005; Berthoin Antal and Sobczak, 2007). As for Brazil, its companies had for a long time been less submitted to the pressure of stakeholder groups than those of other countries, in particular in the periods when the government was in the hands of the military. It was only quite recently that Brazilian companies have started at a broader level to develop more structured discourses and practices on CSR to reply to the expectations of international clients or, increasingly, to local consumers and NGOs that are more attentive to the economic, social and environmental challenges (Schmidheiny, 2006).When the CSR discourses and practices finally became more explicit and structured in France and Brazil over the last decades, the primary focus in both countries was the social dimension rather than the environmental one, much more developed in the Nordic countries, even if the protection of the environment has since then also gained more attention from companies and other stakeholder groups in the two countries. In France, the CSR discourses and practices traditionally focused internally on labour-related issues rather than on external stakeholders (Berthoin Antal and Sobczak, 2007). Even if this situation has changed over the last years with the development of activities in the field of environmental protection or in favour of the community, this focus may explain the fact that French workers' representatives continue to be much more involved in companies' CSR policies than in other parts of the world. French companies have for example been pioneers and continue to be among the most active ones in the negotiation of transnational collective agreements on CSR with global union federations (Sobczak, 2007). In Brazil, the CSR discourses and practices traditionally focused on the role of business in community projects, in particular in the field of education, highlighting thus the government's difficulties in this area.The analysis of the cultural and legal context in France and in Brazil reveals a number of differences, which have shaped the relations between business and society in the two countries and thus the development of the CSR discourses and practices. A first difference concerns the link that is, or not, established with religious values. In France, the principle of secularity has constitutional value and explains that the Catholic Church is traditionally very discreet in economic affairs (Beaujolin and Capron, 2005). In France, the Church has thus not taken any particular initiative in the area of CSR. Some employers' associations that play an active role in the promotion of the CSR discourses and practices, such as the Centre of Young Leaders (Centre des Jeunes Dirigeants, CJD) are rather close to Christian values, but refuse to officially refer to them. In Brazil, by contrast, the Association of Christian Enterprise Directors (Associacao dos Dirigentes Cristaos de Empresas) has had a rather strong influence on the CSR discourse, by defining as early as by 1965 a set of principles for responsible management (Simoes, 2008). More generally, the Catholic Church has to a certain extent influenced the CSR discourses and practices in Brazil through its tradition of assistance towards the poor that may explain the companies' philanthropic ideas.A second difference is linked to the role of public authorities. Whereas France is an example of a country where the public authorities have deeply influenced the emergence and the development of CSR and its content through various legislation and other policies, the case of Brazil highlights on the contrary how the lack or the insufficiency of interventions by public authorities may lead to particularly important expectations among society towards business to fill in the gap. In other words, in both countries, public authorities have played an important role in shaping the CSR discourses and practices, but in two almost opposite ways. In France, the government has used its traditionally strong role in the economy to intervene directly by encouraging companies to define and implement more responsible strategies and by empowering other stakeholder groups. In Brazil, the government's influence in this area was much more indirect, insofar as it was mainly its incapacity to solve the country's social and environmental challenges that created a need for voluntary CSR initiatives.1.1 France: the active role of the public authorities in the development of CSR\n2. The impact of international initiatives on the CSR discourses and practices: While the different cultural and legal contexts continue to influence the CSR discourses and practices in France and Brazil, international actors contribute to a certain harmonization in the area. Multinational companies and their supply chains are a first category of actors that play an important role in this process, insofar as they favour the dissemination of CSR discourses and practices beyond national borders. In France, multinationals with headquarters in North America have played a leading role in the development of innovative practices aiming at promoting diversity and equal opportunities among the employees (Bellard and Ruling, 2001). In Brazil, foreign multinationals have often been among the pioneers in introducing the CSR discourses and practices in the country (Griesse, 2007). To a lesser extent, multinationals with their headquarters in France or Brazil may also have learnt from the CSR discourses and practices of their subsidiaries in other parts of the world (Berthoin Antal and Sobczak, 2007).International NGOs are a second category of actors that have largely contributed to change the traditional CSR discourses and practices in France and Brazil, be it through campaigns aiming at raising the awareness on social and environmental problems among the population and criticizing corporate behaviour, or by developing innovative partnerships with companies in the field of CSR and sustainable development (Chartier and Deleage, 1998). Finally, companies and other stakeholders in France and Brazil have been rather proactive in the integration of the emerging international standards and tools in the area of CSR. This has allowed them to evolve within a short period of time from a situation where their strategies and practices were mostly ignored at the global level to a situation where they actively contribute to the international debate, and even shape new standards, as we describe below.2.1 Development of partnerships between companies and NGOs\nConclusions: This article shows the impact of both national traditions and recent international influences on the CSR discourses and practices in France and in Brazil. While it highlights the tendency to a certain harmonization in the area through the various activities of multinational companies, international NGOs and standardization organizations, it also shows that cultural and legal traditions continue to play an important role in the definition of the relations between business and society. Managers in multinational companies thus have to pay attention when defining their CSR strategies to create a balance between the respect of certain universal values, in particular those integrated in the emerging international standards, and the need to adapt to the different national contexts of their subsidiaries or suppliers. In some countries, such as France, the respect of the national specificities is mandatory by law; in others, such as Brazil, local stakeholders expect it, in particular from multinational companies, which are perceived as having specific responsibilities reflecting their economic powers. In both cases, the best way to find the right balance consist of engaging with the local stakeholder groups that may help the managers to understand the context and to implement effective projects improving the company's positive impact on its natural, economic and social environment. However, more research has to be conducted on the conditions allowing the long-term success of such local partnerships between very different actors, as well as on the ways allowing the managers to learn from these local partnerships to improve their global CSR discourses and strategies.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to illustrate the interplay between national traditions and international influence on the CSR discourses and practices, in the context of two very different national contexts, namely France and Brazil.\nDesign/methodology/approach: - The paper is based on a literature review as well as on interviews with representatives from companies and other stakeholder groups in France and Brazil.\nFindings: - The paper highlights the impact of national traditions, such as the strong intervention of the State in France and the lack of monitoring legislation by public authorities in Brazil, on the CSR discourses and practices in both countries. But it also shows that multinational companies, international NGOs as well as international standardization organizations lead to a certain homogenisation of the CSR discourses and practices.\nResearch limitations/implications: - This exploratory research should be complemented in the future by more in-depth research, for example by comparing the CSR discourses and practices of multinational companies having activities in both countries.\nPractical implications: - The paper highlights the importance for managers in companies having activities in different countries to adapt their CSR discourses and practices to the local contexts, while maintaining a coherent strategy at the international level based on the respect of some universal principles.\nOriginality/value: - Few researches have been conducted on the CSR discourses and practices in France and in Brazil. This paper highlights the main characters of the CSR discourses and practices in both countries, while targeting their common points and differences.\n",
"role": "assistant"
}
] |
HRM formality differences in Pakistani SMEs: a three-sector comparative study
|
[
"SMEs",
"Pakistan",
"HRM formality",
"HRM differences"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Global interlinking of economies has increased pressure for superior organizational performance (Kuruvilla and Ranganathan, 2010). Competitive pressures from institutional changes, technological advances and deregulation increasingly influence sustained growth (Campbell et al., 2012). Until recently, traditional sources (e.g. capital, technology, economies of scale) have been central to the acquisition of competitive advantage but these resources are becoming increasingly imitable (Boxall and Purcell, 2003; Lin and Wu, 2013). Based on resource-based view (RBV) theory (Barney, 1991), human capital can be a valuable source of sustained competitive advantage since the specific resources derived from it are difficult for other firms to imitate due to specialization, scarcity and tacit knowledge (Campbell et al., 2012; Wright and McMahan, 2011).The institutional factors that shape HRM practices that link to sustained competitive advantage have attracted previous attention (e.g. Wood and Lane, 2012; Wu et al., 2014). Guided by the institutional perspective (DiMaggio and Powell, 1983), several studies have found contextual factors (e.g. business sector, size of the firm) directly related to the adoption of formal/best HRM practices (Boselie et al., 2003; Wu et al., 2014). The comparative HRM literature treats the business sector as an influential determinant of HRM formality and best practices (Datta et al., 2005; Jiang, 2009) and argues that the adoption of HRM activities within different SME sectors is usually needs-based (Deshpande and Golhar, 1994; Jiang, 2009).Although large businesses play a crucial role in the economic growth of a country, the role of SMEs in stimulating and strengthening economic indicators carries equal importance (Ayuso and Navarrete-Baez, 2018; Dundon and Wilkinson, 2018). Small firms can be seen as \"growth engines\" that can make both social and economic contributions to the development of a country (Umer, 2012). The economy of Pakistan is also a direct reflection of its SME sector (Soomro et al., 2019) since SMEs represent more than 90% of the total established businesses (PBS, 2016). Despite its economic significance, the SME sector in Pakistan faces a variety of shortcomings, which limit its ability to fully contribute towards national economic progression. These include a lack of business information infrastructures, limited financial literacy and importantly the lack of a strategic approach towards human resource development (Iqbal and Malik, 2019; SBP, 2010).Human resource systems in Pakistani firms are currently going through a developing phase. Firms are beginning to rename their administration departments as HR/Personnel, but there is little HRM-related research in this context that helps to understand the effects of this change (Khan et al., 2014). Informal HRM practices are common across the majority of organizations that lack a systematic approach towards managing their human capital. Consequently, low motivation and high employee turnover are some of the common challenges faced (Ahmad and Allen, 2015).Given the important role of SMEs in strengthening a country's economy (Higgs and Hill, 2018), research on HRM in SMEs remains underrepresented as most research concerns large firms (Botero and Litchfield, 2013; Harney and Nolan, 2014; Wiesner et al., 2007). Since large organizations differ from SMEs in multiple aspects (e.g. business strategies, influence of institutional factors and availability of resources), their HRM practices are often not comparable to the SME context (Sheehan, 2014; Storey, 2002). Moreover, the literature lacks perspectives from SMEs from developing and transitional economies since less empirical work is conducted in contexts such as Pakistan (Iqbal et al., 2018; Soomro et al., 2019).The aim of this empirical study is to unfold the differences in HRM formality (including sub-components of recruitment, selection, training and development, performance appraisal, compensation and benefits) among three SME business sectors (services, manufacturing and trade) in Pakistan. The need for such exploration is inspired by some pertinent gaps in the comparative HRM literature. First, scant attention has been given to the comparative HRM literature in SMEs in terms of sector differences (Dickmann et al., 2008; Edwards and Ram, 2006; Psychogios et al., 2016). Second, the available literature lacks perspectives from trading sector SMEs since comparative studies have mainly focussed only on manufacturing and services. Given the important role of trade sector SMEs in terms of their contribution to national GDP and employment creation (World Trade Organisation, 2016), this study extends the boundaries of comparative HRM literature pertaining to sector differences by incorporating trade sector SMEs. Third, the comparative HRM literature does not take into account the effect of some contextual control variables (e.g. age of the firm) concerning differences in HRM practices across sectors. Given the influential role of certain organizational contextual variables, this study incorporates firm age as a control variable while exploring sector differences in HRM practices.The key contribution of the paper is to show that levels of HRM formality differ across the three sectors such that the services sector has adopted more formal HR practices than manufacturing and trade, whereas the manufacturing sector exhibits greater HRM formality than trade. Further, the influence of sector remains significant when firm age is controlled. Implications for the SME sector in Pakistan are proposed.\nDiversity in HRM formality among services, manufacturing and trade SMEs: Theoretical underpinnings\nMethods: Sample and data collection\nResults: HRM formality differences by sector\nDiscussion and conclusions: The aim of this study was to unfold differences among service, manufacturing and trade firms in terms of HRM formality and its underlying components (recruitment, selection, training and development, performance appraisal, compensation and benefits). The results suggest that service sector SMEs have adopted more formal HRM practices than manufacturing and trade sector SMEs in Pakistan. These findings corroborate other empirical studies that found that the nature of the industrial sector is an influential determinant of HRM practices in firms (Datta et al., 2005; Jiang, 2009; Psychogios et al., 2016). One explanation for this relates to the distinctive characteristics of industrial sectors in which SMEs operate. For instance, service firms are associated with intangible outputs and the involvement of customers in the production of services. Similarly, the need for more skilled people in services compared to other sectors means an increased focus on employee satisfaction in order to minimize labour turnover (Harney and Dundon, 2006). These findings challenge studies that have found no differences in the HRM practices of manufacturing and services sector firms (Guest et al., 2003; Deshpande and Golhar, 1994).The empirical evidence thus far concerning differences in HRM practices among manufacturing and trade sector SMEs is lacking, and this study has found that manufacturing firms employ more formal HRM practices than trade firms. Explanations lie in the varying characteristics of both sectors (Wu et al., 2014; Jiang, 2009) such as a lack of semi-skilled manpower and technological advancements in manufacturing (Tiwari and Saxena, 2012) and a high level of workforce attrition in the trade sector (Almas, 2014).The results also demonstrate that services SMEs in Pakistan have adopted more formal individual HRM practices than manufacturing and trade. Furthermore, manufacturing firms were found to be more formal than trade sector SMEs in terms of the adoption of individual HRM practices/functions. The literature concerning comparative HRM in SMEs reinforces the notion that differences in individual HRM practices (e.g. performance appraisal) among SMEs are explained by distinctions based on industrial sector (Raziq, 2011). For instance, Jackson and Schuler (1995) found that service sector employees experience more formal and systematic appraisal systems and as a result more formal compensation practices. They also observed that customers play a central role in appraisal systems in service sector firms as compared to firms from other sectors. Similarly, Bartman and Lindley (1995) and Raziq (2011) also found more formal recruitment and selection practices in services sector small firms than in manufacturing arguing that service-based firms are more dependent on a skilled workforce than manufacturing or trade. With regard to training and development, Duberley and Walley (1995) found that service sector firms follow more formal training procedures than manufacturing.No conclusions could be drawn on the effects of firm size due to the violation of certain statistical assumptions. However, for age of the firm as a covariate, service-based firms have adopted more formal HRM practices (HRM formality) in comparison with manufacturing and trade sector firms. Also, manufacturing firms followed more formal HRM practices than trade firms. Added to previous research (Storey et al., 2010), our results support firm age as an influential determinant of HRM formality across sectors. An important implication from this is that future researchers should incorporate business sector as a control variable in HRM studies (along with size and age), especially within SMEs.Within an institutional perspective, various dimensions (e.g. culture, competition, contextual organizational characteristics) have been studied to understand their influence on HRM practices (Hoque and Bacon, 2006). In line with the objectives and scope of this study, a comprehensive analysis of comparative HRM practices among three important SME sectors produced findings that are not only under-researched in general (Dickmann et al., 2008; Psychogios et al., 2016) but also unique in the context of Pakistan. Moreover, the scant comparative HRM literature concerning SMEs in Pakistan has only focussed on larger firms associated with manufacturing and services (Raziq, 2011) which lacks generalization. By focussing on sector differences in HRM formality while controlling for the effect of firm age, the study adds distinctive value to the comparative HRM literature since estimating such comparisons while controlling for influential contextual variables (e.g. size, age) is not only rare (Harney and Dundon, 2006) but also unique in the context of Pakistan.With regard to practical implications, the results suggest that industrial sector shapes HRM practices in Pakistan's SMEs. Also, the greater reliance of service sector SMEs on more formal HRM practices is reflected in the economic performance of Pakistan. The growth in the service sector has been more than that in any other sector and has facilitated an economic shift from commodity-producing to service sectors (Pakistan Economic Survey, 2017). Since the positive role of formal HRM practices in enhanced organizational performance through employee satisfaction, lower turnover and increased labour productivity has been established (Nguyen and Bryant, 2004; Paauwe et al., 2013; Zhou et al., 2013), owners/managers of manufacturing and trade SMEs can not only learn from this but also implement more structured HRM practices for superior organizational gains. Lastly, based on the findings, trade sector SMEs were found to be the least formal. Given the important role of trading SMEs in low-income countries in terms of promoting inclusive economic growth (World Trade Organisation, 2016), efforts are needed by SMEDA to prioritize this sector by facilitating their financial, technical and intellectual needs (e.g. access to low interest on borrowings, subsidized ICT infrastructure, innovation funds, tax relief for exporting, management training).In terms of limitations, the sample was drawn from the largest province of Pakistan which represents around two-thirds of SMEs in Pakistan although the results may not be generalizable to firms in other provinces. Social desirability effects may have encouraged respondents to portray a more positive image of the organization by inflating HRM sophistication. However, if this occurred, it seems likely to have affected all sectors in equal measure. A further limitation is that the performance of the firms was not measured. As such, it is not possible to judge whether greater HRM formality was associated with greater organizational performance. While we have shown clear formality differences across sectors, whether these differences have any bearing on performance requires further research.\n",
"role": "user"
},
{
"content": "Purpose: Guided by institutional theory, this empirical paper examines variations in the adoption of HRM practices among SMEs in three different business sectors (services, manufacturing and trade).\nDesign/methodology/approach: Data from 300 owners/managers representing three business sectors were collected through a survey method.\nFindings: The results suggest that service SMEs use more formal HRM practices than manufacturing and trade SMEs. Manufacturing SMEs are more formal than trade firms. Results are not affected by firm age.\nResearch limitations/implications: Social desirability bias may have influenced respondents into portraying a positive image of the organization by inflating HRM sophistication. A further limitation is that the performance of the firms was not measured. As such, it is not possible to judge whether greater HRM formality correlated with improved organizational performance.\nPractical implications: This study shows how the business sector shapes HRM practices in Pakistani SMEs. Findings help to inform Pakistan's Small and Medium Enterprise Development Authority (SMEDA) in dealings with manufacturing and trade firms in terms of improving HRM practices.\nOriginality/value: Given the important role of SMEs in economic development, comparative research on HRM in SME contexts is scarce. Since SMEs are vital for Pakistan's economy, an improved understanding of the sector's approach to human resource development is important. The findings extend the boundaries of prior comparative HRM literature in SMEs by addressing sector influences while controlling for contextual factors.\n",
"role": "assistant"
}
] |
A new stochastic multi source approach to improve the accuracy of the sales forecasts
|
[
"Uncertainty",
"Demand forecasting",
"Monte Carlo simulation",
"Forecasting model",
"Sales forecasts",
"Sales & operations planning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: One of the challenges that companies must face daily is being able to obtain an adequate and timely balance between:\n2. Literature review: There is an extremely rich bibliography in terms of innovative forecasting models for the most diverse sectors of application. Some of the most recent and innovative approaches have been developed for applications in the electrical or energetic field (Li et al., 2012; Jeong et al., 2014; Xiao et al., 2015; Arora and Taylor, 2016) and for the evaluation of investment's profitability or market dynamics (Zhu and Wei, 2013; Bendato et al., 2015a; Bonfiglio et al., 2014). A wide spread of forecasting models deals, similarly to what is discussed in this paper, with the sales of products and services (Guo et al., 2013; Lu, 2014; Bendato et al., 2015b; Fantazzini and Toktamysova, 2015; Ramos et al., 2015; Schneider and Gupta, 2016; Vhatkar and Dias, 2016). The launch of new products/services, for which historical data are obviously lacking, has also led to numerous attempts at innovative algorithms (Chern et al., 2010; Tanaka, 2010; Xiao and Han, 2016).\n3. Description of the methodology: As shown in Section 1, an S&OP process is based, for the development of the subsequent phases, on reliable demand planning.\n4. Test case: The test case was developed in collaboration with the management of a multinational, multi-product company in the food and beverage sector. Initially, a classic trend analysis was conducted on the sales of some products and in different geographical areas, over the 2011-2014 period. The current company's forecasting process is carried out over the four weeks following the current one, and, as a consequence, each forecast is updated three times. For each week, thus, four different forecasts (called Lag 3, Lag 2, Lag 1 and Lag 0) are generated.\n5. Conclusions: In dealing with the problem of drafting a realistic demand plan - a fundamental step of a valid S&OP - the authors focused their attention on forecast accuracy.\n",
"role": "user"
},
{
"content": "Purpose: The aim of this paper is to suggest a new approach to the problem of sales forecasting for improving forecast accuracy. The proposed method is capable of combining, by means of appropriate weights, both the responses supplied by the best-performing conventional algorithms, which base their output on historical data, and the insights of company's forecasters which should take account future events that are impossible to predict with traditional mathematical methods.\nDesign/methodology/approach: The authors propose a six-step methodology using multiple forecasting sources. Each of these forecasts, to consider the uncertainty of the variables involved, is expressed in the form of suitable probability density function. A proper use of the Monte Carlo Simulation allows obtaining the best fit among these different sources and to obtain a value of forecast accompanied by a probability of error known a priori.\nFindings: The proposed approach allows the company's demand forecasters to provide timely response to market dynamics and make a choice of weights, gradually ever more accurate, triggering a continuous process of forecast improvement. The application on a real business case proves the validity and the practical utilization of the methodology.\nOriginality/value: Forecast definition is normally entrusted to the company's demand forecasters who often may radically modify the information suggested by the conventional prediction algorithms or, contrarily, can be too influenced by their output. This issue is the origin of the methodological approach proposed that aims to improve the forecast accuracy merging, with appropriate weights and taking into account the stochasticity involved, the outputs of sales forecast algorithms with the contributions of the company's forecasters.\n",
"role": "assistant"
}
] |
Connecting influence tactics with full-range leadership styles
|
[
"Leadership",
"Influence tactics",
"Transformational",
"Transactional",
"Full-range",
"Passive avoidant"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Transformational and power-and-influence theories of leadership are among the most researched contemporary approaches to leadership (Dinh et al., 2014). However, little research has explored the potential conceptual and empirical overlap between these important theoretical traditions. This paper presents a study that seeks to broaden and deepen potential conceptual and empirical alignment of transformational and power-and-influence theories of leadership by closing methodological gaps left by previous research.\nPrevious studies of leadership style and influence tactics: Two studies have previously connected full-range leadership styles with influence tactics (Charbonneau, 2004; Lian and Tui, 2012). Charbonneau's (2004) study was the first to examine the relationship between influence tactics and transformational leadership. She surveyed 181 military personnel, who provided ratings of 80 peers. Charbonneau only obtained ratings of transformational leadership and use of the core influence tactics, and found that the core influence tactics correlated with an aggregate measure of transformational leadership and mostly correlated with the four sub-facets of transformational leadership that she measured. More recently, Lian and Tui (2012) surveyed 347 employees (mostly supervisors and managers) in Malaysia. They found that transformational leadership was positively correlated with inspirational appeals, consultation, and ingratiation, and negatively correlated with exchange and pressure. In contrast, transactional leadership was negatively correlated with inspirational appeals, consultation, and ingratiation, and positively correlated with exchange, pressure, and legitimating. Although these studies suggest that leadership styles and influence tactics are related, the evidence is incomplete in scope, has methodological limitations, and has not fully addressed the question of what tactics of influence are connected with transactional and passive-avoidant leadership.\nThe present study: This study aimed to enhance the potential theoretical integration of full-range and power-and-influence theories of leadership by addressing the methodological limitations of previous research (i.e. Charbonneau, 2004; Lian and Tui, 2012). First, all 11 influence tactics were measured using the extended Influencing Behavior Questionnaire (Yukl et al., 2008), which is a measure that allows followers to report on the influence tactics used by their leaders. Second, leadership styles were measured using the Multifactor Leadership Questionnaire (MLQ-5X) (Bass and Avolio, 1994) in both leader self-report and follower versions. This allowed for follower-rated influence tactics to be correlated with both follower-rated and leader-self-rated leadership styles.\nMethod: Participants and procedure\nResults: Data screening and assumption testing\nDiscussion: The results of the present study partially supported H1: transformational leadership was positively correlated with leaders' use of the core influence tactics. These correlations were significant for all four of the core influence tactics when transformational leadership style was follower-rated, but only for inspirational appeals when transformational leadership was leader-self-rated. Similarly, H2 was only partially supported, with the influence tactics, exchange and legitimating, but not pressure, correlating with transactional leadership. This partial support for H2 was limited only to correlations between influence tactics and follower-rated leadership styles. Moreover, follower-rated transactional leadership was, unexpectedly, significantly positively correlated with the core influence tactics. H3 was also partially supported, with some negative correlations observed between influence tactics and follower-rated passive-avoidant leadership. Additionally, leader-self-rated passive-avoidant leadership was only negatively correlated with the use of the consultation influence tactic. Taken together, the results for the three hypotheses provide insight into RQ1 concerning whether correlations between influence tactics and leadership styles would be similar when leadership styles were follower-rated vs self-rated. The correlations were weaker, and less often significant, when leadership styles were rated by the leaders themselves.\nLimitations and unresolved questions: Self-rated leadership style was included in the present study to overcome a critical methodological limitation of previous studies. However, in doing so, the results obtained raise more questions than they provide answers; specifically: whether correlations between follower-rated leadership styles and influence tactics were inflated by common methods variance; whether leaders lack insight into their own leadership styles and their connection with how their influence attempts are perceived by their followers; and whether non-significant correlations resulted from inadequate statistical power.\nConclusion: The present study adds important new information to the literature on theoretical and empirical connections between full-range and power-and-influence theories of leadership. This study confirms that followers perceive connections between the way their leaders lead and the ways in which their leaders attempt to influence them. Practically, leaders should use the core influence tactics and apprising, as these tactics accounted for more than 60 percent of the variance in followers' perceptions of them as transformational leaders. However, this study goes beyond previous studies in ways that may be helpful to organizational development practitioners and researchers. First, it adds to what is known from previous studies by including the passive-avoidant leadership style and additional influence tactics that were either not previously measured at all or were not previously related to transformational and transactional leadership. Second, follower-rated influence tactics were not strongly related to leader-rated leadership styles. Thus, additional research is needed to separate the potential explanations for the weak relationships this study found between follower-rated influence tactics and leader-rated leadership styles. Finally, the results of this study challenge Cerni et al.'s (2014) theoretical position that transformational and transactional leadership would be exclusively related to the use of specific tactics of influence by leaders.\n",
"role": "user"
},
{
"content": "Purpose: Full-range leadership theory and power-and-influence approaches to leadership emerge from different theoretical traditions, but may overlap conceptually and practically. Previous research has found connections between full-range leadership styles and leaders' influencing behaviors. However, this research has been conceptually and methodologically limited, neither examining all full-range leadership styles nor all common influence tactics, and measuring only employees' perceptions. The purpose of this paper is to address these limitations.\nDesign/methodology/approach: Follower-rated leadership styles and influence tactics, and leader-rated leadership styles were surveyed for 160 pairs of leaders and followers (n=320).\nFindings: Core influence tactics and apprising were correlated with, and predicted, follower-rated transformational and transactional leadership, but confirmatory factor analysis provided a more nuanced view of these relationships. Rational persuasion (negatively) predicted passive-avoidant leadership. There were few significant correlations between leader-self-rated leadership styles and followers' ratings of leaders' influence tactics.\nOriginality/value: This study extends upon previous studies by more comprehensively connecting full-range leadership styles and influence tactics, and doing so with both leader-rated and follower-rated leadership styles. The results help to provide a clearer picture of the overlap between full-range and power-and-influence theories of leadership. The mismatches between results from leader-rated and follower-rated leadership styles raise conceptual, practical, and methodological questions for future research.\n",
"role": "assistant"
}
] |
Sustainable development in an industrial enterprise: the case of Ontario Hydro
|
[
"Sustainable development",
"Transformational leadership",
"Organizational change",
"Social change",
"Canada"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Sustainable development has been identified as a key social, organizational and managerial concern for the new millennium (Agenda 21, 1992; Brundtland, 1987; Schmidheiny, 1992). The notion of sustainable development was articulated in the World Conservation Strategy (1980). It was defined more deeply and discussed widely as a result of the Brundtland (Committee) Report (1987). It was established as a program for international action through Agenda 21 (1992).\nContext and key concepts: Context\nResearch strategy and research methods: Before outlining the research strategy and methods we used to examine the events and process we briefly introduce Ontario Hydro and its strategy.Ontario hydro: brief background\nOntario hydro: case study: Formulation\nRenewable energy technology (RET) program: Formulation\nAnalysis and interpretation: The pathway of SED strategy\nDiscussion and conclusions: Industrial leadership or participation in sustainable development involves recognition of the need to commit to transformational change. It also requires the skills to facilitate and participate in collaborative processes with many actors, internal and external to the organization, who have an interest in the economic, environmental and social aspects of the organization's products, services and technologies. These processes are designed to envision more sustainable industrial systems, to identify the basis for action and change toward sustainable development and to provide a continuing forum through which emerging understandings of this vision and practice can be shared and used to revise the direction of change.This research investigated the formulation and implementation of a strategy for leadership in sustainable development by an industrial organization. There was no previous empirical research on the organizational pathways and processes of transformation used by industrial organizations to effect leadership in sustainable development that we could draw on when our study began. Indeed, no industrial organization has gone as far as Ontario Hydro in the authenticity of its interpretation of sustainable development or the ambition of its strategy. We used a conceptual framework, which anticipated that Ontario Hydro's ambition to lead the transformation of the Ontario system of energy development and use toward sustainable development would require it to engage in a complex process of paradigmatic change and learning. This would involve learning and action in collaboration with many actors in the organization as well as other actors in the overall system of energy development and use in Ontario. Ontario Hydro embarked on this complex social and industrial experiment but did not succeed in achieving its stated ambition. We identify a number of factors that contributed to this outcome but our main focus has been on the process of transformation.Our findings indicate that the designers of Ontario Hydro's strategy clearly recognized the need to embed stakeholder collaboration and partnership in their strategy. When the strategy was formulated they used these processes within the organization. However, external stakeholders were not engaged in formulating the strategy even though the intent was to bring about system-wide change through collaborative action. Moreover, the internal stakeholder collaboration, did not continue, once the strategy was approved.Since, the study began, the importance of integrating stakeholder inputs into company strategies directed toward environmental management and sustainable development has become an important strand in the literature on organizations and the natural environment (Roome, 1994; Hart, 1995; Hartman et al., 1999; Sharma and Vredenburg, 1998; Clarke and Roome, 1999).This work suggests that the integration of stakeholder inputs in organizational decisions derives from the systemic character of sustainable development issues and the requirements of product stewardship (Roome, 1994). Hart (1995) confirms the need for stakeholder inputs as a component of product stewardship and highlights the skills needed to integrate those inputs into organizational decisions. He maintains that these skills are socially complex and rare. Our study supports the notion that skills to integrate stakeholder inputs are based on the competence to facilitate collaborative multi-stakeholder processes. It also supports the claim that these are socially complex and rare. We also observed that the facilitation of these processes, whether internal or external, has to contend with multiple, shifting interpretations and experiences of sustainable development. These processes need to be held together by an integrating vision and by a commitment from stakeholders to remain involved to the process.Sharma and Vredenburg (1998) identify the importance of trust between stakeholders and companies if collaborative relationships are to flourish. In the case of Ontario Hydro, the failure to include external stakeholders in the formulation of the SED strategy created a gap between the rhetoric of commitment to collaboration and partnership and the fact that this commitment was then launched on an unsuspecting set of stakeholders. We saw evidence that this undermined the trust-base for future partnerships between Hydro and these stakeholders. Indeed, in the stakeholder dialogues that took place on the SED report the notion of partnership was construed by external stakeholders, to represent Ontario Hydro's agenda rather than the outcome of a collaborative exercise led by Hydro.Trust was also an issue internally within Hydro. Although the original SED Task Force involved managers at Ontario Hydro in a participative process, none of them was able to adopt this model within their own work to implement the strategy. This indicates the importance of the skill to maintain commitment to process and to develop trust in the capacity of participation to deliver outcomes that are relevant to SED, to Ontario Hydro and its business units. This means building managerial confidence in working collaboratively across traditional business and organizational lines as well as with external stakeholders to generate joint action directed toward agreed objectives for SED. In the absence of this trust, and without the necessary skills to manage the process, Ontario Hydro quickly reverted to its more formal and hierarchical approach to decision-making and managerial choice. As soon as the process was lost, the content of SED became caught in conflict between competing interpretive paradigms, each with their advocates within the corporation.Hartman et al. (1999), drawing on empirical work by Clarke and Roome (1999) on learning-action networks, recognize the reciprocal nature of stakeholder company relationships, which foster significant changes in how industry and society view one another and the environment. Our study indicates that managers at Ontario Hydro were unwilling or unable to engage in processes of dialogue with external stakeholders that might yield these new insights and new relationships. Managers found it problematic to work across internal organizational boundaries to share experiences and interpretations of sustainable development. This provides a possible key to understanding the issues surrounding sustainable development as a new paradigm for industrial organizations. It also helps establishes the challenge of sustainable development for management theory and practice.The vantage points from which managers addressed transformation and change, is consider central to our understanding of sustainable development at Hydro. Managers adopted a perspective determined by their function, their business unit or the organization. These vantage points lead to questions such as what is our position in the energy system of Ontario and how do we maintain that? Which stakeholders are salient to Ontario Hydro and why? How do we maintain our position and interests? How are our organizational or economic priorities best served? However, leadership in sustainable development as systems change re-frames these questions. It places emphasis on the economic, environmental and social sustainability of systems. In the Ontario Hydro case the system was energy development and use in Ontario. From this perspective the questions are different. For example, what does a sustainable system of energy development and use look like? Which actors are salient to the system of energy development and what are their interests? What skills and processes are needed to bring them together into collaborative action to transform this system to our common vision of sustainability? How are organizational priorities, economic, environmental and social interests in this system sustained? How are organizational priorities sustained?In accepting the ambition of leadership to transform the Ontario energy system to sustainability Ontario Hydro moved from a purely organizational perspective toward a wider systems perspective. The evidence from the case is that senior managers at Ontario Hydro were unable or unwilling to alter their perspective in line with the ambition of the SED strategy.The case study also suggests that the processes of innovation we saw were highly distributed through the organization and its system. They apply to the formulation and implementation phases of strategy, to strategy and its supporting programs and operational activities. They span all organizational levels from those dealing with strategic issues down to operational choices. They involve corporate center as well as operational business units. They are particularly important at the interface between levels, especially the interface between corporate strategy, the management of R&D and technology and business units. They are necessary between business units and between internal managers and external stakeholders. In other words sustainability as systems change respects none of the conventional boundaries developed by organizations to help them to manage their normal, non-sustainable activities. In that sense, sustainable developments appears as organizational and social innovation requiring the leadership of many individual actors to facilitate the complex process of negotiated transformation.\n",
"role": "user"
},
{
"content": "Purpose: - This paper seeks to present a longitudinal case study of Ontario Hydro - an industrial organization that used sustainable development as the basis for a strategy of social and organizational transformation.\nDesign/methodology/approach: - The paper describes the complex factors that influenced the formulation and implementation of this strategy.\nFindings: - The findings indicate the advanced ambition and authenticity of Ontario Hydro's strategy, even though it was formulated some ten years ago. The study suggests that the strategy was abandoned for reasons that include the gap between the processes identified in the strategy and the processes followed in practice, the absence of platforms to discuss and agree the meaning and practice of sustainable development within the company and its wider system, and the scarcity of skills to facilitate sustainable development as a process of multi-actor innovation. Consequently, the concept of sustainable development was not translated into practices that had shared meaning for the many actors involved in the energy system of Ontario. While based on a case study of one organization, the findings appear to speak to more general issues of sustainable development as the management of organizational and contextual change.\nPractical implications: - The paper indicates much about the process of organizational change to effect more sustainable practices within a company and its social context.\nOriginality/value: - No other organization has pursued a strategy for sustainable development with the same claim to authenticity as that of Ontario Hydro, where the strategy was cross-referenced to Agenda 21 and developed with input from some of the main architects of Agenda 21. Moreover, few studies of sustainable development in the literature span a period as long as this case.\n",
"role": "assistant"
}
] |
The role of Yin-Yang leadership and cosmopolitan followership in fostering employee commitment in China: A paradox perspective
|
[
"China",
"Japan",
"Cosmopolitanism",
"Yin-Yang",
"Paradox",
"Employee commitment",
"Cross-cultural leadership"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The cross-cultural and expatriate leadership literature relies heavily on cultural value differences between societies in the investigation of cross-cultural challenges (Stahl et al., 2017; Stahl and Tung, 2015). The predominant approach is through assessing the scores of cultural dimensions (e.g. those from Hofstede and the GLOBE), where differences in cultural values, or cultural distance (Kogut and Singh, 1988) between societies are thought to affect cross-cultural leadership effectiveness (Dickson et al., 2003; Mendenhall et al., 2008; Wang et al., 2012). While leadership studies adopting the cultural values approach have generated many valuable insights in the cross-cultural management and international business fields, there is increasing awareness that cross-cultural studies should also be approached from perspectives other than differences in societal cultural values (e.g. Beugelsdijk et al., 2017; Kirkman et al., 2006; Leung and Morris, 2015; Tung and Verbeke, 2010).\nYin-Yang leadership: Contingency theories of leadership state that the most appropriate leadership approach is dependent on context (e.g. Vroom and Jago, 1988). That is, managers can improve employee attitudes and performance by exhibiting an appropriate leadership style that is suitable to a given context. In the cross-cultural management literature, it is often argued that the ideal leadership style in a particular context is highly contingent on national culture (Dorfman et al., 1997; House, 1971; Javidan et al., 2006; Larson et al., 1974). For instance, knowledge of cultural dimensions and leadership behaviors expected by members of a given society can inform managers how to adapt their style (Javidan et al., 2006). The implication is that expatriates need to adopt leadership behaviors that are relevant to the host country's local culture, in order to maximize the effective leadership of local employees (Dickson et al., 2003; House et al., 1999).\nCosmopolitan followership: There is growing awareness in the leadership literature that the follower's role and characteristics are critical, but under-investigated, in the leadership process (Kelley, 2008; Uhl-Bien et al., 2014). Similarly, the cross-cultural and global leadership literature has identified key leadership capabilities, including being culturally relevant (Dorfman et al., 1997; House, 1971; Javidan et al., 2006), and managing contradictions and tensions (Zhang et al., 2015). While the role of the leader has been extensively discussed, the cross-cultural leadership literature has been surprisingly silent on the follower's role. We argue that followership plays an important role in cross-cultural contexts. In the case of Chinese employees who work in a foreign subsidiary of MNCs, the firm's routines and practices may not be culturally or organizationally compatible with the local Chinese context.\nMethod: Sample\nResults: Table I shows the means, standard deviations and inter-correlations of the study variables. Table II presents the results of the multiple regression analyses which test our research hypotheses. We used hierarchical multiple regression analyses. In model 1 we entered the control variables, including gender, tenure, employee's Japanese language proficiency, supervisor's Chinese language proficiency, and job satisfaction. In model 2 we added our independent variables, Yin leadership behaviors, Yang leadership behaviors, and cosmopolitanism. In model 3 we added three two-way interaction terms, beginning with our main variable Yin-Yang leadership (Yin leadership behaviors x Yang leadership behaviors). We also included two other two-way interaction terms of Yin leadership behaviors x Cosmopolitanism and Yang leadership behaviors x Cosmopolitanism as it is necessary to include all possible combinations of two-way interaction terms that comprise the three-way interaction in model 4. Finally, in model 4 we added the three-way interaction term of Yin leadership behaviors x Yang leadership behaviors x Cosmopolitanism to test our H3, the interaction effect of Yin-Yang leadership and cosmopolitanism on employee commitment. All of the interaction terms were calculated based on the standardized values of each component. The b coefficients of the final model, model 4, are used to determine support for the research hypotheses.\nDiscussion: Traditional bipolar cultural frameworks (e.g. Hofstede's and the GLOBE's) that allocate national cultures from low to high along a few measurable cultural dimensions have intellectual roots in Aristotelian dualism or Western formal logic. Western epistemological systems that facilitate mental separation of complex, interrelated phenomena into simplified opposites of \"either/or,\" and do not leave room for seeming opposites to coexist, have limitations particularly in the contemporary global business environment (Chen, 2008; Keller and Lewis, 2016; Lee, 2017; Li, 2016; Tung, 2008; Tung and Verbeke, 2010). The Eastern epistemology of the Yin-Yang principle, however, which embraces paradox and dynamism (Keller and Lewis, 2016; Li, 2016), offers a more robust approach to managing complexity. Drawing on the paradox perspective and the Yin-Yang principle (Faure and Fang, 2008; Li, 2012, 2014, 2016), we conceptualized \"Yin-Yang leadership\" as leader behaviors derived from a two-step process of translation and contextualization that embrace and respond to these opposites and \"dualities\" in Chinese culture, and proposed it as culturally relevant leadership in contemporary China. We demonstrated in our sample that leadership behaviors that respond simultaneously to the cultural needs of nurturing and building relationships while focusing on business development in a fast-paced, changing institutional environment contribute to employee commitment. We also demonstrated that Chinese employees' cosmopolitanism, the follower side of paradox management, also contributes to employee commitment. Details of our contributions are outlined below.\n",
"role": "user"
},
{
"content": "Purpose: Utilizing a paradox perspective, the purpose of this paper is to investigate the leadership-followership dynamic in foreign firms in China, specifically, the extent to which Yin-Yang leadership behaviors of Japanese expatriates and cosmopolitanism of Chinese employees influence employee commitment.\nDesign/methodology/approach: Data were collected through an online survey of Chinese employees who directly report to a Japanese supervisor in a Japanese subsidiary in China. Based on responses from 97 Chinese employees in three Japanese subsidiaries in China, the authors test if their cosmopolitan orientation and perceived Yin-Yang leadership behaviors of Japanese supervisors are related to employee commitment.\nFindings: Yin-Yang leadership and cosmopolitan followership have a positive effect on employee commitment. Further, cosmopolitanism moderates the link between Yin-Yang leadership and employee commitment such that the follower's cosmopolitanism compensates for lower levels of Yin-Yang leadership, especially a relative lack of Yin leadership behaviors.\nResearch limitations/implications: Results suggest that Yin-Yang leadership and cosmopolitan followership work together as a two-way street of cultural adaptability to build employee commitment, highlighting the interplay between leadership and followership in multinational enterprises. Future research should attempt to further refine the Yin-Yang leadership construct, and to gain a larger sample representing multiple expatriate nationalities to corroborate the relationships found in this study.\nOriginality/value: The study applies a context-based approach to developing culturally relevant leadership, through analyzing both the emic and etic concepts of culture in China. In doing so, the authors extend the application of paradox theories to the cross-cultural leadership literature utilizing the Yin-Yang principle, which is particularly relevant in societal contexts where rapid and dramatic change brings to the fore competing values, needs and employee preferences.\n",
"role": "assistant"
}
] |
Highly focused document retrieval in aerospace engineering: User interaction design and evaluation
|
[
"Semantics",
"Information retrieval",
"User studies",
"User interfaces",
"Aerospace engineering",
"Jet engines"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Organisational memory, the ability of an organisation to record, retain and utilise information from the past to bear on present activities (Stern, 1995), is a key issue for large organisations. The possibility of observing and reflecting on the past is particularly valuable in highly complex domains as it can inform and sustain decision-making. Civil aerospace engineering is one example: the life cycle of a jet engine can last 40-50 years from initial conception until the last engine is removed from service. During this long product lifetime a vast amount of information is accumulated. When a new engine is designed, it is important to reflect on previous technical solutions, to identify what has been successful and what instead needs to be revisited. In this context, even if a procedure for capturing newly created information has been introduced recently (e.g. online databases), the potential value of legacy data of the past 15/20 years is high.The value is, however, only potential as it has to be balanced by the effort to spend to get to the relevant piece of information. A room (or a computer) full of documents in no particular order would discourage even the keen searcher. Easy access is therefore paramount to make organizational memory useful and used.Another issue is the currency of legacy data. In the case of organizational memory currency does not relate to how recent the information is, but on the validity it has despite the elapsed time. It has much to do with context, more specifically if the context in which the information was created matches (or is important) to the current context of information use then its currency (and its value) is high (Stern, 1995). Contextualized information goes a step further from being a description of data or facts and becomes knowledge (Ackoff, 1989). Retrieving knowledge gives an advantage on retrieving information as the first step for interpretation (information in context) is already outlined. Retaining, organizing and retrieving knowledge poses a number of challenges over traditional information management (i.e. retention, organization and retrieval of information).This paper describes the first phase of an effort to introduce digital technology in the capturing, organising, searching and sharing of operational experience in a complex organisation. The vision is an integrated tool that supports: knowledge acquisition, organisation, retrieval and sharing of corporate memory, knowledge and expertise. The case study reported is that of Rolls-Royce plc and their aerospace engineering activity. This paper in particular focuses on the retention of legacy information, its automatic organization into knowledge (i.e. information in context), its retrieval and use.The paper is organized as follows. Section 2 provides a short overview on systems for organizational memory, focussing on semantic solutions. Section 3 describes the domain in terms of information practice and type of documents. Section 4 summarises the requirements of highly focused search, while section 5 describes the design of the K-Search system. The user evaluation and general results are discussed in section 6 and 7 respectively. Section 8 reflects on the findings and concludes the paper.\n2. Systems to support organizational memory: The need for technology that supports the capturing, classification and finding of information and knowledge in an organization over a period of time has been recognized for quite some time. One of the first systems to have explicitly addressed the issues of organizational memory is Answer Garden (Ackerman and Malone, 1990) that structured into a browsable hierarchy answers provided by experts to problems posed to a mailing list. The classification was manually done; the evaluation showed a clear benefit as users could find answers straightaway and new knowledge was cumulated as new questions were received. This approach has held interest until now (Pipek and Wulf, 2003), but in recent years Web 2.0 technologies to support the capturing and classifying of knowledge while at the same time motivating users to contribute their knowledge has stimulated new research. Wikis have been naturally considered as collaborative environments for creating and editing knowledge (White and Lutters, 2007). Various projects have enriched the Wiki model with semantics (Millard et al., 2006, Krotzsch et al., 2006), focusing on collaborative production of semantic knowledge (Alquier et al., 2009). A drawback of these approaches is that they focus on capturing; while the retrieval is limited to browsing through organized hierarchies of links (interested readers can find a review of Semantic Media Wikis in Millard et al. (2008)). Moreover, the need for manual labour makes the use of these techniques for archival material too expensive to adopt in any organization.At the opposite end of the spectrum lies the retrieval of structured knowledge instead of sparse information. Semantic search has been developed for this purpose and makes use of ontologies and knowledge bases to drive and execute the user's queries, returning the results with ontology-related visualizations (e.g. Maedche et al., 2003; Popov et al., 2004; Castells et al., 2007). What distinguishes semantic search systems is whether they keep (and display) a reference to original documents (by storing triples linked to the documents with the annotations (Guha et al., 2003)), or are limited to storing and displaying ontology triples (Castells et al., 2004; Maedche et al., 2003). Most of these research systems, however, are limited to semantic search only, missing the advantage of full keyword search (Lei et al., 2006), but enterprises have started to understand the advantage of adopting semantic technology to improve the performance of their retrieval systems. However, the potential of semantic technology is not exploited and its use is limited to matching query terms to similar concepts (synonyms)1.\n3. Information in aerospace engineering: 3.1 Information production and use\n4. Requirements: The analysis of the user and their task, and the analysis of the corpus resulted in a list of requirements that directed the design of both user interaction and retrieval functionalities. The search system should:1. Make legacy data available for contextual search via concepts at limited cost.2. Support the composition of complex, context-rich queries in an easy to understand and simple way.3. Retrieve contextualized information, i.e. only documents where the term occurs in the context specified by the user in the query should be returned.4. Retrieve with high recall, i.e. all documents containing knowledge (information in context) relevant to the query.5. Retrieve with high precision, e.g. only documents where FMU was removed and not those where FMU was just tested.6. Support users in quickly changing their search strategy or focus.7. Accommodate individual perspectives: different user classes (e.g. service engineers or designers) and their specific terminology should be supported.8. Support the statistical analysis of retrieved results via dynamically generated charts that are:* customizable - the user selects dimensions to explore alternative views on the retrieved data; and* interactive - support drilling into specific aspects captured in the chart.9. Provide multiple views on the same retrieved set (e.g. multiple charts, many documents opened simultaneously) to support data exploration.The next section describes the design rationale, i.e. how requirements drove the K-Search system design.\n5. K-Search design: 5.1 The design of the search functionality\n6. User evaluation: 6.1 Method\n7. The effect of the user's goal: The flexibility of the interface in accommodating searching styles emerged in several situations. There was no prescription on how to compose a query: some users started with keyword search and then switched, others started with concepts, while others combined immediately the two from the start.The exploration of the results seems based on individual attitude and preference too. Some participants used the chart or the highlighted concepts as query expansion tools; all opened several documents (in tabs) and moved between them to perform comparisons.The clearest indication of individual use of a common system emerged when analysing the self-selected task. Different users looked for different content in the same documents: at design time ERs are important for the facts they report, but some participants were interested in the cost of the event and not in its occurrence. K-Search failed to satisfy those user class needs, but simple re-training to extract new data allowed those needs to be satisfied in the following version of the system.Differences emerged between groups. Participants were classified as service engineers (32 per cent), designers (39 per cent) and others (e.g. cost engineers, 29 per cent).Designers searched mainly for instances of components (45 per cent), problems (15 per cent), and engines (19 per cent), concepts (9 per cent), or event (7 per cent) were present in a minority of queries.Service engineers looked to the data from the point of view of finding events (35 per cent) and engines (32 per cent), in a much smaller proportion from the point of view of problem (15 per cent) and component (18 per cent). No interest was shown for concepts (0 per cent)Participants belonging to other groups had a less distinctive search strategy: 37 per cent started from the component, 37 per cent from the engine, 16 per cent from an event, 10 per cent from concepts and none from a problem (see Figure 8).Service engineers and designers differ also in search strategy. Service engineers clearly preferred hybrid search, 61 per cent, with 24 per cent preferring semantic and 15 per cent keyword search.Designers favoured keyword search, 43 per cent, hybrid was used in 30 per cent of cases and semantic only in 27 per cent. The preference for using keyword search could be explained by the potential it offers to explore the data irrespective of classification: designers were able to investigate components from many and diverse points of view.Other groups used semantic search most, 66 per cent, 24 per cent hybrid, and 15 per cent keyword. A possible explanation is that semantic search is driven by the concepts in the ontology and therefore provides an easier way to search unfamiliar data.In summary, the user and their task are fundamental to capturing the actual context of use. This in turn is fundamental to directing the design, both interaction and functionalities, that is able to accommodate a wide variety of styles and goals.\n8. Discussion: The study was successful in showing the possibilities technology offers for the automatic capturing of archival material into a usable format that can be easily accessed and used by everyone in an organization. This is an indispensable step as the value of legacy documents is uncertain and the cost of manually acquiring knowledge from them is prohibitive (e.g. reading them one by one and inputting the core information in a database). The value of such an effort was made explicit during the user evaluation as more than one participant asked to be able to use the system in their everyday job: one person ended up printing a few of the retrieved documents as relevant for a task in hand but which they had not previously discovered. Having the historical archive available to everyone opens up new uses of this information, particularly by employees who would not normally access it. For example, those planning business cases would not look at ERs individually but at digested summaries; being able to break down the summaries into single events and access the facts directly allows in-depth analysis that was not possible before.The outcome of the user evaluation was used to fix minor issues on the user interface For example, search for range of dates, and re-run the information extraction module to capture new knowledge that emerged as relevant. The system was then delivered to Rolls Royce for actual use. The data set has been progressively enriched with other kinds of documents selected from the large Rolls Royce archive. By re-creating electronically the rich legacy, the organizational memory is progressively composed into a coherent and organic set that overcomes the fragmentation that occurs when paper copies or different databases are used by different departments in charge of storage.As with all automatic acquisition techniques, those used by K-Search may also miss some relevant information or misrepresent others, with the consequence that errors are stored in the database used5. Users accustomed to the certainty of manually created and validated centralised databases, (nearly) always correct and exhaustive, need to be made aware of the limitations of these techniques. When questioned about this issue, participants stated that the advantage of having archives easily accessible outweighs any issues of correctness, accuracy and exhaustiveness.The quality of the automatically acquired data depends on the initial format of the archive and on the number of steps applied to process archival documents into the repository. The original form of the Event Reports used in this study was Microsoft Word; this was then converted into HTML, on which information extraction was applied. The conversion Word-HTML is a potential source of noise (i.e. errors), that impact on the following information extraction, which in turn can fail to detect features. It goes without saying that the higher the number of the steps the greater the number of the errors. During the acquisition of new archives we came across a repository of files that contained scanned documents on which manual annotations had been made. The original Word documents were no longer available and to include this set into K-Search required converting the scan into HTML. Both the conversion from scan to PDF and then from PDF to HTML introduced errors. However, once more, being able to search those highly important documents overcame any drawbacks. What is worth pointing out is that strategies could be put in place to help users discover data mismatches and support updating and corrections that will eventually converge the data set to a correct and complete repository (Petrelli et al., 2005).Since K-Search was introduced for everyday use in Rolls Royce, it has evolved into a suite of tools (K-Tools) for capturing, managing and organizing knowledge using semantic technology. The whole knowledge lifecycle is now supported: the users can define their own knowledge representation and use self-defined forms to formalize and capture new knowledge (using K-Forms (Bhagdev et al., 2008b)) that is then stored and available for sharing and reuse via K-Search. By adding a way to capture new knowledge, K-Tools allows the recovery of organisational memory from the past and combines it with newly captured knowledge, creating a sustainable and consistent way of linking past information and current issues.\n9. Conclusions: To study a case of interactive information retrieval requires considering other elements beside the system: its user, their task, and the data are all equally important in setting up the context of use. When designing such systems, decisions have to be taken on which features are important and what should the system do; evaluations then have to be set up in order to test if the system is as effective as expected.We showed how the analysis of users and data affected the design of the system in terms of both search functionalities and interaction. An innovative search paradigm that combines ontology-based and keyword-based search in a hybrid system was devised and tested with users in the aerospace industry.The purpose of the user evaluation was twofold: to measure usability and to investigate how the different user groups would use the system. The results show the interaction design was successful while the data needed to be processed again to accommodate different users' interests. It also showed how users approach the system in very different ways, sometimes showing characteristics of a group, sometimes as singular individuals.Since the study reported here, K-Search has been updated: different document types have been added and the user interaction has been modified where needed (e.g. suggesting possible values for a selected component). After a monitored trial, K-Search is currently in use at RR in Derby.\n",
"role": "user"
},
{
"content": "Purpose: - This paper seeks to describe the preliminary studies (on both users and data), the design and evaluation of the K-Search system for searching legacy documents in aerospace engineering. Real-world reports of jet engine maintenance challenge the current indexing practice, while real users' tasks require retrieving the information in the proper context. K-Search is currently in use in Rolls-Royce plc and has evolved to include other tools for knowledge capture and management.\nDesign/methodology/approach: - Semantic Web techniques have been used to automatically extract information from the reports while maintaining the original context, allowing a more focused retrieval than with more traditional techniques. The paper combines semantic search with classical information retrieval to increase search effectiveness. An innovative user interface has been designed to take advantage of this hybrid search technique. The interface is designed to allow a flexible and personal approach to searching legacy data.\nFindings: - The user evaluation showed that the system is effective and well received by users. It also shows that different people look at the same data in different ways and make different use of the same system depending on their individual needs, influenced by their job profile and personal attitude.\nResearch limitations/implications: - This study focuses on a specific case of an enterprise working in aerospace engineering. Although the findings are likely to be shared with other engineering domains (e.g. mechanical, electronic), the study does not expand the evaluation to different settings.\nOriginality/value: - The study shows how real context of use can provide new and unexpected challenges to researchers and how effective solutions can then be adopted and used in organizations.\n",
"role": "assistant"
}
] |
Resale prices in franchising: insights from franchisee perspectives
|
[
"Retailing",
"Brand image",
"Franchising"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Franchising is defined by the European Code of Ethics for Franchising (2008) as:\n2. A brief literature review of resale pricing in franchising: The literature on resale pricing in franchising remains scarce. Some scholars have examined the question of resale pricing from a legal perspective. For instance, Blair and Esquibel (1996) focused on maximum resale price restraints in franchising in relation to anti-trust laws. Basset and Perrigot (2015), rather than limiting their legal analysis to maximum resale prices, studied franchisee restrictions within the context of current legislation and regulations, including recommended resale pricing, maximum resale pricing, fixed resale pricing and minimum resale pricing and franchise contract clauses and their connection to franchisee obligations. In their study, they examined various forms of recommended resale pricing and provided a discussion of resale price-setting policies, in general, along with suggestions for setting specific resale price-setting policies. They also urged that franchisees reject certain practices within the franchising sector, such as pre-tagging prices on products, pre-printing prices on order forms and advertising and pre-recording prices in cash registers.\n3. The legal framework for resale pricing in franchising: 3.1 The principle: the freedom of the franchisee in terms of setting resale prices\n4. Research methodology: To assess the perceptions of franchisees regarding resale pricing and, more specifically, their perceptions of their freedom to set resale prices; their perceptions of recommended resale prices; their perceptions of directly imposed resale prices; and their perceptions of indirectly imposed resale prices via underlying threats; via pre-tagged prices on products; via pre-recorded prices in cash registers; and via advertising campaigns, we used a qualitative approach and in-depth interviews. Qualitative approaches have been adopted in previous literature on franchising. For instance, Weaven and Frazer (2007) and Perrigot and Herrbach (2012) interviewed franchisees from several chains and several industries to investigate multi-unit franchising and plural form, respectively.\n5. Findings: 5.1 Franchisee perceptions of their freedom to set resale prices\n6. Discussion: 6.1 Summary of findings\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to offer a novel perspective on resale pricing in franchising, i.e. from a franchisee perspective, by combining legal and managerial considerations in the European context. The objective is to assess franchisee perceptions regarding resale pricing in their chains.\nDesign/methodology/approach: The authors adopt a qualitative approach and use 46 in-depth interviews with franchisees covering retail and service industries in the French market.\nFindings: Many of the interviewed franchisees believe that joining a franchise chain involves respecting the recommended resale prices. For some of the franchisees, in link with the chain uniformity, imposing uniform resale prices throughout the chain represents a strength, because customers who visit different stores within the franchise chain expect to find consistent pricing. Moreover, many franchisees consider that their franchisors have some know-how that they use to set correct resale prices, taking into account the profit margin.\nResearch limitations/implications: This research contributes to the literature on resale pricing in franchising, as well as the franchising literature in general, by combining legal and managerial considerations, adopting a franchisee perspective, covering retail and service industries and focusing on French and European markets.\nPractical implications: This research can be viewed by franchise experts, franchisors, franchisees and franchisee candidates as a synthesis of resale price-related legal aspects, adopted practices and potential conflicts in franchise chains in the French market. It also highlights price-related practices to be avoided to prevent potential conflicts.\nOriginality/value: The subject of resale pricing in franchise chains is a hot topic, because of its link with customer attraction, chain uniformity, franchisor know-how, franchisee autonomy and the legal dimension.\n",
"role": "assistant"
}
] |
Female labour force participation and economic development
|
[
"Economic development",
"Female labour force participation",
"Panel fixed effect estimation",
"Demographic transition",
"Interactive modelling",
"Arellano",
"Bond estimates"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: This paper presents a theoretical model and empirical results related to women's labour force participation and the process of economic development. Boserup (1970), Goldin (1995), and Mammen and Paxson (2000) recognise U-shaped relationships between female labour force participation rates and per capita income around the world, as Figure 1 shows for a cross-section of 40 countries in 2005. That is, women's labour force participation rates first fall and then start to increase again during the process of economic development; the female participation rates are relatively high in low-income countries (e.g. Tanzania, Madagascar, Zambia), relatively low in middle-income countries (e.g. Egypt, Tunisia), and then relatively high in both upper middle (e.g. China, Brazil) and high (e.g. Canada, Italy, the USA) income countries. Table I provides an overview of all these countries.\n2. Theoretical framework: The labour force participation rate can be derived from a choice model between consumption and working time, subject to the usual budget constraint that (the present value of) consumption does not exceed (the present value of) income. At the micro level, the decision to participate in the labour market is a dichotomous random variable, equal to 1 if the decision is positive and 0 if it is negative. If we start with country level, instead of individual, data, the variable consists of the proportion LFPi of women who belong to the female working age population in country i (i=1, ..., N) who decide to participate. We present a theoretical framework to identify key determinants of the individual labour force decision, before we explain the transition from the micro level to the country level.\n3. Empirical analysis: implementation: The data we use for the empirical analysis refer to 40 countries and 45 years (1960-2005)[3]. We selected observations in five-year intervals (1960, 1965, ..., 2000). Because the data set is not complete, we obtained 326 observations. The countries belong to different income classes (see Table I), so the sample should cover every part of the U-shaped relationship between the female labour force participation rate and economic development. The dependent variable is female labour force participation rates in 10 five-year age groups (15-19, 20-24, ..., 60-64). In addition, we report the aggregated results (15-64 years) as a kind of benchmark.\n4. Empirical analysis: specification tests: Because female labour force participation may have a negative feedback effect on fertility rates, we treated the fertility rate and the product term between the fertility rate and the capital-to-labour ratio as two separate endogenous explanatory variables. We tested several potential instrumental variables, and eventually selected the birth rate lagged five years in time, measuring the number of childbirths per 1,000 of the population, the infant mortality rate (up to 1 year of age), and the mortality rate under five years. Data on the birth rate were extracted from the World Population Prospects (United Nations, 2006), and data on infant mortality rates came from the World Development Indicators (World Bank, 2007). Summary statistics of these variables are presented in the bottom rows of Table II.\n5. Empirical analysis: results: The share of employment in agriculture appears to have a positive effect on the female participation rate in every group (significant for five age groups); the interaction effect between the share of employment in agriculture and the capital-to-labour ratio reveals a negative effect (significant for all age groups). The positive sign of the agricultural variable might be explained by the women's ability to combine farm work with family responsibilities in low-income and developing economies. We explain the negative sign of the interaction term with two rationales. First, a rising capital-to-labour ratio represents a structural shift, from employment in agriculture to industry and services. Consequently, the agricultural sector employs fewer women, but they cannot immediately move to other sectors, due to their lack of skills and knowledge. Second, when an economy develops, the mechanisation of farm activities increases labour productivity, so fewer people are needed in the agricultural sector. In this respect, Boserup (1970, 1990) and Gladdis and Klasen (2014) argue that industrialisation marginalises women by hindering their participation in wage work. Our results corroborate the hypothesis that agricultural employment has a positive effect in lower-income countries and a negative effect in higher-income countries. The point at which the marginal effect, which is the first-order derivative for the total female working age population with respect to agriculture, changes its sign amounts to $9,210. This point reflects the real-world situation: in countries of Africa and South Asia, the capital-to-labour ratio is lower than the level at which the marginal effect of the share of employment in agriculture changes sign. In these two regions of the world, we also find the highest rate of women employed in the agricultural sector, namely, 69 and 61 per cent in 2007, respectively. Schultz (1990) finds that changes in the sectoral composition of the labour force trace the U-shaped relationship between female labour force participation rate and economic development. Our results are consistent with this finding.\n6. Conclusion: To test whether women's participation in the labour force first declines and then rises with economic development, we have presented an economic-theoretical model of female labour force participation indicating that the capital-to-labour ratio is a better proxy for economic development than GDP per capita. We also have presented empirical evidence in favour of a more flexible functional form than those that appear in previous studies. Because female labour force participation may have a negative feedback effect on the fertility rate, one of the main explanatory variables, we also tested for the exogeneity of the fertility rate. However, we must reject this hypothesis, so we used instrumental variables to obtain consistent parameter estimates.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present a theoretical model, which is aggregated across individuals to analyse the labour force participation rate, and empirical results to provide evidence of a U-shaped relationship between women's labour force participation and economic development.\nDesign/methodology/approach: The U-shaped relationship is investigated by employing a panel data approach of 40 countries around the world over the period 1960-2005. It is investigated whether the labour force behaviour of women in different age groups can be lumped together by considering ten different age groups.\nFindings: The paper finds evidence in favour of the U-shaped relationship. For every age group and explanatory variable in the model, a particular point is found where the regime of falling participation rates changes into a regime of rising participation rates.\nResearch limitations/implications: To evaluate this relationship, microeconomic analysis with primary data can also provide significant insights.\nSocial implications: Every country can narrow the gap between the labour participation rates of men and women in the long term. Fertility decline, shifts of employment to services, part-time work, increased opportunities in education, and the capital-to-labour ratio as a measure for economic development are the key determinants.\nOriginality/value: In addition to the U-shaped relationship, considerable research has been carried out on demographic transition. This paper brings these two strands of literature together, by econometrically investigating the impact of demographic transition on female labour force participation given its U-shaped relation with economic development, i.e., turning points for different explanatory variables are calculated and their implications for economic growth are discussed.\n",
"role": "assistant"
}
] |
Interorganisational collaboration at UK World Heritage Sites
|
[
"Tourism",
"Business planning",
"Heritage",
"Conservation",
"United Kingdom"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction and context: Interorganisational collaboration\nRationale and methodology: Within the tourism context, Jamal and Getz (1995, p. 188) have defined IOR collaboration in the following terms:... a process of joint decision-making among autonomous, key stakeholders of an interorganizational/community tourism domain to resolve planning problems of the domain and/or to manage issues related to the planning and development of the domain.Using this as a benchmark to assess the extent of interorganisational collaboration in the management of World Heritage Sites, the evidence for this paper comes from content analysis of WHS Management Plans and in particular the programmes of action that detail specific objectives. The management plans are now considered to be the most important strategic framework for integrative site management between organisations. As Rodwell (2002, p. 15) highlights, the plans are also \"promotional documents\" written for a specific target audience with the purpose of either achieving World Heritage Site status, or, sustaining it at the monitoring stage, and as such have a high international profile.Of the 23 sites within the UK, Management Plans for two sites - Durham Cathedral and Castle and Blenheim Palace - were unavailable. In relation to the remaining sample of 21 World Heritage Site Management Plans, only those plans that detailed the organisations responsible for implementing the specific objectives were included. This reduced the sample to 12 management plans, broadly divided into two categories:1. complex sites (5) - City of Bath, Derwent Valley Mills, Hadrian's Wall, Maritime Greenwich and Liverpool Mercantile Maritime City; and2. single focus (7) - Fountains Abbey, Giant's Causeway, Ironbridge, Neolithic Orkney, Royal Botanic Gardens, Kew, Stonehenge and Tower of London.The fact that only 12 plans specified the organisations responsible for implementing objectives was surprising given that \"many consider the management plans [...] to be exemplars in their field\" (Rodwell, 2002, p. 18). The management plans generally share a commonality in the focus of objectives in four key areas:1. statutory and policy;2. conservation;3. sustainable tourism and visitor management; and4. sustainable traffic and transportation objectives.Tourism and visitor management objectives generally refers to the provision of visitor facilities, interpretation and education, the carrying capacity of visitors and the need to monitor and survey visitors and the wider tourism context. In general terms, sustainable traffic and transportation objectives refers to increasing public transport, improving and increasing cycle routes, walking routes and pedestrian access, reviewing park and ride schemes, car parking provision, signage and information, increased usage of vehicles with zero emissions and preparation of green transport plans.As a small number of key agencies are responsible for the protection and conservation of world heritage sites, objectives in these categories do not by their nature tend be \"collaborative\". It was also noted by ICOMOS in its monitoring report on site management that:... individual agencies can plan and manage their own responsibilities effectively, but lack machinery for taking decisions going beyond their own areas of responsibility [...] This produces a high standard of maintenance and physical care, but a lower standard of planning for visitors access and facilities, because that often involves several agencies (ICOMOS, 1995, p. vii).Hence, the content analysis focuses solely on objectives in the latter two categories (sustainable tourism and visitor management, and sustainable traffic and transportation) and singles out only those objectives that are detailed as being implemented by two or more named organisations. Given the non-statutory and voluntary nature of the plans, this is, in effect, the primary evidence that collaboration is intentional. Interorganisational activity in relation to these two categories is profiled under the Sustainable Tourism Objectives.The methodology comprised three stages. Stage 1 produced a general profile of intended interorganisational collaboration illustrating the involvement of key stakeholders. Stage 2 produced a deeper analysis of the data and details the nature of true collaboration by eliminating the owners and lead organisations from the data. In each instance the data is presented in three tables:1. a profile and frequency table of the types of organisation involved in the implementation of sustainable tourism objectives at each site;2. a ratio table indicating the ratio number of organisations involved in the implementation of each objective; and3. a ratio table indicating comparison by site category.Ratio analysis is a tool commonly used to measure business performance in relation to financial objectives. Regardless of the size of the business, ratios can help to identify strengths and weaknesses and enable comparison with industry averages. Non-financial and qualitative ratios have also been used to measure performance in other areas and sectors, often discussed in terms of benchmarks. For example, Simatupang and Sridharan (2004) identify a number of collaborative practices that contribute to performance improvement between suppliers and retailers. For the purposes of this research, the unit of analysis is each sustainable tourism objective. In order to compare levels of intended interorganisational collaboration, the ratio is calculated by weighting by the types and frequencies of organisations involved in the joint implementation of objectives. A more detailed description of the method of ratio calculation is provided in the findings.Stage 3 of the methodology is an analysis of the management mechanisms detailed in each plan that would facilitate and encourage collaboration. Comparison of these across sites will give some indication of the level of commitment towards collaboration.\nFindings: Stage 1: profile and frequency of interorganisational collaboration\nConclusions: It is clear from the content analysis of sustainable tourism objectives in the World Heritage Site Management Plans that commitment to collaboration is wide-ranging among the 12 sites examined. A major weakness of the study is that it is based on secondary data and this places several limitations on the research. Firstly, it is not possible to deduce from the management plans how meaningful, or indeed how fruitful interorganisational collaboration will be, at those sites that are committed to it. Secondly, the analysis has focussed on the intentions stated in the management plans, which could be quite different to what is implemented in reality. Thirdly, the methodology, which compares collaborative intention using ratio analysis, is one-dimensional and would need to be supported by more extensive fieldwork, as a ratio does not provide any detail on the nature of the true nature of collaborative activity.Without being too critical about the quality of management plans, it must, however, be taken into consideration that the concept of collaboration is still relatively new to the heritage and tourism sectors, and that these plans represent small steps on an infinite journey of management (sites of universal significance will be preserved for all time) that, in effect, only started in 2005. What the content analysis reveals is that the first stages of effective collaboration have been set in motion: various organisations have been brought together (Timothy, 1999), coordination and cooperation is occurring (Ladkin and Bertramini, 2002) and structures are in place (Table VII), albeit along a wide spectrum of activity (the management plan for Fountains Abbey has only two sustainable tourism objectives that will be jointly implemented whereas the City of Bath has 19; see Table IV). In terms of the interface between WHSs and the \"interorganisational domain\", six key players emerged as important stakeholders in collaborative management:1. local authorities;2. tourist boards;3. education bodies;4. community representatives;5. private sector organisations; and6. transport organisations.However, the data has shown that there is much more scope for collaboration between WHSs and these sectors. Arguably it is only once the plans have been published that comparison and evaluation can be made among sites. The fact that the documents will be in the public domain will hopefully inspire a desire to identify and emulate best practice.Reasons for lack of IOR collaboration among the UK's WHSs\n",
"role": "user"
},
{
"content": "Purpose: - This study aims to identify the extent of intended interorganisational collaboration in the implementation of management objectives at World Heritage Sites (WHSs). These sites represent a unique management challenge in terms of the number and diversity of organisations involved both in terms of ownership and in the interface between the site and the wider tourism environment.\nDesign/methodology/approach: - The research is based on content analysis of 12 management plans. Analysis focuses on the types and frequency of organisations involved in the joint implementation of sustainable tourism objectives. Ratio analysis is used to develop a ratio of collaboration which provides a measure of the numbers of organisations involved and the level of collaborative activity across sites.\nFindings: - There is higher commitment towards interorganisational collaboration at complex sites, comprising a range of attractions, compared with single-focus sites. This was unsurprising, given that complex ownership of sites necessitates collaboration. However, when data relating to site owners was eliminated to reveal a truer picture of collaboration, the findings indicated that single-focus sites were committed to jointly implementing a higher proportion of objectives compared with complex sites organisations. A number of mechanisms are in place to facilitate collaboration, but again, these are more prominent in the management plans of complex sites. The findings indicate that interorganisational collaboration is currently underutilised in the strategic management of WHSs.\nResearch limitations/implications: - The study is exploratory, given that this is the first time that World Heritage Sites have had published management plans. Measuring comparative activity using ratio analysis is one-dimensional. The research is also based on intended activity over the next five years and would need to be followed up by further research based on the reality of implemented objectives.\nOriginality/value: - The study assesses the extent of intended interorganisational collaboration in the public-sector context of WHSs and contributes to the literature in both areas of study.\n",
"role": "assistant"
}
] |
Cooperation and partnering in facilities construction - empirical application of prisoner's dilemma
|
[
"Partnership",
"Game theory",
"Buyer‐supplier relationships",
"Construction industry",
"Sweden"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: During recent years, the interest for partnering arrangements has increased in many countries in the context of construction (Bresnen and Marshall, 2000; Chan et al., 2003; Ng et al., 2002) and facilities management (FM) (Jones, 1995; Okoroh et al., 2001; Roberts, 2001). Partnering arrangements can be divided into two main types; short-term agreements regarding a specific project (project partnering (PP)) or long-term agreements concerning a series of projects or transactions (strategic partnering). Both these types have been suggested to be suitable alternatives in order to transform the traditional adversarial relationships into cooperative ones. Since, traditional contract (TC) forms give little incentive for cooperation (Kadefors, 1997), both practitioners and researchers claim that construction projects often suffer from opportunism and a lack of cooperation between clients and contractors (e.g. contractors' search for extra work and clients' rejections of contractors' suggestions regarding alternative solutions). Alternative project governance forms, like partnering, aim to address problems in resolving disputes and enhance cooperation through mutual goals and effective routines to handle conflicts (Cheng et al., 2000; Barlow et al., 1997).According to Lazar (2000), game theory (GT) and the game prisoner's dilemma (PD) are useful tools to analyze inter-organizational relationships in construction projects as they are very similar to such games. GT is the analysis of rational behavior in situations where decision makers with different goals participate and where interdependence between outcomes is involved. Inter-organizational relationships is a popular research topic, for which GT is well-suited (Camerer, 1991). The PD game is especially suitable to model cooperative behavior in buyer-supplier relationships since it closely resembles the structure of an exchange relationship (Zagare, 1984; Hill, 1990). GT has been criticized for being hard to use and test in real life. GT applications are often purely theoretical and mathematical (Jehiel, 2001), while empirical work is very uncommon (Camerer, 1991). In studies where GT is applied to specific situations, the values of variables (e.g. pay-offs and discount parameters) are often assumed and set by the researcher rather than empirically collected from real life (Reardon and Hasty, 1996). Many game theorists, therefore, call for a more practical brand of GT (Camerer, 1991). Since, the most insight for practitioners would be expected to come from an empirically grounded GT (Camerer, 1991), it was seen as valuable to investigate the implications that specific theoretical applications have in real life contexts (Jehiel, 2001). This paper attempts to achieve a more practical brand of GT, based on empirical data, in order to increase the usefulness of PD as a tool for understanding decision-making behaviors in buyer-supplier relationships.Since, traditional construction relationships are argued to be adversarial it should be interesting to investigate if the opportunistic behavior of clients and contractors in construction projects can be explained by GT and if different contract forms lead to different outcomes in PD games. There are several important game theoretic variables (especially pay-offs and the discount parameter) that influence cooperation. To empirically apply PD, these variables need to be operationalized and measured; the degree to which this was done in earlier research was unsatisfactory. The purpose of this paper is to investigate if a lack of cooperation in construction projects may be explained by game theoretic reasoning, based on the empirical application of PD. To do this, three different contract forms will be analyzed: TC, PP and lasting relationships (LR). In order to make the empirical application possible, a method to operationalize and measure the variables pay-offs and discount parameter is developed. When they have been measured, the empirical data will be used to analyze the players' decision-making behaviors regarding cooperation and defection.\nPrisoner's dilemma: PD is a two-person nonzero-sum game that can be applied to describe the conflict between individual and collective interests in many different political and economical situations, such as inter-organizational relationships (Zagare, 1984). Each player can choose to cooperate or to defect, resulting in one of the four pay-off outcomes (Figure 1). Owing to rational calculations, the players can be caught in a \"catch 22\" situation, i.e. a no-win situation. Each player, in pursuing his own selfish ends, will defect from cooperation even though they are both better off to cooperate. A game is defined as a PD when T>R>P>S and R+R>T+S. T is temptation to defect, R is reward for mutual cooperation, P is punishment for mutual defection, and S is sucker's pay-off.Factors influencing cooperation in PD\nMethodology: This paper aims to develop a method to operationalize and measure the factors pay-offs and discount parameter, so as to make an empirical application of PD possible. In this section, both theoretical and empirical methods are described and discussed.Choice of type of game\nOperationalization and empirical measurement of pay-offs: Empirical data regarding pay-offs were obtained through interview questions that described hypothetical situations of the four possible outcomes in a PD. The amounts of the pay-offs in construction projects depend somewhat on the contract terms. In this paper, projects with three different contract forms are analyzed and described; TC, PP, and LR.Pay-offs for traditional contracts\nOperationalization and empirical measurement of w: The value of the discount parameter (w) mostly depends on the probability to meet again in future games and how high the player values future pay-offs (Axelrod, 1984). w is constituted of several variables, which should be measured on their own. Therefore, the operationalization of w involves creating a formula that contains the most important variables affecting w.Empirical values of variables affecting w\nAnalysis of cumulative values: In TC, client values are R=20, T=23, P=1, and w=0.83. The cumulative value of an infinitely repeated mutual cooperation will then be R/(1-w)=118, and the value for an initial temptation to defect (T) followed by mutual defection (P) will be T+wP/(1-w)=28. Owing to the large importance of future pay-offs, the rational choice for the client is cooperation. Contractor values are R=11, T=14, P=6, and w=0.41. The cumulative value of an infinitely repeated mutual cooperation will then be R/(1-w)=19, and the value for an initial temptation to defect (T) followed by mutual defection (P) will be T+wP/(1-w)=18. Owing to the small importance of future pay-offs, the cumulative values of cooperation and defection are almost equal. Thus, it is difficult for the contractor to make the decision, for which reason their behavior is hard to predict. A rational player will choose cooperation in some situations and defection in others.In PP, client values are R=22, T=23, P=1, and w=0.83. The cumulative value of an infinitely repeated mutual cooperation will then be R/(1-w)=129, and the value for an initial temptation to defect (T) followed by mutual defection (P) will be T+wP/(1-w)=28. Owing to the large importance of future pay-offs, the rational choice for the client is cooperation. Contractor values are R=13, T=14, P=6, and w=0.5. The cumulative value of an infinitely repeated mutual cooperation will then be R/(1-w)=26, and the value for an initial temptation to defect (T) followed by mutual defection (P) will be T+wP/(1-w)=20. The cumulative values of cooperation and defection for PP differ more than in TC. The rational choice for the contractor is cooperation. However, because of the rather small difference (26-20 = 6), it is uncertain if the contractor will always choose cooperation.In LR client values are R=22, T=23, P=1, and w=0.84. The cumulative value of an infinitely repeated mutual cooperation will then be R/(1-w)=138, and the value for an initial temptation to defect (T) followed by mutual defection (P) will be T+wP/(1-w)=28. Thus, the rational choice for the client is cooperation. Contractor values are R=13, T=14, P=6, and w=0.84. The cumulative value of an infinitely repeated mutual cooperation will then be R/(1-w)=81, and the value for an initial temptation to defect (T) followed by mutual defection (P) will be T+wP/(1-w)=46. In LR, cooperation is the rational choice for the contractor since it leads to significantly higher earnings than defection.In Table I, the client's cumulative values for cooperation (R/(1-w)), defection (T+(wP)/(1-w)), and the empirical inputs (R, T, P and w) are summarized, for the three forms of contracts (TC, PP and LR). For all forms of contracts, the cumulative values for cooperation (118-138) are drastically larger than for defection (28), for all forms of contracts. Thus, it is always rational for the client to cooperate.In Table II, the contractor's cumulative values for cooperation (R/(1-w)), defection (T+(wP)/(1-w)), and the empirical inputs (R, T, P and w) are summarized. In TC-projects, the cumulative values for cooperation and defection are rather similar (19 and 18, respectively), rendering the choice of action for the contractor to take not very obvious. In PP-projects, the cumulative value for cooperation (26) is somewhat larger than for defection (20); thus it should be rather rational for the contractor to cooperate. However, because of the small difference, it is uncertain that the contractor will always choose cooperation. Only in LR-projects is the cumulative value for cooperation drastically larger than for defection (81 and 46, respectively). Thus, only in LR-projects is it always obviously rational for the contractor to cooperate.\nConclusion: Eriksson (2001) discovered that players in the construction sector tend to connect different games, since contractors hope to be a part of the client's next project, though the parties have not entered into long-term contracts. In this paper, construction projects correspond to infinitely repeated games of PD. To separate projects with different forms of contracts, different values of pay-offs and the discount parameter are calculated. These two are the most important factors influencing the choice between cooperation and defection, if the players are assumed to be profit maximizing, since these factors affect the cumulative values of pay-offs. To allow empirical measurements of pay-offs and the discount parameter, these factors need to be operationalized. In this paper, a formula for calculating the discount parameter has been created and the factors have been empirically measured through interviews. These values have in turn been used to calculate cumulative values for the two players. Different forms of contracts lead to different cumulative values of cooperation and defection.An infinitely repeated PD is the most favorable form for cooperation to emerge. In spite of this, cooperation is not always an obvious and rational choice for all forms of contracts. According to game theoretic reasoning, defection is rational when the cumulative value for defection is higher than the corresponding value for cooperation. In the empirical example in this paper, the cumulative values for cooperation and defection for the contractor are almost equal, in projects with TC forms (see Table II). This explains why the behavior of the contractor is difficult to predict, since the contractor does not know if the players will meet again in future games. For the client, cumulative values of cooperation are always significantly larger than the corresponding values for defection, for all contract forms (see Table I), given that the client knows that he may play against the contractor again if he wants to.PP, which is often argued as a solution to a lack of cooperation in the construction sector, does not necessarily secure mutual cooperation. According to the empirical example in this paper, the cumulative values of cooperation and defection for the contractor may not differ that drastically in partnering projects, see Table II. Thus, it is uncertain that the contractor will always choose to cooperate in PP. These results support earlier research discussing the short perspective in the construction industry. The strong emphasis on individual projects favors a narrow perspective, both in time and scope (Dubois and Gadde, 2002). Since, relationships are mostly focused on short-term for the duration of one single project, the parties attempt to lever what they can out of the existing contract, resulting in opportunistic behavior in adversarial relationships (Voordijk et al., 2000, Cox and Thompson, 1997). Only in projects within LR is cooperation always the obvious and rational choice for the contractor. This is because long-term contracts both decrease the uncertainty for future games and increase the value of future outcomes. The best way to enhance mutual cooperation is, therefore, to make sure that players meet each other again and again in subsequent projects. This can be obtained by long-term contracts or by using a very small pool of suppliers that are contracted on regular basis.The main contributions of this paper are:* It has been shown that GT and PD when it is operationalized allows for significant insights into the nature of partnering in construction and FM.* That the factor discount parameter, rather than the factor length of the game, should be used to distinguish projects (transactions) with different lengths of contracts. According to Hill's (1990) and Brandenburger and Nalebuff's (1996) reasoning about connections between games, this issue is probably not unique for construction projects. Also other types of buyer- supplier relationships within FM probably correspond to this.* It has been shown how the two factors pay-offs and discount parameter can be operationalized and measured in real life transactions, through interviews. To measure the discount parameter a formula comprising of several variables has been created, and empirical data has been used to calculate cumulative values of cooperation and defection.* Furthermore, the paper gives a possible game theoretic explanation for the inconsistent behavior of actors in the Swedish construction sector, caused by lack of long-term contracts. According to game theoretic reasoning, one cannot expect actors to cooperate if this is not rational. Cooperation is probably not always a rational behavior in single construction projects.* To increase the incentives for cooperation the actors should try to decrease the difference between the pay-offs T and R, increase the chance for future games (repeat prob.), and/or increase the profits for LR (LR profits). To do this the actors should adopt a long-term perspective regarding their inter-organizational relationships, facilitating players to meet regularly.Since, the values used are based on empirical data collected from very few respondents, they should be viewed as illustrative empirical examples, rather than statistical generalizations. Future research should explore the possibilities to use quantitative methods to collect the data needed for calculations of the cumulative values. Statistic generalizations could then be accomplished to receive possible explanations of decision-making behavior in large contexts, such as FM.\n",
"role": "user"
},
{
"content": "Purpose: - To investigate if game theoretic reasoning may be used to explain a lack of cooperation in buyer-supplier relationships within construction and facilities management. In order to make an empirical application of the prisoner's dilemma (PD) game, possible important variables are operationalized and empirically measured.\nDesign/methodology/approach: - Empirical data concerning pay-offs and the variables in the discount parameter formula (created in this paper) have been obtained through interviews with clients and contractors in the Swedish construction sector.\nFindings: - This paper suggests a way to operationalize pay-offs and the discount parameter, making empirical measurements possible. Owing to differences in pay-offs and the discount parameter, different forms of contracts will affect cooperation. Cumulative values of cooperation are much higher in lasting relationships than in occasional transactions. Thus, the best way to facilitate cooperation between rational players is long-term contracts.\nResearch limitations/implications: - Since, the values used are based on empirical data collected from a few respondents, they should be viewed as illustrative empirical examples, rather than statistical generalizations.\nPractical implications: - From a game theoretic perspective the practice of project partnering may not solve problems regarding lack of cooperation. To increase the incentives for cooperation, the actors should work together in long-term relationships instead of focusing on single projects. Long-term strategic partnering is, therefore, beneficial for the construction and management of facilities.\nOriginality/value: - This paper makes empirical application of the PD game possible by operationalizing and empirically measuring game theoretic variables that previously have been given values set by the researcher rather than by the players in the game.\n",
"role": "assistant"
}
] |
The incorporation of market orientation in the school culture: An essential aspect of school marketing
|
[
"Market orientation",
"Education",
"Schools",
"Culture"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The research on school marketing has been receiving increasing attention in recent years as more and more researchers have illuminated marketing mechanisms within educational institutions, including schools, colleges and universities (e.g. Bagley et al., 1996; Bell, 1999; Bunnell, 2005; Foskett, 2002; Hemsley-Brown and Oplatka, 2006; James and Phillips, 1995; Oplatka, 2002, 2004, 2006; Oplatka et al., 2002).A common finding across many studies worldwide points to the significant role of a relationship marketing approach (RM) in devising and implementing a marketing plan and strategy in schools (and other educational organizations). A RM approach puts emphasis on nurturing relationships, especially with existing customers, and the development of supportive market networks (Brown et al., 1994). An underlying assumption is that attracting, maintaining and enhancing customer relationships is an important determinant of the customer's overall satisfaction with a service. Forms of RM have been observed in schools worldwide (e.g. Bell, 1999; Oplatka et al., 2002).It is widely held that a RM approach begins with a commitment to marketing orientation (MO) and to developing an organizational culture that is customer-driven, and focuses on the quality of the service (Narver and Slater, 1990). MO frequently underpins the development and implementation of successful relationship marketing strategies in any organization (Helfert et al., 2002). If a school can develop or improve its degree of MO, then it should also be able to achieve improved levels of RM. To the best of our knowledge, the literature on educational marketing has not paid sufficient attention to MO and explored its implications for schools.This paper presents the major features of MO and its benefits for schools; suggests an inventory to measure the degree of MO in a school, and provides strategies to incorporate elements of MO into the school culture. It is based on current theory and research on MO in non-education sectors and on the research on educational marketing.\nMarket orientation: a key element in marketing the school: Many managers today recognize that the ability to succeed in the marketplace requires more than just sales techniques - customers rarely respond to sales pitch, but rather they want their circumstances to be acknowledged and their needs to be satisfied. Successful marketing is now much more targeted, and the meeting of customers' needs is of high value. Hence, business and service organizations seek to achieve a competitive advantage in their dynamic environments, at least in part, by being market-driven, i.e. by anticipating, understanding and responding to the preferences and behaviors of customers (Jaworski et al., 2000).The marketing literature is replete with definitions and perspectives of MO, yet there is much agreement about the key concepts (Harris, 2002; Helfert et al., 2002; Kohli et al., 1993; Narver and Slater, 1990). At the core of this concept is the significance of customer orientation. Accordingly, customers' needs, desires and particular circumstances, e.g. lifestyles ought to be the main focus of the market-oriented organization. In this sense, MO is the degree to which an organization generates and uses intelligence about the current and future needs of customers; develops a strategy to satisfy these needs; and implements that strategy to meet those needs and wants.MO takes into account the influence of competitors and incorporates inter-functional coordination. It encourages the generation of intelligence - or the use of data about competitors, and integrated cross-functional processes, in addition to the execution of a strategic organizational response to market opportunities. All these activities are directed towards creating and satisfying customers through continuous needs-assessment.In the school context, we believe that a focus on current and prospective customers (parents, students) should take precedence over the two other functions of MO. In our view, MO is a set of beliefs that puts customers' interests first, but at the same time raises the school's awareness of the need to obtain information about competitors and establish cross-departmental activities to satisfy customers' needs, in order to gain a competitive edge in the turbulent, competitive environment.MO as an aspect of organizational culture\nA measure of market orientation in the school: In light of several measures of MO suggested in the business literature (e.g., Hajjat, 2002; Harris, 2002; Helfert et al., 2002; Kohli et al., 1993), a measure of MO in schools which relies on reports of the perceptions of school staff is suggested here. Our aim is to provide school management with a means of measuring the degree of MO in their school, to provide a baseline assessment - crucial information that would be required if school management decided to enhance the dimension of MO in the school culture.In our view, MO is a continuum rather than a dichotomous concept. It has three components: customer orientation, competitor orientation and inter-functional coordination. Based on this conceptualization, the MO Inventory is a 31-item measure of MO in a school and is presented here to be used by school management for identifying and measuring the extent to which the staff in the school behave in accordance with the principles of MO.The first component of MO is \"customer orientation\" which comprises 20 items which measure the extent to which staff in the school (including: managers; middle managers; teachers; administrators etc.) focus on customers' needs, desires and concerns. These items are measured using a 6-point Stapel Scale (resulting in a possible total score ranging from -60 to +60 on an interval scale) (see Table I).Using a similar approach the perceptions of competitor schools, by members of the schools staff, can also be measured. \"Competitor orientation\", the second component of MO, is often neglected in educational institutions, chiefly because educators sometimes perceive it to be incompatible with moral values of education (Oplatka and Hemsley-Brown, 2004). The next 7 items aim to measure the extent to which members of staff within the school are aware of the strengths and weaknesses of competitors and how far they believe that this information could contribute to the marketing of the school (scores range from -21 to +21) (see Table II).The final component of MO - \"interfunctional coordination\" within a school, also needs to be measured. Here, the behavioral expressions of inter-departmental coordination of marketing activities, as well as staff's beliefs about their responsibility for school marketing and customer orientation, are measured (scores range from -21 to +21) (see Table III).How should the answers be analyzed?\nIncorporating market orientation into the school culture: One approach to linking MO as a philosophy to more practical issues is to look at it as an aspect of organizational culture. Culture refers to the collection of deep-seated, unwritten system of shared values, attitudes, beliefs, ways of thinking, artifacts and norms within a school, which dictate its policies and practices (Kluchhon, 1951). Further, organizational culture usually involves the goals of the school, the preferred means to reach the goals, and a set of principles which maintain the identity and integrity of the school. A marketing culture is one in which the philosophical principles of MO rule, in policy and in practice. In other words, this culture reflects an enthusiastic service orientation, high awareness to competition, and acceptance of inter-functional coordination.Homburg and Pflesser (2000) conceptualize market oriented organizational culture as a construct including four components. We describe here the first three of them and match them to the school culture:1. Organization-wide shared basic values supporting MO. At the core of the market oriented school culture is the emphasis given to customer service (process oriented) rather than to academic achievements (results oriented) exclusively. A shared set of beliefs and values in market-oriented school centers around organizational pride, customer service, service quality, innovation, open-minded, sense of mission, entrepreneurship, outperforming competition, striving for excellence. In particular, schools sharing the values of an open internal communication are more likely to be market-oriented, because market information is not kept by the school management but is disseminated across the school.2. Organization-wide norms for MO that provide links between the value-level and the practice one could be the extension of services on offer; decentralized decision making which enables fast market oriented processes; coordination of inter-departmental activities, or inter-departmental implementation of innovations, which is the key for adding value for customers. A cultural value of openness goes along with knowing the customer well, doing what is necessary for accomplishing the organization's customer friendliness and service, learning from mistakes made, being attentive to complaints and considering one's own work as important in accomplishing the school's service.3. The market-oriented culture is embedded with perceptible artifacts that promote the MO among school staff. For example, stories about exceptional behaviors of a teacher with a strong service commitment, welcoming reception area, special events for parents at key points in the calendar, or regular awards for customer-oriented teachers. The language used in the staff room of market-oriented schools includes sentences and words such as: \"in our school we do care for the well-being of students\", \"satisfied parents promote learning and school image\", or \"we are attentive to the personal distress of students\" (Oplatka et al., 2002).Designing an effective approach to cultural change\nConclusion: This paper sets the stage for studying the aspects of MO in schools and related educational institutions and for suggesting practical strategies to incorporate MO culture in this sort of organization. While school management has been encouraged to adopt MO and change the school culture accordingly, researchers in the field of educational management will find MO a useful concept for increasing our understanding of school-environment relationships, on one hand, and of cultural aspects of the school organization, on the other.Subsequent research on MO in educational institution should begin with an enquiry of the degree of MO in schools, colleges and universities, as well as the particular antecedence and consequences of MO in this kind of organization. There is a great need for research which explores the influence of school principals and college managers on the incorporation of MO in these institutions, and teachers' responses to the principal's attempts in this respect.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to present the major features of market orientation (MO) and its benefits for schools, suggests an inventory to measure the degree of MO in a school, and provides strategies to incorporate elements of MO into the school culture.\nDesign/methodology/approach: - An instructional, technical approach which is based on empirical literature both from business and service marketing and from the emergent research on educational marketing is taken in this article.\nFindings: - The paper analyzes the implications of MO for the management of school-environment relations, and provides an inventory to measure the degree of MO in individual schools. In addition, a stage by stage approach to incorporating MO into the school culture is broadly discussed, with a focus on the principal's key role in this process.\nPractical implications: - The paper concludes by suggesting some implications for future research on MO in schools and other educational institutions and highlights the significance of MO for our understanding of school marketing in the era of competition and choice.\nOriginality/value: - As MO frequently underpins the development and implementation of successful organization-environment relationships, the current paper is a first attempt to help principals and administrators incorporate MO into their school, thereby capitalizing on the great advantages of market-oriented organizational cultures.\n",
"role": "assistant"
}
] |
Exploring reduced global standards-based inter-organisational information technology adoption
|
[
"Survey",
"Information management",
"Collaboration",
"Case study",
"Supply chain management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Most scholars support the idea that integrated supply chains hold the key to operational excellence, and that information technology (IT) can play a crucial role in facilitating the integration between supply chain partners (Wiengarten et al., 2013; Rai et al., 2006; Vickery et al., 2003). Drawn by the promise of improved performance through the integration of supply chain partners and processes, firms make regular investments in inter-organisational IT (Cooper et al., 1997; Bagchi and Skjoett-Larsen, 2003; Mortensen and Lemoine, 2008). Scholars typically argue that the more firms possess the \"right\" predictors, the more they operate under suitable circumstances, in other words, the more firms will use inter-organisational IT to benefit their supply chains (Jeyaraj et al., 2006; Fichman, 2004). This research paradigm is largely supported by progressive, stage-based models of technology diffusion that account for linear adoption patterns (Ranganathan et al., 2004). At the same time, evidence that investments in inter-organisational IT lead to significant increases in firm performance is inconsistent (Bakker et al., 2008; Sriram and Stump, 2004; da Silveira and Cagliano, 2006). Indeed, a handful of studies have questioned the assumption that IT and more integration is desirable and linked to more successful supply chains (e.g. Furneaux and Wade, 2011; Autry et al., 2010; Mortensen and Lemoine, 2008). These contradictory results pose a challenge for theories of supply chain integration and IT adoption and raise two important interrelated research questions:\nBackground and literature review: The basic premise of our research is that firms rely on inter-organisational IT to cooperate with other entities in the supply chain. Scholars generally agree that to obtain the maximum overall benefit from a supply chain, trading partners should be linked electronically (e.g. de Vries and Boonstra, 2012). The potential benefits of these technologies are manifold; Chou et al. (2006) provide an overview of 36 outcome criteria that have been used to evaluate investments in inter-organisational IT. We thus position our study on GS1 standards-based inter-organisational IT in the broader supply chain integration and collaboration literature alongside many other studies in this area, although without limiting ourselves to specific performance criteria (e.g. da Silveira and Cagliano, 2006; Wiengarten et al., 2013; de Leeuw and Fransoo, 2009; Snider et al., 2009; Abdinnour-Helm et al., 2003).\nResearch design: To address the first research question, we begin with a quantitative assessment of changes in the implementation of GS1 standards-based IT between 2001 and 2011. This evaluation allowed us to measure changes in the implementation of these technologies over time, and assess whether firms really move unidirectionally towards ever more IT. Subsequently, to address the second research question, we inductively collected data and performed an empirical study based on cases of manufacturers in a supply chain context.\nPhase 1 (RQ1): survey research: Initial data collection\nPhase 2 (RQ2): multiple case studies: We used case studies for a number of reasons. Case studies helped us to understand the complex processes that shape implementation of these technologies within the identified companies, and to build new theory from practice (McCutcheon and Meredith, 1993). The case study approach was also crucial to provide conceptual insight into the changes identified through the survey research phase. Further, case studies allowed us to get close to data, enabling rich and inductive description (Halinen and Tornroos, 2005), which is necessary to comprehend the complexities associated with implementing GS1 standards-based inter-organisational IT across firm boundaries. Finally, given the time between the two surveys, the case studies helped to control for the many contingent factors that may have affected these firms between 2001 and 2011. We specifically asked respondents about their behaviour in using GS1 technologies (such as whether the GS1 standards-based inter-organisational IT the firm used had changed) not only over the past ten years, but also the past five. Five years is the typical strategic planning horizon for many firms, and an assessment of GS1 standards-based technology implementation at these intervals facilitates our analysis of the changes between T1 and T2 (Barringer and Bluedorn, 1999).\nDiscussion: IT implementation is intended to facilitate and streamline processes in a supply chain, which are often messy, ill defined, and fraught with practical constraints. Scholars generally agree that to obtain the maximum overall benefit from a supply chain, trading partners should be linked electronically (e.g. de Vries and Boonstra, 2012). However, instead of relying on current theories to explore a theoretical, linear trajectory of IT implementation, we adopted a mixed-methods, sequential research design to extend the understanding of standards-based inter-organisational IT implementation over time.\n",
"role": "user"
},
{
"content": "Purpose: The information technology (IT) literature is mixed regarding the benefits of inter-organisational IT, but shifts towards less adoption of IT remain largely an unexplored area of research. The purpose of this paper is to address this gap by shedding light on the question of whether or not such shifts exist, and if so, why they occur.\nDesign/methodology/approach: A mixed method, sequential research design was adopted. A quantitative study based on two cross-sectional surveys conducted in 2001 and 2011 yielded responses from 62 matching firms that had implemented IT based on global GS1 standards. Survey results led to the identification of a group of organisations that provided evidence of reduced standards-based inter-organisation IT over time. Subsequently, four theory-building in-depth case studies explored why changes towards less IT occurred.\nFindings: Results from the quantitative study show that measurable changes take place in IT implementation over time (both increasing and decreasing). The case studies show that changes towards less implementation of IT are largely explained by a complex interaction of multiple factors lending support to a situational model of IT adoption.\nResearch limitations/implications: The empirical investigations were limited to Australian manufacturers. Further studies should extend the generalisability of the findings and study the phenomenon in different contexts, as well as across firms that increased, decreased, and maintained their level of IT over time.\nPractical implications: The study helps managers to identify the IT that best suits their strategic objectives. Further, this research increases managers' understanding of how to better use IT as firms weigh opportunities to invest in new supply chain technologies and as a result may choose to reduce investment as a strategic alternative.\nOriginality/value: The paper provides the first systematic insight into reduced IT standards use and implementation. The authors offer a more granular understanding of why firms may choose to reduce usage of IT over time.\n",
"role": "assistant"
}
] |
The critical success factors for stakeholder management in the restoration of built heritage assets in the UK
|
[
"UK",
"Critical success factors",
"Stakeholder management",
"Built heritage assets",
"Restoration projects"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In his book, Understanding Historic Building Conservation, Michael Forsyth (2008) recounted the pre-history of English architectural history. According to him, the development of architecture in the UK can be traced back to the prehistoric Britain, about 3,500 BC, where ancient monuments - barrows, stone circles, henges, etc. - were built mostly around ditches, forming enclosures used probably as multi-purpose gathering places, combining the functions of livestock pen, trading center, church, feasting area and ceremonial arena. Roman invasion in the first century imposed the Mediterranean style of architecture which was followed by the Saxon buildings and the medieval age buildings of the fifth century comprised mainly of churches. However, a massive adoption of Gothic architecture by churches in Britain between the twelfth and sixteenth century led to the replacement of most medieval buildings. The Tudor and the Elizabethan architecture led the seventeenth and eighteenth centuries while the nineteenth century saw the revival of the Gothic architecture referred to as the Victorian Art and Architecture (Ross, n.d.). Different people perceive the architectural aspect of their past in different ways. An illustration of this situation is seen in the play authored by John Steinbeck. In an excerpt of the play, two women who were being evacuated from their abode, and were disallowed from taking along their artifact, lamented: \"how can we live without our lives? How will we know it's us without our past?\", and the evacuators replied \"No. Leave it. Burn it\" (Steinbeck, 2006, p. 111). Byatt (1990, p. 116) argues that the life of the \"past\" is \"the business of every thinking man and woman.\" Moreover, David Lowenthal (1997, pp. 38-52) outlined the benefits of the past to include familiarity, reaffirmation and validation, identity, guidance, enrichment and escape. This study focuses on the architectural aspect of our past, sometimes referred to as heritage assets.\nLiterature review: Attributes of heritage assets\nMethodology: Research epistemology and approach\nData analysis: Sample characterization\nFindings and discussion: There is yet to be seen any academic work where stakeholder management with respect to restoration projects was specifically discussed. Hence, this research does not seek to directly challenge existing knowledge, but to present CSFs for stakeholder management which could be adapted in restoration projects. The CSFs identified through review of literature were categorized into seven groups using PESTLE technique. Although the collected data were not analyzed under these categories, to explicitly demonstrate how the research findings support the theories from literature, the findings were discussed under the PESTLE categories.\nConclusion: Since no significant research has been done to examine the CSFs for stakeholder management in restoration of built heritage assets (BHAs), this study provides a starting point toward what could become a major area of research interest in the near future by presenting the results of a questionnaire survey to investigate the perception of heritage sector practitioners in the UK on the CSFs for managing stakeholders in restoration projects. A total of 20 CSFs were synthesized in the research and were proven to be reliable using statistical method. Data were collected from restoration practitioners, including project managers, architects, heritage consultants, building surveyors, quantity surveyors, planning officers, engineers and facility managers, all practicing within the UK. Findings of the study show that all the CSFs are regarded as critical by most respondents for stakeholder management in restoration projects. In other explore relationships between the opinions of the respondents on the rankings for the CSFs, responses were grouped based on employer characteristics, employee (or respondent) characteristics and project characteristics. It was found that, irrespective of the groupings, \"Understanding statutory requirements for the specific restoration or renovation project\" and \"Obtaining necessary permits from statutory bodies and local council\" are considered most important for stakeholder management. Respondents across the groups agree that \"Understanding the cultural significance of the heritage built asset to the stakeholders\" is also very important in managing stakeholders in restoration projects. On the other hand, respondents across the groups assigned relatively low rankings to \"Engaging the support of local businesses and communities to whom benefits arising from the project do not relate to heritage or conservation matters.\" Likewise, the four social factors, \"Formulating a clear statement of project missions,\" \"Balancing multiple (stakeholders') priorities,\" \"Collaboratively working with other stakeholders at every stage of the restoration or renovation project\" and \"Formulating appropriate strategies to manage stakeholders\" received relatively low rankings across the groups.\n",
"role": "user"
},
{
"content": "Purpose: Built heritage or historic assets (BHAs) constructed in the pre-nineteenth century in the UK are perceived to have certain characteristics which instill cultural significance in them and have seen them become valuable to the economy of the country. The heritage sector makes significant contributions to the UK economy through provision of tourist attractive sites, construction and servicing of heritage assets, heritage conservation, research, and commercial activities carried out within and around heritage assets. These benefits have seen them draw considerable interests from diverse stakeholders within and outside the heritage sector. Hence, a lot of attention is drawn toward restoration of such assets, from stakeholders of different interests, ranging from advocacies for no alteration to complete alteration of the heritage assets. As with construction projects, conflict of interests amongst stakeholders affect the outcome of restoration projects and the purpose of this paper is to examine the critical success factors (CSFs) for managing the stakeholders to achieve the projects' objectives.\nDesign/methodology/approach: This paper examines the views and experiences of practitioners in the heritage sector who have been involved with BHA restoration projects. A total of 32 CSFs for stakeholder management, obtained through rigorous reviews of literature, were subjected to a severe scrutiny with eight restoration experts to determine the importance of the CSFs in restoration projects. The outcome of the exercise was a modified list of 20 CSFs which were further tested on 52 restoration practitioners in the UK using a structured questionnaire to determine the degree of importance of each of the CSFs in restoration projects and their relationships as perceived by the practitioners.\nFindings: The results of the analyses performed on the data show that most of the CSFs were perceived by restoration practitioners as truly critical and vital for successful management of stakeholders in restoration of BHAs. The results also indicate that there is a strong consensus amongst over 50 percent of the practitioners on the rankings of the CSFs.\nPractical implications: The identified CSFs could be used by the restoration practitioners as a \"road map\" for the development of appropriate solutions for successfully managing stakeholders associated with the promotion and BHAs restoration assets.\nOriginality/value: Although CSFs for stakeholder management in construction have been studied by many scholars, no specific research could be identified prior to this study to have been done in defining the CSFs for stakeholder management in restoration projects.\n",
"role": "assistant"
}
] |
Waste in health information systems: a systematic review
|
[
"Lean",
"Quality improvement",
"Systematic review",
"Waste",
"Health information systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Health information systems (HIS) effectively reduce cost in health care delivery and medical error, and increase clinicians' productivity (Oliver, 2007; Goroll et al., 2009; Westbrook et al., 2009; Devine et al., 2010). However, despite reported benefits of HIS, its impact on patient care is inconsistent (Herrick et al., 2010). Rising doubts regarding positive HIS achievements are attributed to a number of weaknesses. HIS issues include error generation, inability to \"fit for purpose\" and support clinical tasks and work-flow efficacy, integration problems and the use of paper-based systems (Yusof, 2015; Yusof et al., 2012; Paul et al., 2012). The term waste is used to refer to these types of non-value added activities in an information systems (IS) environment (Bell and Orzen, 2010; Hicks, 2007). HIS may be more efficient and capable of producing output required by organisations, clinicians and patients if waste is reduced (Hicks, 2007; Taylor, 1982).\nLean transformation: Lean is a methodology for continuously improving process quality in customer products or services. Process quality is improved by identifying and eliminating waste using Lean tools. Waste refers to any non-added value activity that occurs in processes; including transportation, error, delay, over-processing, over-production, movement and inventory.\nMethods: The systematic review consisted of formulating the research question; developing the research protocol; selecting primary studies, quality assessment and data filtering and synthesis; and report writing. A qualitative approach using the content analysis method was employed.\nResults: Study characteristics\nDiscussion: Selection of studies for systematic review\nConclusions: Lean has been applied in the health care domain but studies on waste identification and removal in HIS are very limited. Waste identification and removal methods are essential, given that consideration of waste is fundamental in Lean implementation. Ten waste categories and their relationships related to HIS were identified; seven of them are originated from traditional waste models, while three were adapted from other domains (over-engineering, confusion and unexploited knowledge). These waste categories may guide future implementation of Lean transformations in HIS by identifying the root cause, waste relationships and overall process mapping. The structured waste categories may subsequently improve the alignment of human, organisational, and technology to eliminate waste and improve HIS efficiency. It was also discovered that the selection of Lean tools should be made based on the timing (top-down approach and sequence in unfolding the problems), purpose and characteristics of waste that need to be removed. Lean intervention has a high potential in improving HIS quality through waste identification and removal. However, the lack of evidence in proving the effectiveness of Lean methods calls for more work. Future work can be conducted by employing hands on searches and searching other databases, conducting empirical studies and including other IS in different domains using a holistic, socio-technical approach.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to discuss a systematic review on waste identification related to health information systems (HIS) in Lean transformation.\nDesign/methodology/approach: A systematic review was conducted on 19 studies to evaluate Lean transformation and tools used to remove waste related to HIS in clinical settings.\nFindings: Ten waste categories were identified, along with their relationships and applications of Lean tool types related to HIS. Different Lean tools were used at the early and final stages of Lean transformation; the tool selection depended on the waste characteristic. Nine studies reported a positive impact from Lean transformation in improving daily work processes. The selection of Lean tools should be made based on the timing, purpose and characteristics of waste to be removed.\nResearch limitations/implications: Overview of waste and its category within HIS and its analysis from socio-technical perspectives enabled the identification of its root cause in a holistic and rigorous manner.\nPractical implications: Understanding waste types, their root cause and review of Lean tools could subsequently lead to the identification of mitigation approach to prevent future error occurrence.\nOriginality/value: Specific waste models for HIS settings are yet to be developed. Hence, the identification of the waste categories could guide future implementation of Lean transformations in HIS settings.\n",
"role": "assistant"
}
] |
The impact of social media on L&D: and what we should be doing about it
|
[
"Social media",
"Technology",
"Learning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: I was caught up in a Twitter storm recently. Someone drew attention to the fact that one of the slides I use in a presentation on neuroscience concerning the link between cognitive functionality and physical exercise had been rubbished by an anonymous Twitter figure who challenged gullible and naive assumptions about neuroscience. The claim was that my image was \"made up,\" i.e. not based on proper data and, therefore, the message it carried was disputed. I pointed out that this was not a random image but one produced by a world-famous cognitive kinesthesiologist at the University of Illinois.\nSo what next?: How do learning and development (L&D) professionals get involved then in social media in a meaningful way that will benefit themselves, their organizations and the employees of that organization? Drawing from my recently published book (Paine, 2014) in which I explore the extent that corporate learning has changed in the past five years, and what the big trends in corporate learning are, I offer ten conclusions about what L&D should do with social media:\nIn conclusion: Every organization wants to build staff who are resilient and can learn for themselves. What they no longer need are staff who are content to wait for the next piece of drip feed knowledge, or feel that the organization somehow has a duty of care to keep each individual relevant and skilled without any input or effort on their part. By embracing social media you put paid to this kind of obsolete model of learning and you create an enabling and facilitating model which shifts the burden of responsibility firmly onto the learner. The learner is then able to build a personal knowledge framework and therefore assure his or her own skill set, usefulness and enduring competence.\n",
"role": "user"
},
{
"content": "Purpose: This article aims to provide a succinct summary of the actions learning and development (L&D) professionals should take to get involved in and make the most of social media, both for themselves and their organizations.\nDesign/methodology/approach: This viewpoint is based on the author's book published in 2014, which explores how corporate learning has changed in the past five years and what the big trends are currently.\nFindings: Social media is here to stay and will continually reinvent itself at perhaps even greater speed. L&D professionals in particular must quickly learn to embrace the technology to assist learners to build a personal knowledge framework and thereby assure their usefulness and enduring competence.\nOriginality/value: By embracing social media you create an enabling and facilitating model, which shifts the burden of responsibility firmly onto the learner.\n",
"role": "assistant"
}
] |
Canadian researchers advise law firms to ensure employees feel empowered to negotiate work-life balance (WLB)
|
[
"Qualitative research",
"Role negotiation",
"Legal profession",
"Work-life balance (WLB)"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Canadian researchers have encouraged employers to promote work-life balance (WLB) by ensuring their employees feel empowered to negotiate their roles, especially with their colleagues. The advice was one of the most important conclusions from qualitative analysis of interviewers about WLB with 42 lawyers at large, law firms.\n",
"role": "user"
},
{
"content": "Purpose: The purpose was to find out how lawyers at high-profile legal firms managed WLB.\nDesign/methodology/approach: The researchers conducted interviews with 42 lawyers at two law firms in a large West Coast city. Both participating law firms focus on corporate law and employ around 100 lawyers. Interviews took place on site over a three-month period. They lasted between 20 minutes and an hour. Questions covered general experience in the profession, as well as balancing work and non-work lives.\nFindings: The answers revealed the tensions between work and non-work experiences. Lawyers were driven to work long hours and expected to respond quickly to clients' needs. But they had diverse attitudes to WLB. They could broadly be divided into three categories - \"work-centric,\" \"non-work centric,\" and \"dual-centric.\" Their life values were also strongly correlated with gender. Only dual-centric and life-centric female lawyers had actively negotiated alternative work arrangements\nOriginality/value: There has been very little qualitative research into workplace attitudes to WLB\n",
"role": "assistant"
}
] |
As learning systems go digital and social, can we keep up?
|
[
"OD",
"Learning",
"Organizational culture",
"Social media"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: It is time to stop debating whether social media has a place in learning or not; that bit is a given. We now need to focus on how you harness it to add value to your learning environment.\nHow did we get started?: Bromford is a UK social enterprise, employing 1,200 colleagues, that inspires people to be their best. Through our range of housing and support services, we help customers and colleagues to achieve ambitions and succeed at reaching personal goals. Whether that is getting in to training, volunteering or employment; owning or renting a home; getting life back on track; or simply living independently - we are here. Sixty thousand people live in our 27,000 homes, and we provide support and employment services to 10,000 people.\nWhat have we learned?: From experience, when it comes to strategy my recommendation would be do not waste time pondering if it is better to build a social media savvy learning culture and capability first or get the infrastructure in place. You are where you are, and this raises a wider question of which is the cart and which is the horse? For established organizations who do not have the luxury of a blank canvass, (we did not) I think it is more important to take stock of what you have and take it from there. I hope the following five sections, which are based on the lessons we have learned, will help you further develop, or perhaps start, your own learning and development (L&D) social media journey.\nImpact on internal relationships. Take stock of the operating landscape [...]: Get close to your Information Technology (IT) crowd - Streaming video is a key element of social learning. Whether it is colleagues uploading their own clips or whether it is instructional video, understand if and how this can be streamed. No matter how excited you get about introducing this element to your learning environment, if your videos are constantly buffering they will not add value and will just annoy the hell out of everyone, detracting from what you are trying to achieve. You will need to understand the IT strategy for things like fire walls and the plans for devices too.\nImpact on working with suppliers: Procurement processes will have a cost/quality equation. To get a great social learning system, you will need to put a high value on the less tangible things such as flexibility and collaboration.\nImpact on leadership development: Do not forget your own back yard. All leadership development activity should include how your leaders take responsibility for supporting a digital, social and self-service world. We are very clear here that our leaders need to play their role when it comes to developing a social learning culture. Not forgetting onboarding and induction programmes too - what are they saying is expected of our people? Our CEO says if you do not embrace social media now you would not survive in a leadership role before too long. This may seem harsh but it is a reality that needs to be reflected in learning and development activity.\nImpact on what we measure: If the old adage is \"what gets measured gets done\" then make sure that your strategy includes the right measures. It is no longer relevant or effective to focus on how much time you invest on average learning days or the equivalent monetary value for each colleague; new metrics on outputs are required. These are still evolving for us because we want to equate to our bottom line but we have started to measure things like: how job ready our people are, how quickly they become job ready, time saved comparing how we used to deliver things to how we do now.\nImpact on wider organizational culture: So tactically where has this led us so far? We have an online learning platform that holds \"job ready\" pathways which for individual roles link out to learning interventions. The interventions are a mix of social learning through an online peer-to-peer learning community, individual e-learning modules, online activities to prepare people for workshop sessions and workshop booking. Rarely do people show any love for a system but we have actually received some fan mail!\n",
"role": "user"
},
{
"content": "Purpose: This case study aims to show how one award winning organization, Bromford in the UK, has engaged with and implemented multiple social media strategies that have positively impacted the very fabric and culture of the organization.\nDesign/methodology/approach: The author shares the lessons learned in Bromford's social media journey so far - a journey that has resulted in numerous social media awards.\nFindings: The article describes the impact that social media has had on the organization in five areas: internal relationships; working with suppliers; leadership development; what the organization now measures; and organizational culture.\nOriginality/value: This paper will be of value to all directors, leaders, managers and learning and development (L&D) professionals who are grappling with the rapid changes brought about by the ever-moving social media landscape. The experiences and advice of the author are shared in an accessible and practical way.\n",
"role": "assistant"
}
] |
The market reaction to the appointment of outside directors: An analysis of the interaction between the agency problem and the affiliation of directors
|
[
"Boards of Directors",
"Market surveys"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nI. Introduction: The separation of the ownership of a company from the control of its resources provides the potential for a conflict of interest between owners (principals) and managers (agents); the managers may pursue objectives that are not compatible with those of the shareholders. This conflict of interest is termed the \"agency conflict\"; Jensen and Meckling (1976) apply agency theory to modern corporations and outline the associated agency costs. The finance literature identifies a number of governance mechanisms - both internal and external - that are used to align managers' objectives with those of the shareholders. Some of these important mechanisms include monitoring by board of directors, executive compensation schemes, ownership-related arrangements, the legal system, the market for corporate control and a firm's debt policy. Shleifer and Vishny (1997), Allen and Gale (1999), Denis and McConnell (2003) and Roe (2004) provide lengthy discussions on a number of these governance mechanisms.The board of directors is considered to be the first line of defence against the misappropriation of wealth by management within an agency setting. However, the effectiveness of boards, when they are dominated by inside (executive) directors, has been a controversial issue in the academic literature. These executive directors are considered part of the agency conflict because they take positions inside the company and thus become a group of people who need to be monitored. Also, management poses a strong influence on the selection process of the board. Therefore, the independence of corporate boards in decision making has been questioned. It is argued that boards which have control rights over corporate resources often abuse them (Jensen, 1986, 1993; Berkovitch and Narayanan, 1993). Outside (non-executive) directors are often used as a device to increase the independence of the board (Shleifer and Vishny, 1997). Their presence is thought to improve the effectiveness of internal controls (Mace, 1971; Weisbach, 1988).This study investigates how the market responds to the appointment of outside directors using a sample of these appointments by New Zealand companies. Section II of the paper discusses the monitoring role of outside directors in corporate boards. Section III explains the appointment of outside directors in the New Zealand corporate environment. Section IV describes the sample selection procedure. Section V reports findings of the study. The last section offers some conclusions.\nII. The role of outside directors in corporate governance: Fama (1980) and Fama and Jensen (1983) argue that the outside directors perform a monitoring role better than their inside counterparts. Boeker (1992) finds that boards dominated by non-executive directors are more likely to remove the CEOs if their entities are performing poorly while Perry and Shivdasani (2001) find that outsider-dominated boards are more likely to initiate restructuring programmes following a significant performance decline. According to Bhagat et al. (1987), outside directors are motivated to fulfil their duties because of a desire to maintain their reputations and a fear of shareholder actions (such as lawsuits) if they fail to monitor the behaviour of managers.There are also a number of arguments against the effectiveness of outside directors. Zahra and Pearce (1989) propose that the effectiveness of outside directors is constrained by factors such as a lack of time and the absence of an appropriate level of expertise. Lorsch and MacIver (1989) assert that non-executive directors fear challenging difficult decisions made by management and hence the presence of outside directors does not contribute to corporate performance. Jensen (1993) sees the governance role of non-executive directors as ambiguous; not only is the management able to interfere with the selection procedure of the board but the motivation of non-executive directors to uncover unethical behaviour of management is weak since they hold smaller equity positions than their executive director counterparts. According to Coughlan and Schmidt (1985), the agency problem cannot be solved by having outside directors in corporate boards because managers may withhold crucial information from directors. Furthermore, under the pressure of performance evaluation from outside directors, managers may become risk averse and may concentrate on short-term targets at the expense of strategic investments that benefit their companies in the long run. Baysinger and Hoskisson (1990) argue that the dominance of outside directors limits the ability of the board to contribute to the strategic direction of the firm.The value of outside directors, if any, should be reflected in share prices at the time of their appointments. Rosenstein and Wyatt (1990) analysed a sample of 1,251 appointments of outside directors made by US companies over the period from 1981 to 1985. Their study revealed a statistically significant positive abnormal return of 0.13 per cent during the two-day announcement period. When regression analysis was used including the characteristics of these directors, the authors uncovered positive coefficients for all three types of groups studied (i.e. corporate outsiders, financial outsiders and neutral outsiders); the coefficients for financial and neutral outsiders being significant. Thus, the study indicated that investors perceived the appointment of outside directors as a valuable addition to corporate boards.A legitimate question to be asked is whether the degree of the potential agency conflict influences the market reaction to the appointment of outside directors. This question was addressed in the UK investigation of Lin et al. (2003). They hypothesised that the market response should be conditional on the magnitude of the agency problem faced by the firm and the monitoring incentives of the appointees. They analysed a sample of 714 such appointments made for the period from July 1993 to December 1996 and reported zero cumulative abnormal returns during the three-day announcement period; the finding remained unchanged even if the sample was sub-divided into two groups: confounded and non-confounded announcements. Also, the status of appointees - independent, affiliation with high monitoring incentives and affiliation with management - did not seem to influence the size or significance of any abnormal returns earned; all three sub groups earned insignificant abnormal returns during the announcement period. However, when they regressed cumulative abnormal returns on board ownership (which proxied for the severity of the agency problem), appointee status and a vector of control variables, the authors uncovered a statistically significant negative coefficient for board ownership. In addition, the interaction between board ownership and appointee status generated positive coefficients with some of them being significant. This study highlighted the importance of integrating the interaction between the agency problems and monitoring incentives when examining the market response to the appointment of outside directors.\nIII. Governance features of New Zealand firms: New Zealand companies have governance features that make an investigation of outside directors interesting. First, New Zealand boards are characterised by a higher percentage of outside directors than their US and UK counterparts. An early study by Fox (1996) revealed that, between 1980 and 1993, outside directors occupied 75 per cent of the positions on an average New Zealand board. This rate has increased to 85 per cent from July 1999 to June 2004 (see Section V)[1]. The comparable figures for the US and UK companies - as reported by Rosenstein and Wyatt (1990) and Lin et al. (2003) - are 43 per cent and 66 per cent respectively. Second, since the New Zealand capital market is small and has a relatively inactive market for corporate control, New Zealand investors seem to rely heavily on (and trust) internal control systems such as boards of directors. According to Korn/Ferry International (2000), no companies in New Zealand have been seriously questioned by their investors at annual general meetings on corporate governance issues. Third, many New Zealand companies have a concentrated institutional ownership and therefore, perhaps New Zealand boards do not need an increased representation of outside directors. In the sample used this study, among the shareholdings of top ten investors, financial and corporate institutional investors held 33.53 per cent and 22.81 per cent of the equity respectively while individual investors owned a minority 7.91 per cent. Under this situation, one would expect New Zealand stock market's reaction to the appointment of outside directors to be less pronounced than in the US and UK markets. Fourth, a strong link between company performance and CEO compensation is not apparent in New Zealand[2]. (See Andjelkovic et al. (2002) and Gunasekarage and Wilkinson (2003) for empirical evidence.) Therefore, the presence of outside directors can be viewed as a useful monitoring device. Fifth, the existing empirical evidence suggests that the fraction of outside directors in New Zealand boards has a significant positive influence on company performance (see Hossain et al., 2001). Finally, in a small country like New Zealand, outside directors might have either financial or business affiliations with the companies that they serve. The Securities Commission, in its handbook on corporate governance, admits that seeking a high level of independence is difficult in New Zealand due to the availability of a small pool of qualified and experienced directors (p. 11). Of particular interest to this study, out of the 517 appointments of outside directors made by New Zealand companies during the period of July 1999-June 2004, 323 appointees had some affiliations with the incumbent management of the firm while another 127 appointees had financial affiliations. It can be questioned whether these individuals can provide an effective monitoring mechanism thereby possibly rendering the appointment of outside directors as value destroying actions in the eyes of outside investors.The objective of this study is to investigate whether investors perceive the appointments of outside directors as value enhancing decisions. In particular, we examine whether any abnormal returns are associated with the degree of agency conflict and the status of the appointees (i.e. whether they have any affiliations with the company or not). If a company has adequate governance mechanisms in place to mitigate agency conflict, these appointments may not have any significance for investors. On the other hand, the incentive that these appointees have for monitoring management may depend on their status; independent and financially affiliated outside directors may have more incentives to monitor management than those who are affiliated with the management. In this setting, any interaction between agency conflict and the appointee status may also be important to investors in determining the value of these appointments. Therefore, in examining the market response to the appointment of outside directors, we explore the issues of whether the market responds to such announcements in an unconditional fashion or whether there is a reaction when controlling for agency problems and an interaction effect between the agency problem and appointee status.\nIV. Sample details: We investigate the appointment of outside directors disclosed by New Zealand companies during the five-year period from July 1999 to June 2004. The data on the company that announced the appointment, the announcement date, information on any affiliation of these directors etc. came from the Investment Research Group (IRG) database. Our initial sample consists of 517 appointments of outside directors. We excluded 54 announcements because of inadequate share price data to estimate the expected returns. For the remaining 463 observations we searched IRG during the 11-day period from five days before to five days after the announcement to uncover whether any confounding events that may contaminate the information contained in these appointments were present[3].Tables I and II provide information on these confounding events. There were 63 announcements of corporate control events such as mergers, acquisitions and investment projects, 61 announcements of changes in managerial positions, 48 announcements of other company news such as the issue of annual reports, the publication of earnings and the release of dividend news and 31 announcements of changes in executive directors. Further 41 observations were also excluded to arrive at the clean sample as these announcements were made by financial companies.Previous studies claim that the extent of any affiliation which the directors have with the company that they serve has an influence on the performance of that company because their monitoring incentive depends on the nature of this affiliation (Rosenstein and Wyatt, 1990; McWilliams and Sen, 1997; Perry and Shivdasani, 2001; Lin et al., 2003). Adopting the approaches in these studies, we classify the outside directors into three main groups:1. management affiliated (those who either previously held an executive position within the firm or currently holding a position in another firm which has business affiliations with the appointing company([4];2. financially affiliated (an individual who is (or was) a director, an employee or the owner of a company that could potentially have supplied debt or equity capital to the appointing company[5]; and3. independent (no affiliation). We carefully examined the information provided by IRG and cross-checked it with National Business Review publications and the business sections of newspapers to achieve an objective classification system with respect to the affiliations of outside directors.The incentive to monitor management may depend on the affiliation status of the appointee. Independent directors are the most objective group, but may lack the necessary expertise. Financially affiliated directors have a strong incentive to monitor managers to protect their (or their financial institutions) capital in the company (Easterbrook, 1984; Lin et al., 2003). The management affiliated directors may have the weakest incentive to monitor as they may have strong associations with the team of managers currently employed in the firm, but such disadvantages may be off-set by the contribution that they can make to the company using their firm/industry specific expertise or knowledge.The distribution of the types of directors across the five-year period is provided in Table II. Management affiliated directors make up 62 per cent (61 per cent) of the full (non-contaminated) sample and dominate the number of appointments in each year. Financially affiliated directors account for 25 per cent (27 per cent) and the remaining 13 per cent (12 per cent) are made up of independent directors. Thus, the majority of the appointments are not independent directors. The sample shows a consistent distribution across the five-year period.\nV. Empirical findings: The unconditional market reaction to the appointment of outside directors\nVI. Conclusion: This study examines the response of New Zealand investors to the news of an appointment of outside directors to company boards. It employs the conventional event study methodology and generates event-period abnormal returns using two specifications of expected return. The market does not respond unconditionally to these appointments; the event period cumulative abnormal returns remain insignificant for both full and non-contaminated samples irrespective of the approach followed to generate them. A sub-classification of the sample into three groups as independent, financially affiliated and management affiliated directors did not provide strong evidence that the abnormal returns across these three groups differ. However, in New Zealand, the appointment of management affiliated directors seems to be associated with significant abnormal returns than the appointment of independent or financially affiliated directors.The interaction effect between appointee status and the agency problem does not have any influence on the market response to the appointment of outside directors. Instead, all three types of affiliations (independent, financially affiliated and management affiliated) remained influential in generating abnormal returns when the effects of the severity of the agency problem and the other governance characteristics are taken into account. Board composition emerged as the most powerful governance mechanism in this New Zealand study. This coefficient was negative and significant in all the regression models estimated.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this study is to examine the character of any market response to the appointment of outside directors. The main propositions tested are: whether the stock market responds unconditionally to these appointments or whether the market response is conditional on the degree of the agency problem faced by the firm and the affiliation of the appointees.\nDesign/methodology/approach: - The authors use a New Zealand sample of the appointments of outside directors during the period from July 1999 to June 2004. The unconditional market response is examined analysing the abnormal returns generated by the appointing companies during the three-day announcement period. The influences of the agency problem and the affiliation of directors are tested by employing multiple regressions.\nFindings: - The findings provide strong support for the second proposition; the market considers the degree of the agency problem faced by the firm and the affiliation of outside directors in responding to these appointments. The percentage of outside directors in the board emerged as the strongest governance mechanism which, together with firm size, posed a significant inverse influence on announcement period abnormal returns. A strong interaction effect between appointee status and the agency problem was not present.\nOriginality/value: - The mere appointment of outside directors may not please the firm's investors. Such appointments are more useful for companies with severe agency conflicts; even if such a conflict is present, the affiliations that these outside directors have with the executives and the operations of the appointing companies may need to be considered in determining the value of such appointments.\n",
"role": "assistant"
}
] |
The financial impact of a withdrawn ISO 9001 certificate
|
[
"ISO 9000",
"ISO 9001",
"Event study",
"Certificate withdrawal",
"Decertification",
"Loss of certificate"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: ISO 9001 is a coherent set of standards developed in 1987 by the International Organization for Standardization (ISO) with the purpose of facilitating international trade (Marquardt, 1999). The principles of ISO 9001 were derived from Total Quality Management, a broader and more demanding quality philosophy than ISO 9001. The less demanding principles and standards of ISO 9001 allowed it to become an easier starting point (Gotzamani and Tsiotras, 2001; Najmi and Kehoe, 2001) for companies that want to develop a standard (undifferentiated) quality system (Conti, 2004; Boiral, 2011). These characteristics of ISO 9001 also contributed to make it the most popular quality management system in the world (Viadiu et al., 2006). Indeed, ISO 9001 standards can be used as guidelines by organizations of any size, in any industry and in any country (Yung, 1997; Marquardt, 1999; Boiral, 2011). Organizations may choose to simply follow the standards, without any formal external supervision, or they may choose to have their quality system audited by an independent agency and, by doing so, they may become ISO registered companies (Boiral, 2011). Many organizations choose to have their quality systems registered, because \"of the perception that an independent confirmation of conformity adds value\" (ISO, 2012). In fact, most of the studies evaluating the impact of ISO 9001 conclude that certification has a significant effect on performance (e.g. Chow-Chua et al., 2003; Tzelepis et al., 2006; Chatzoglou et al., 2015). Similarly, the majority of studies suggest that certification can be a source of competitive advantage, with only a limited number of studies concluding the opposite (e.g. Conti, 2004; Casadesus and Karapetrovic, 2005).\n2. Literature review: 2.1. Previous studies on ISO 9001 and the financial impact of certification\n3. Methodology: 3.1. Data and summary statistics\n4. Results: Table IV summarizes the results of the main test. In Panel A, ROA is employed as the performance measure and its dynamics is tracked over a number of periods. To give an example, when \"Initial Year\" is 2007 and \"Final Year\" is 2008 (i.e. the first row of Panel A of Table IV), the abnormal performance is computed for the sample firms in 2008 when the base year is 2007. As can be seen, the mean (median) abnormal performance value computed for such period is 0.5 percent (0.8 percent). These results suggest that the actual performance of the sample firms (as measured by ROA) is marginally higher than that one should expect in the period under scrutiny. Yet, the reported abnormal performance is not statistically significant at conventional levels as the p-value of the Wilcoxon signed-rank test (i.e. this paper's main statistical test) and the t-test for this period is always higher than 10 percent. As such, the statistical evidence suggests that in the period that immediately precedes the loss of ISO certification, the sample firms do not exhibit an especially high or low performance as compared to their expected performance. Expanding the analysis to the remaining entries of Panel A of Table IV is important since it sheds light on what happens once the sample firms lose their certification. As can be seen, none of the Wilcoxon signed-rank tests (or the parametric t-tests) are significant at the 10 percent level in the post-event periods.\n5. Robustness tests: The main results presented on Section 4 are computed using the methodological choices that Barber and Lyon (1996) show yield well-specified test statistics in all situations they consider. Yet, it is important to verify to what extent the previous findings are robust to methodological changes related to the selection of the control firms.\n6. Discussion and conclusion: Previous studies suggest that, in general, there is a positive and significant impact of ISO 9001 certification on the financial performance of firms, and that such benefits are likely to persist or accrue for several years (Corbett et al., 2005; Heras et al., 2002a). However, despite ISO 9001 having become the most popular quality management system in the world, the number of decertified firms has increased considerably in recent years. Yet, to the best of the authors' knowledge, no study has addressed the consequences of an ISO 9001 certificate withdrawal. This paper starts shedding light on this issue by exploring the abnormal financial performance of 143 Portuguese firms that lost their ISO 9001 certification in 2008. Results show that decertified firms do not perform differently from their matched counterparts in the post-event period. Therefore, this paper reveals that, contrary to certification, ISO 9001 decertification does not lead to any significant financial abnormal performance.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to assess to what extent the loss of the ISO 9001 certification affects the decertified firms' financial performance.\nDesign/methodology/approach: Using standard event study methods, this paper matches a sample of 143 Portuguese companies that lost their ISO 9001 certification with similar non-event counterpart firms (according to return-on-assets (ROA) and size) and compares the performance of these two groups of firms using financial data collected from the AMADEUS database.\nFindings: Results show no statistical significant differences in the financial performance (as measured by ROA, return-on-sales (ROS) and sales growth) between companies that lost their ISO 9001 certification and their matched firms. Although the literature suggests that certification improves firms' performance and that the benefits of certification may last over long periods of time, this paper's results suggest that, after decertification, companies do not exhibit over or underperformance in their operations vis-a-vis comparable firms that do not undergo the same event.\nOriginality/value: As far as the authors are aware, this is the first study assessing the impact of ISO 9001 certificate withdrawal on the decertified firms' financial performance.\n",
"role": "assistant"
}
] |
Masking the bullwhip effect in retail: the influence of data aggregation
|
[
"Bullwhip effect",
"Data aggregation",
"Hierarchical linear modeling",
"Retail supply chain management",
"Seasonality"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Researchers have long studied how demand variability propagates through supply chains. This work was framed in the context of production smoothing, which is the notion that companies smooth production, relative to sales, by using inventory as a buffer (Holt et al., 1968; Cachon et al., 2007). While a rather simple and intuitive hypothesis, the preponderance of empirical evidence did not seem to support the notion that demand uncertainty is reduced in higher supply chain echelons (Blanchard, 1983; Blinder, 1986; Blinder and Maccini, 1991; Miron and Zeldes, 1988).\nTheory and method: Bullwhip can be the result of both rational and irrational managerial reactions to demand fluctuations (Forrester, 1961; Burbidge, 1991). For example, managers' inability to fully consider the systemic impact of their ordering policies (e.g. cycle times, quantities) as they respond to demand shocks and stock-outs can result in amplified order variance (Kahn, 1987; Naish, 1994; Lee et al., 1997). In addition, attempts to capture logistical scale economies through order batching (Eichenbaum, 1989; Burbidge, 1991), as well as responding to retail promotions (Ettouzani and Yates, 2012) may also result in heightened bullwhip. Finally, bullwhip may also be attributed to overtaxed worker cognitive capabilities (Bruccoleri et al., 2014).\nBullwhip measurement: In our research, we consider a retail distribution network, where a supplier distribution center (DC) fulfills orders from the retailer's DCs (Figure 1). Each retail DC then fulfills orders from a set of retail stores that are clustered within its designated geographic region. Given that bullwhip is defined as \"the phenomenon where orders to the supplier tend to have larger variance than sales\" (Lee et al., 1997, p. 546), we measure the bullwhip effect at the point where product orders are placed by the retailer to the supplier (i.e. retail DC). First, we observe product sales at each DC, which correspond to the aggregate sales of all stores serviced by a particular DC. We let D\nAggregation and seasonality: In the calculation of the bullwhip measures above, it is expected that important aspects, such as data aggregation, seasonality, and the autocorrelation of demand, might influence the results.\nEstimation procedure: Data collection\nAnalysis: Observed bullwhip\nDiscussion: Currently, the bullwhip effect is well-known and its potential effects on firms and supply chains have been widely elucidated. However, the measurement of the bullwhip effect has not received substantial attention in the literature, especially in regards to the effects of data aggregation and seasonality. The findings of this research confirm that data aggregation can mask the extent of the bullwhip effect and that seasonality in the data has a dampening effect, resulting in both research and practical implications.\nLimitations and future research: Although this research empirically studies the effects of data aggregation and seasonality on bullwhip measurement, the nature of the data utilized in the analysis does present some limitations to the generalizability of the results. Such limitations, however, offer a number of relevant opportunities for future research.\n",
"role": "user"
},
{
"content": "Purpose: The accurate measurement of demand variability amplification across different nodes in the supply chain, or \"bullwhip effect,\" is critical for firms to achieve more efficient inventory, production, and ordering planning processes. Building on recent analytical research that suggests that data aggregation tends to mask the bullwhip effect in the retail industry, the purpose of this paper is to empirically investigate whether different patterns of data aggregation influence its measurement.\nDesign/methodology/approach: Utilizing weekly, product-level order and sales data from three product categories of a consumer packaged goods manufacturer, the study uses hierarchical linear modeling to empirically test the effects of data aggregation on different measures of bullwhip.\nFindings: The authors findings lend strong support to the masking effect of aggregating sales and order data along product-location and temporal dimensions, as well as the dampening effect of seasonality on the measurement of the bullwhip effect.\nResearch limitations/implications: These findings indicate that inconsistencies found in the literature may be due to measurement aggregation and statistical techniques, both of which should be applied with care by academics and practitioners in order to preserve the fidelity of their analyses.\nOriginality/value: Using product-weekly level data that cover both seasonal and non-seasonal demand, this study is the first, to the author's knowledge, to systematically aggregate data up to category and monthly levels to empirically examine the impact of data aggregation and seasonality on bullwhip measurement.\n",
"role": "assistant"
}
] |
High or low season? Contrasting launch timing considerations for big-budget and low-budget entertainment products
|
[
"Seasonality",
"Small firms",
"New product launch",
"Film industry"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The timing of new product release affects new product performance (Calantone et al., 2010; Ottum, 1996) and represents an important element of a firm's new product commercialization strategy. Accordingly, business scholars have sought to address various aspects of launch timing, including first- versus second-mover strategies (Miles and Snow, 1978), release timing into emerging markets (Tegarden et al., 2000), launch timing using real-options (Miller and Folta, 2002), how input cost fluctuations may influence the launch-timing-performance relationship (Fisch and Ross, 2014) and the organizational variables that may impact launch timing in general (Srivastava and Lee, 2005).\nBackground and theoretical design: Seasonality in the entertainment industry\nEmpirical investigation: We test our hypotheses using data on small independent films, which are typically new products unfettered by the restrictive conventions that shape studio productions. Independents compete against well-resourced Hollywood studios (henceforth, \"the studios\"), which focus increasingly on mega-productions costing over $100 million (Elberse, 2013, p. 18). Independent films are produced on a limited budget of about $10 million on average, and they face more constrained commercialization and distribution options (McDonald and Wasko, 2008). Importantly, independent filmmakers do not currently display clear strategic behavior in favor of high- or low-season releases.\nResults: Table II shows the descriptive statistics for our focal sample of independent films and the comparison samples of big- and low-budget studio releases. As expected, independent films and big-budget studio productions differ substantially on all the measures. More surprising, low-budget studio films not only share characteristics of the other two categories but also display unique features. In particular, they appear to have lower quality on average than indie releases, with 36 per cent of critics rating them positively versus 60 per cent for indie releases. In this respect, they are similar to studios' big-budget productions, which garner 38 per cent of positive critic ratings on average. Low-budget studio films differ notably in achieving 10 times higher distribution intensity at the opening, on average, and spending 10 times more on prerelease advertising than independent films, although these numbers are two to three times smaller than for big-budget studio films. This points to a distinct marketing strategy on the part of low-budget studio films, and justifies our treatment of them as a separate film category.\nDiscussion and conclusion: Timing new product introductions with respect to seasonality presents a difficult dilemma for producers in short-life-cycle industries. On the one hand, high seasons are characterized by greatly amplified consumer demand with the promise of positive spillovers for all new offerings. On the other hand, large producers launch and relaunch in high seasons their a priori strongest products supported by intensified advertising. Unfortunately, there are no published practitioner accounts or scholarly research to help small producers specifically with seasonal launch timing decisions.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to investigate the moderating impacts of seasonality on the effectiveness of new product commercialization strategies in short-lifecycle markets. The authors contextualize their theory in the vast and culturally significant entertainment industry sector and contrast the effects between independent films and big budget movies.\nDesign/methodology/approach: This study uses an econometric modeling approach.\nFindings: This study finds that unlike new films by well-resourced studios, which must launch in a high season for best performance, independents can generate more revenue in low seasons under certain conditions. The study shows how seasonality moderates the effectiveness of new films' commercialization strategies and how new product outcomes are different for small independent products than for big-budget productions with regards to distribution duration, advertising expenditure and product characteristics.\nResearch limitations/implications: This research extends the literature on launch timing, which examines various strategic tradeoffs. In contrast with the few extant studies whose concern is sensitizing to the effects of seasonality (Siqueiraet al., 2016), this research treats seasonality as an exploitable opportunity that can be strategically factored into business planning for small producers. Accordingly, this is the first study to theoretically and empirically investigate the moderating relationship between seasonality, marketing decisions, product characteristics and performance.\nPractical implications: To achieve useful specificity, the study constructs its discussion around the highly seasonal entertainment industry sector. The study shows that seasonality moderates the effectiveness of new films' commercialization decisions and that the strategic outcomes are different for small independent products than for major studio productions in particular.\nOriginality/value: In contrast with extant research whose concern is sensitizing to the effects of seasonality, our research treats seasonality as an exploitable opportunity that can be strategically factored into business planning. Accordingly, ours is the first study to theoretically and empirically investigate the moderating relationship between seasonality, marketing decisions, product characteristics and performance.\n",
"role": "assistant"
}
] |
Business process improvement in services: case studies of financial institutions in Thailand
|
[
"Thailand",
"Financial institutions",
"Customer services quality",
"Six Sigma",
"Operations improvement",
"Quality management",
"Service quality"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: In spite of recent problems, the financial services sector continues to play a key role in the global economy, both in the developed world and increasingly, in emerging economies. Competition in financial services has increased, with higher customer expectations, new technologies, business models and service offerings. Operational efficiency, low cost and customer satisfaction (CS) are important goals, to stay competitive and sustain long-term profitability. Hence, financial services has emerged as an important area of business process improvement (BPI) application in both developing and developed economies, aiming to improve efficiency, reduce operating cost, and deliver world-class customer service (Antony, 2004b).Established BPI approaches such as total quality management (TQM), business process reengineering (BPR), Lean and Six Sigma were successfully adopted many years ago by manufacturing industry; and more recently have become popular in the service sector. There is some literature in this area (Allway and Corbett, 2002; Swank, 2003), but research related to BPI adoption in this sector remains relatively limited, and tends to lack consideration of employee and customer perspectives.Although key scholars have long suggested the importance of service quality as a prerequisite for satisfying customers (Cronin and Taylor, 1992; Rust and Oliver, 1994), the linkage between BPI and CS in the literature remains unclear. Several studies have examined the nature of this linkage (Crosby, 1991; Boaden and Dale, 1993; Mehra and Ranganathan, 2008), but have not provided a complete or coherent formulation of the key causal relationships. This suggests a need for continuing endeavour to explore the relationships between BPI initiatives and CS, considering the importance of service quality principles.Sidorova and Isik (2010) examined BP research from a multidisciplinary perspective, analysing the abstracts of more than 2,700 academic articles. They emphasised the role of BPI in the achievement of organisational goals, including CS, and recommended that the different aspects of BPI be recognised as such, and developed in an integrated manner, leading to longer lasting benefits. They considered that: \"The role of BP researchers is in identifying such synergies and potential conflicts among various approaches and organizational initiatives involving BP change\" (Sidorova and Isik, 2010).In typical business practice, much of the emphasis in BPI adoption has been on efficiency improvement and cost reduction, relatively little in enhancing the quality of service to satisfy customers. The outcomes of initiatives were typically assessed on process (e.g. cycle time, reduction of waste), and cost considerations (Flynn et al., 1994; Yasin and Alavi, 2007). The absence of a systematic and rigorous framework to evaluate the outcomes of BPI initiatives, especially from the customer perspective, is unhelpful for practitioners.The authors considered that there is indeed a research gap and a need for further theory, to explain the process by which BPI initiatives lead to CS. The intended contribution of this paper is to describe the development of a theory model to explain the effects of BPI initiatives towards CS. Such a theory will provide valuable practical guidance in helping managers to:* specify improvement areas that directly impact on customers' needs and expectations; and* assessing BPI outcomes from the customers perspective.\n2 Literature review: 2.1 Service quality and improvement perspectives\n3 Research methodology: 3.1 Case study design and data collection\n4 Empirical findings: This section presents key empirical findings from the case study analysis. These are outlined below, in terms of the main stages of a BPI project to distil the critical findings.4.1 Project initiation\n5 Development of the theory model: An initial theory model was developed in the first phase of the project, then refined and validated using the empirical findings from the other two research phases.Refinement and validation\n6 Discussion and conclusions: The extant theory models in service quality do not provide a clear formulation of the relationships between specific BPI initiatives (e.g. projects) and CS. Our research aimed to help fill this gap, by exploring how Thai financial institutions adopt and conduct BPI, to improve service quality. Based on this study, the authors developed a theory model based on service quality principles, to explain BPI adoption outcomes at the operational level. We have not sought to provide a general theory relating all extant BPI approaches to impacts on service quality and CS. Rather than emphasising improvement in internal efficiency, the service delivery processes, cost, or CS measurement, this model aimed to provide an understanding of how BPI adoption and consequent process improvement impacts on customers' needs and expectations. Our empirical findings highlighted the important role of BPI adoption for improving service quality, in terms of technical and functional features, which are considered positive drivers of BPI impacts towards CS.In order to distinguish and clarify the contribution of the proposed model, it is worth reviewing its purpose, context and content, in comparison with two important service quality improvement theories: the SPC and ROQ models. The purpose of our model was to explain the effects of BPI adoption at the operational level and to provide a basis for the evaluation of outcomes. In this respect it differs from well-established service quality models such as the SPC and ROQ, which aim to explain the broader business effects (e.g. revenue, market share, profitability), rather than considering the importance and results of the specific BPI initiatives and projects that help improve operations processes. The context of the proposed model is that of a financial services organisation seeking to adopt specific improvement approaches at operational (perhaps departmental) level, to improve services and their delivery. In this respect also, it differs from the SPC and ROQ models which present generic high-level cause-and-effect relations at an organisational level. The SPC and ROQ pay little attention to detailed processes, and hence do not greatly assist the manager who wishes to introduce a specific BPI initiative. Neither of these existing models provides (or could provide) the basis for a BPI evaluation framework.The model content comprises the theory elements, overall model structure and the linkages between elements. The authors' appreciation of the different purpose and context of the proposed model led them to set the existing models aside and start afresh. Theory elements were conceptualised and synthesised from the observed phenomena of BPI adoption in Thai financial services, not abstracted from models in the literature. The fact that some elements appear with similar names is unsurprising, and does not indicate derivation from existing models. The structure of the model was designed from first principles and refined using an end-to-end business process viewpoint. For example, the decision to separate the outcomes measures pertaining to internal and external quality dimensions was supported by the typical organisational structure of financial services, which is split between front- and back office operations.Compared with existing theory models, the proposed model focuses on a smaller but more elaborate context: understanding the adoption/implementation of specific BPI initiatives for improving service quality in financial services, rather than attempting to cover the larger scope of organisational service quality improvement. Indeed, the proposed model can be considered complementary to both the SPC and ROQ frameworks, by adding more detailed issues and implications of specific improvement initiatives at the operational level.As with most empirical research, there are limitations and a need for further work. First, the case studies undertaken in this research were drawn from the Thai financial services sector, in which the number of institutions that have adopted BPI methodologies is small. This could impact the potential generalisation of the proposed theory model, to cover financial services in other contexts or indeed other types of service offering. Second, while the primarily qualitative approach employed is valuable for theory building, and specific quantitative efforts at triangulation were made, further quantitative research will be required for complete verification of the proposed linkages. Also regarding future work, an evaluation framework was developed, based on the theory model, which has not been presented in this paper. Further refinement, testing and validation of the evaluation framework are required, and will be undertaken in the next phase of the research. This work aims to provide practitioners with a practical, actionable evaluation framework, with a suitable theory basis, which can be applied to BPI projects in the financial services sector.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to explore how financial institutions adopt business process improvement (BPI) for improving service quality, to enhance customer satisfaction. To explain this adoption, it was necessary to develop a theory to explain the linkages between BPI initiatives and customer satisfaction.\nDesign/methodology/approach: - Case study was used as the research strategy, following the theory-building process suggested by Eisenhardt. In total, three phases of data collection were employed, with expert interviews for theory enhancement and validation.\nFindings: - The authors identified three main stages of BPI adoption. A theory model was developed and refined using the empirical findings, to provide understanding of the outcomes of BPI initiatives. This model is compared with existing service quality models.\nResearch limitations/implications: - The research outcomes represent an extension of existing service quality approaches, to consider the BPI adoption process as well as broader organisational issues. However, data were collected only from Thai financial institutions, which might impact the generalisation of the results.\nPractical implications: - The proposed theory model is developed at an operational level, and specifically aims to provide managers with adoption guidance and a practical foundation for further development of operational-level assessment, assisting a more systematic evaluation of the outcomes of specific BPI initiatives at the project level.\nOriginality/value: - The paper provides empirical evidence of BPI adoption in a financial services context. A theory model is presented based on service quality principles, to help explain BPI adoption outcomes at an operational (e.g. project) level, which provides a different perspective to that of existing service quality models.\n",
"role": "assistant"
}
] |
Tertiary-level internet users' opinions and perceptions of cyberhate
|
[
"Gender",
"Ethics",
"Virtual community",
"Community",
"Human-computer interaction (HCI)",
"Information society"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The Universal Declaration of Human Rights underlines human beings' freedom, equality, dignity, and rights. In today's world, no one can question the rights of all humans to live without being exposed to discrimination, including hate behavior and harassment. Impeding these basic rights because of how a group of people may be viewed by others, or how they view themselves, is discrimination. As a matter of fact from a western European perspective (Bleich, 2011), it is undeniable that hate speech inciting an immediate breach of peace by means of racist, discriminatory, offensive, abusive, and threatening ideas are among the biggest problems challenging social peace throughout the world. There is less agreement, however, as to the content of hate speech and material (Timofeeva, 2003). Cohen-Amalgor (2011) offers an extensive definition of hate speech as bias-motivated, hostile, malicious speech aimed at a person or a group of people because of some of their actual or perceived innate characteristics. The author also labels the discriminatory, intimidating, disapproving, antagonistic, and/or prejudicial attitudes targeting gender, race, religion, ethnicity, color, national origin, disability, or sexual orientation as hateful behavior. Another frequently used definition is the one outlined in the Council of Europe's Committee of Ministers Recommendation 97(20): \"the term 'hate speech' shall be understood as covering all forms of expression which spread, incite, promote or justify racial hatred, xenophobia, anti-Semitism or other forms of hatred based on intolerance, including: intolerance expressed by aggressive nationalism and ethnocentrism, discrimination and hostility against minorities, migrants and people of immigrant origin\" (British Institute of Human Rights, 2012).\n2. Methods: A sectional research design and correlation survey method was adopted throughout the study to examine the opinions and perceptions of tertiary-level internet users regarding cyberhate by descriptively exploring the existing situation from various perspectives.\n3. Results: This section includes the findings in line with the research questions posed in this study. The table below depicts the average points and standard deviation values received from opinion-based statements in the Cyberhate Perception Scale. The Likert scale used within these items was as follows: (0) no idea, (1) definitely no (2), (3), (4), (5), and (6) definitely yes.\n4. Discussion: This study aimed to explore the both the perceptions of the internet users regarding discriminative and hateful content based on identities and the predictors of their perceptions toward cyberhate in line with the research questions posed in the beginning. To achieve this objective, cyberhate perception scale was developed and utilized to gather data in two different countries as the USA and Turkey.\n5. Conclusion: The results of this study indicate that culture, age, SES, political perspective, social media use affect internet users' perceptions regarding cyberhate. The results of this research support the idea that in order to predict socio-psychological and physical outcomes of cyberhate, we should identify how people perceive and react to hateful content and behaviors on the internet. By extending our knowledge of how internet users' perceive cyberhate, the findings of this research provide insights for developing educational programs and instructional materials to raise awareness against cyberhate.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to elaborate the opinions and perceptions of internet users regarding online hate speech, and bring cyberhate to the attention of internet users and policy stakeholders.\nDesign/methodology/approach: A sectional research design and survey method was adopted throughout the study to examine the opinions and perceptions of internet users regarding cyberhate by descriptively exploring the existing situation from various perspectives. The participants of the study were determined by purposive sampling methods to attain maximum variety among internet users who are considered as highly literate in technology use. The data were collected through a personal data form and a survey (Cyberhate Perception Scale) from 372 internet users living in Turkey and the USA.\nFindings: The findings of the study revealed that the majority of participants do not perceive cyberhate as a part of freedom of speech and they believe that online hate behaviors, which they also consider to be a violation of human rights, should be deemed illegal and be punished accordingly. The findings, which were discussed in line with the existing research, indicated some significant predictors of internet users' perceptions on cyberhate.\nOriginality/value: As cyberhate is an understudied area that raises concerns in terms of internet user exposure, the objective of this research is to understand tendencies about the opinions and perceptions of internet users regarding online hate speech, and bring cyberhate to the attention of internet users and policy stakeholders.\n",
"role": "assistant"
}
] |
Organisational factors influencing sustainability of local non-governmental organisations: Lessons from a Ghanaian context
|
[
"Sustainability",
"Non‐governmental organizations",
"Management",
"Leadership",
"Ghana"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Increasingly, it is being accepted in many developing countries that the state can no longer be the sole provider of goods and services for eradicating poverty and ensuring sustainable development among its people. Accordingly, with the increasing demands on governments in many developing countries to meet the needs of their citizenry, non-governmental organisations (NGOs) are taking active and complementary roles in harnessing the potentials of the people for national development (Lekorwe and Mpabanga, 2007). Consequently, in the last decade the support to, and interest in NGOs has grown in many developing countries. The persuasion was that NGOs were more flexible, adaptive and quick to respond to peoples' needs than governments (Lekorwe and Mpabanga, 2007). Further, they could deliver higher quality social services and programmes to the very poorest sectors of the society in a cost-effective and efficient way for sustainable development (Clark, 1991; Fowler, 1988). Not surprisingly, by 2004, Ghana had already registered 3,000 NGOs to operate in different parts of the country (Adam, 2004).Interestingly, the expanded role of the NGO sector had been accompanied by heightened expectations of accountability and requirement for demonstrable evidence that they can significantly contribute to national development as earlier thought. In most developing countries such as Ghana, NGOs have been associated with several weaknesses including lack of experienced manpower, limited financial assistance, having short range objectives, yielding to political influences and corruption (Schiampo-Campo and Sundaram, 2001; Shastri, 2008). Schiampo-Campo and Sundaram (2001), for instance, argued that not all NGOs have the time and expertise to efficiently manage all their programmes, or even to ensure full involvement of communities. These perceived weaknesses, perceptibly, present a major threat to organisational sustainability of many local NGOs in developing countries. That is, the ability or capacity of the NGOs to endure - internal and external factors that affect their operations and survival (Broekhuis and Vos, 2003; Geelsa, 2010). Anecdotal evidence shows that many NGOs in Ghana fold up within the first two years of their registration for reasons not clearly understood. For those that survive, only a few remain active and functional, and the worse affected are local NGOs.\nResearch objectives: Given the background, the objective of the study was to provide an understanding of the key factors that can influence the sustainability of local NGOs based on a Ghanaian context. The intent is to provide suggestions for improving the sustainability of local NGOs.\nResearch methodology: The research adopted a case study approach to allow for an in-depth understanding of a complex organisational phenomenon such as sustainability, which according to Kane (1999) involves an enormous variety of activities, often without clarity and distinctiveness in an organisation. The Western Region of Ghana was selected for the study because it is one of the worst affected regions in the country in terms of local NGO sustainability, and the terrain was well-known and accessible to the researchers. Anecdotal evidence suggests that although several NGOs have registered to operate in the region, only few of them have survived and are functional. The assumption is that those that have survived have been influenced by certain factors worth understanding. As such, the choice of the study area was influenced by Marshall and Rossman's (1999) view that an ideal case should be accessible, offer the researcher the opportunity to encounter many of the processes, people, programmes, interactions and/or structures that are relevant to the research question or objective, and provide credible data for the analysis of the phenomenon.To select key informants, a combination of simple random sampling and purposive sampling procedures were used. Two districts, namely, Mpohor Wassa East and Wassa West were selected from the 17 districts in the Western region, because they were the most accessible. To select an NGO for the study, the determining criteria were that it was seen as having being in operation for five or more years, and having the majority of the staff - particularly senior staff - who had worked for the organisation from its inception, and are in position and willing to share their experiences, and having good archival records. Based on the above criteria, a total of 28 key informants from 20 NGOs were selected from the two districts for the study. Beside those from the NGOs, a total of 32 beneficiary informants, at least two from each selected NGOs participated in the study. The latter provided supplementary information for clarification and validation regarding the activities of the NGOs where necessary.The data were collected mainly through interviews and documents. A structured interview schedule was used to generate quantitative data that summarize the characteristics of the participants and the NGOs with respect to sustainability. This quantitative data were analysed with descriptive statistics including frequencies, means and percentages. The use of statistics here were primarily to show the differences in views of the key informants in relation to key sustainability factors, and not for the purpose of generalizing the findings to all local NGOs in Ghana. The qualitative data generated from the interviews and documents were analysed by summarizing, coding and establishing relationships among the factors influencing the survival of the participating NGOs. To ensure a high quality case study, several strategies including data triangulation, establishment of chain of evidence and explanation-building analytic strategy were adopted to improve the credibility of the findings.\nResults and discussion: The context of the case\nConclusions: Based on the findings of this case study, several conclusions can be drawn. The employees of local NGOs in the study area were dominated by youthful males below the age of 40 years with moderate level of working experience ranging from three to four years. They were mostly middle-level management personnel.The findings suggest that to improve organisational sustainability, local NGOs can broaden their scope of sustainability factors to include human resource management especially leadership, funding, programme development, general management and material resources. The above factors, although not mutually exclusive, were perceived to affect local NGOs' sustainability in the study area. But more importantly, it can be concluded that supportive leadership and adequate funding were the two most essential ingredients and determinants of local NGOs' sustainability in the study area.On human resource capacity, it can be concluded that managerial leadership, technical competence and staff motivation are three most critical factors that can influence the survival of local NGOs. The study suggests that local NGOs can achieve this through advertisement, peer recruitment and regular on-the-job training especially in leadership, project and prudent financial management and lobbying. Particular to funding, the study reveals that it is becoming increasingly difficult for local NGOs to attract funding, especially, from foreign sources to support their programmes. Accordingly, the study suggests that local NGOs in similar context to those studied can improve their funding situation by spreading their network to include domestic (local) sources. Vigorous lobbying from board of directors, management, friends, churches and philanthropists, with good project proposals, and the provision of for-fee consultancy services represent some of the measures which local NGO can adopt to improve their ability to raise local funds.Based on the findings, it can also be concluded that the survival of local NGOs in similar context as those studied can be improved if they are able to develop and implement need-based and demand-driven programmes, supported by good leadership, transparency, accountability, and commitment to meet felt and expressed needs of beneficiaries. Further, the study highlighted the need for management capacity that promotes learning organisation principles (e.g. shared vision, teamwork), and the provision of quality material resources for organisational efficiency and effectiveness.\n",
"role": "user"
},
{
"content": "Purpose: - Despite the role of non-governmental organisations (NGOs) as actors in development, the issue of local NGOs' sustainability remains a major concern in many developing countries such as Ghana. This study seeks to determine key factors that can influence local NGOs based on a Ghanaian case study.\nDesign/methodology/approach: - The research used a combination of quantitative and qualitative research approaches.\nFindings: - The study identifies that the availability of funds, quality material resources, supportive leadership, development of needs-based and demand-driven programmes, and effective management can have a significant influence on the sustainability of local NGOs. Significantly, leadership emerged as the most important factor to organisational sustainability of local NGOs. Equally, the availability and quality of material resources for work were least critical to local NGO sustainability, although they were all very important.\nPractical implications: - The case study suggests that local NGOs can adopt several measures to ensure their sustainability. These include the writing of good needs-based and demand-driven project proposals, ensuring transparency and accountability, leadership training and lobbying for resources including funds, especially, from local sources.\nOriginality/value: - Although the findings are based on Ghanaian local NGO sustainability scenario, the lessons can be useful to NGOs within similar contexts.\n",
"role": "assistant"
}
] |
Barriers to repeat patronage: the impact of spectator constraints
|
[
"Consumers",
"Marketing strategy",
"Sporting events"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The international prominence of spectator sport, its prevalence and fiscal importance, continues to draw attention from economists and marketers (Bambridge et al., 1996; Mason, 1999). The industry's magnitude is evident in global revenues with $15 billion reported in television rights alone (Back et al., 2004), and consumer spending where a third of the US adult population spent $13.6 billion to attend events (US Census Bureau, 2004). Interest in professional sport is also high. For instance, a readership survey of US and Canadian adults (55 percent of who peruse newspapers daily) reports sports as second only to front page news as the most widely read section (Newspaper Association of America, 2003). Even with significant numbers passing turnstiles and using media, understanding the ebb and flow of spectator behaviour continues to challenge industry researchers and practitioners. Some investigators focus on the role service, event atmospherics or team attachment has on repeat patronage (Hightower et al., 2002; Wakefield and Sloan, 1995). While another stream of research in Europe and the US centres on the effect media use has in the consumption of the sport product (Jeffres et al., 2003; Pritchard and Funk, 2006).Understanding what energizes behaviour often leads market researchers to consider the \"why\" behind consumption and the psychological motives driving purchase behaviour (Kahle et al., 1996). Although motivation research answers, \"Why do people buy?\". It does not address an equally formidable question, \"Why do people stop buying?\". Notwithstanding early interest from marketing practitioners, constraints-to-consumption have drawn little attention from the field (Lepisto and Hannaford, 1980). Motivation's weak explanation of consumer behaviour certainly gives momentum to the idea something intervenes to prevent motivated patrons from acting consistently, yet work on impediments is sparse. Assael (1995), p. 94) believed that regardless of being strongly disposed, consumer decisions to continue could be \"terminated at any stage because of outside constraints\". Unfortunately, at this stage these \"obstacles\" hinder motivated intentions from ending in repurchase remains underdeveloped (Oliver, 1999, p. 36). In reality, studies on the factors behind discontinuance are scarce. This void then is the focus of the current study, which examines constraints to determine how they impede motivated consumers from continuing.\nLiterature review: Bagozzi (1993, p. 216) argues attitude-based explanations of consumption are poor, and advises consistent predictions will remain elusive until attention focuses on how \"mechanisms intervene between attitudes or preferences on one hand and behaviour on the other\". Using a general theory of action, he suggests motivation's ability to explain behaviour improves when volition's role in the sequence is considered. Volitional processes represent the degree of freedom patrons perceive in their decision to act. Psychologists suggest people \"feel and behave more positively\" when they perceive greater control in a situation (Faranda, 2001, p. 1261). Control is akin to volition, as both perceptions moderate whether strong, positive attitude translates into action. Like Bagozzi's ideas on volition, other attitude researchers commend control's inclusion in models that seek to explain consumer behaviour (Ward and Barnes, 2001). Gotlieb et al. (1994) took this tact when he tested control's moderation of satisfaction in patron decisions to reuse health service providers. In similar fashion, Pritchard et al. (1999) observed volition moderating brand commitment's explanation of loyal behaviour, with a weak view of constraints (i.e. fewer constraints, more freedom) allowing stronger patronage patterns to occur. This suggests volition entails two related perceptions, that customers are not only free from constraints but also free to act (i.e. they control the decision) when moving toward repurchase.Reflecting this idea in a motivation - behaviour sequence means showing how constraints prevent consumers from acting consistently (Bagozzi, 1992). Our view is that constraints form perceptions of volition which then moderate patronage (Bagozzi et al., 1990), since those aware of barriers inhibiting their patronage would naturally feel less free to continue. The following literature sketches out prior work on consumer motivation and the nature of constraints to ongoing patronage.Constraints\nMethodology: The objective of the data collection was to collect spectator responses about their attitude and behaviour toward consumption. Testing the impact of constraints in a motivation-behaviour sequence could proceed once valid measures for each variable in Figure 1 were in hand.Sample\nResults: Respondent commentary on what constrained their attendance were submitted for content analysis by three researchers (n=308). Using descriptions from earlier classifications, researchers initially coded responses into one of eight categories (see Table II). Reliability assessment used Cohen's (1968) kappa to determine the degree of agreement between researcher ratings (Hughes and Garrett, 1990). Cohen's kappa is an intra-class correlation derived from the following definition: observed agreement - expected agreement/(1-expected agreement). Values range from 1, indicating perfect agreement, to values of 0, which indicate little conformity between ratings. With inter-judge reliabilities >0.80 deemed satisfactory (Keaveney, 1995), kappa comparisons demonstrated consistent researcher classifications: R1-R2=0.98, R1-R3=0.93, R2-R3=0.92. Next, category classifications assigned patrons to one of three constraint groups:1. No constraints (n=40).2. Internal constraints (n=78).3. External constraints (n=190).Examining the constraint groups\nDiscussion: Although little effort has focused on processes intervening between attitudes and action, some authors believe adding barriers can improve explanations of consumptive behaviour (Patterson, 2004). To date most research concentrates on positive barriers that encourage people to continue (Jones et al., 2000) rather than on negative factors that prevent patronage (Lepisto and Hannaford, 1980). The current study considers the latter by examining how constraints impede spectator attendance. Content analysis identified the primary reasons behind patron decisions not to attend sports events. Over a half of the sample cited external barriers such as work commitments, and a quarter recorded internal reasons like personal or family priorities restricting their patronage.After initially identifying constraints, the study tests a motivation-behaviour sequence. Direct and indirect routes to patronage garnered support. However, motivation's link to attendance favoured an indirect route via media use. This finding appears to reflect a dual role where media can both compete with and fuel spectator attendance (Jeffres et al., 2003; Pritchard and Funk, 2006). This matter is a sensitive issue for the entertainment industry. Discussions over whether media broadcasts hinder or assist patronage regularly surface (Yin, 2003; Zhang and Smith, 1997), prompting management to revisit blackout policies that struggle with ideas of lost revenue at the turnstile (Bambridge et al., 1996).Understanding how or why patronage shifts between attendance and media use is important. Our final hypothesis gives this some attention when it examined how constraints interfere with consumption. Despite some theorists speculating barriers moderate A-B (Crompton et al., 2005), our results found the presence of constraints failed to alter product attitude such as motivation or involvement, but did significantly hinder behaviour. Those without constraints consistently attend more games and use media more frequently, whereas the reverse was true for those who identified barriers. This finding is coincides with Assael's (1995) view that constraints moderate behaviour but do not necessarily affect attitudinal precursors to patronage. Patterson's (2004, p. 1305) suggestion that consumers engage in \"reward/cost\" evaluations is another plausible way to think about why patrons change their behaviour. Although our inquiry did not directly cover this, subsequent work could calculate whether volitional processes actually do weigh motivational benefits against the costs (effort) of overcoming barriers to patronage (Bagozzi et al., 1990). In sports this sort of process would mean that spectators though motivated, might at times use media as a more convenient (less effortful) alternative (Yin, 2003). But that when an important match comes (divisional rivals, play-offs, etc.) fans become more willing to negotiate obstacles as rewards outweigh barriers (costs) to attend.Practical implications\nLimitations: First, while motivation and involvement measures performed relatively well in the study the explanation of the outcome (attendance) was limited. This suggests subsequent work could include other factors that account for repeat patronage (e.g. Bauer and Sauer, 2005; Hightower et al., 2002). Although Patterson (2004) and our study advocate including barriers to predict behaviour, both used nominal groups to assess impact. Determining constraints with open-ended data alone can restrict consideration to type (nature) and not the potency or magnitude of those perceptions. Much like Keaveney's (1995) exploratory work on switching behaviour, we believe similar effort that grounds the nature and measurement of constraints is necessary. A further limitation lay in the use of cross-sectional data to study ongoing behaviour. In this sense, constraints could vary over time or by season (Crompton and Kim, 2004). Attending to this may require using research designs (e.g. experiments) or samples (e.g. longitudinal panels) that expressly focus on how motivation and constraints fluctuate and alter patronage over time (Winer, 1999).A final limitation of the study is that we narrowly define and examine constraints in terms of factors that prevent repurchase. This may prove misleading as patrons may engage or modify their behaviour in ways beyond attendance and media use. Capturing substitution in leisure consumption may mean we need to consider a wider evoked set with a range of entertainment alternatives (Timmermans et al., 1992). Widening the focus of research beyond hedonic consumption is also necessary, as constraints to purchase extend well beyond this arena. Consumption of need-based utilitarian products or services might offer very different accounts to the one rendered here (Babin et al., 1994). For example, barriers to using medical services may constitute significant internal or external cost (e.g. perceived risk, financial), yet a patron's sense of need (e.g. type of surgery, appendectomy versus cosmetic) may consistently outweigh those constraints (e.g. it is expensive but we will take out a loan).\nConclusion: To summarize, the study reintroduces the topic of constraints and provides a description of barriers patrons encounter that prevent them from attending sport events. The practical advantage for marketers of assessing both motivation and constraints lies in their joint depiction of how best to influence patronage. In our mind, the work offers a preliminary step in this direction, clarifying why and how fans continue or discontinue.\n",
"role": "user"
},
{
"content": "Purpose: - The reason patrons cease to attend sporting events is not well understood. The purpose of this paper is to examine how factors motivate and inhibit patronage from continuing.\nDesign/methodology/approach: - A random sample of a sport franchise's fan database (n=308) is drawn. First, respondent data refines measures and tests a structural equation model of direct and indirect links to patronage. Next, content analysis classifies spectators according to self-stated barriers to continuance. These groupings then check the moderating role constraints have on patron attitude and behaviour.\nFindings: - Structural work confirms both direct and indirect links but notes that consumption primarily took an indirect route, with motivational desires rousing fan involvement and media use before increasing attendance. Group differences verify constraints and limit patronage but do not dampen product-related attitude.\nResearch limitations/implications: - The study helps clarify the connection between media use and attendance, describing how constraints impede spectator consumption. Study limitations include a focus on one hedonic service setting and the use of cross-sectional data to examine ongoing phenomena.\nPractical implications: - Negotiating barriers to repeat purchase remains largely overlooked as a foundation for guiding strategy. Practical implications consider integrating both motives and constraints when marshalling efforts that build continuance.\nOriginality/value: - Despite early interest from marketing practitioners, factors that inhibit patronage have drawn little attention. This study employs content and path analysis to address the matter.\n",
"role": "assistant"
}
] |
Exploring antecedents of extended supply chain performance measures: An insight from Indian green manufacturing practices
|
[
"Performance management",
"Manufacturing",
"Supply chain management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Supply chain management was earlier viewed as a mean of movement of materials and goods. It was just considered as a process by means of which the organizations used to strategize. The best value supply chains focusses on increasing key results and not just merely movement of the product from manufacturing unit to customer. Mehra (2005) suggested that due to the rapidly changing global market there is a great demand to analyze the emerging trends that can impact the business. The supply chain is a key component of planning in all types of business; hence it must be constantly observed. The activities like developing, manufacturing and marketing a product may throw a serious challenge before any organization. In order to ensure a sustainable competitive advantage the companies should strive for supply chain excellence and should constantly monitor their current processes. The following aspects are some of the emerging trends in supply chain management:\n2. Literature review: We have adopted \"SLR\" in our paper as suggested by Tranfield et al. (2003). The present section is divided into two stages and its phases.\n3. Conceptual model development and research design: 3.1. Conceptual model\n4. Data analysis and findings: In this section researcher present exploratory factor analysis (EFA) conducted on raw data collected during pilot study to check construct validity and measure reliability of parsimonious and orthogonal factors obtained during second varimax rotation.\n5. Concluding remarks: The conclusions drawn from factor and regression analyses are as follows.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore the antecedents of Indian firms practicing green manufacturing practices and their impact on extended supply chain performance.\nDesign/methodology/approach: In this paper the authors have adopted systematic literature review (SLR) to derive research objectives and questions. In order to answer research questions the authors have proposed a theoretical framework and derived research hypotheses. The authors identified constructs and items through extant literatures and developed a structured questionnaire which was pretested before using for final survey. The data were collected in two phases and also performed wave analysis to check non-response bias to avoid any significant impact of non-response bias on statistical analysis. The data has been used to conduct exploratory factor analysis using varimax rotation which reduces variables into five parsimonious and orthogonal factors. The factor analysis output was further used as an input of regression analysis.\nFindings: The factor analysis output has further validated the findings from literature review. The factor analysis output suggests that total quality management (TQM), supplier relationship management (SRM), R\nResearch limitations/implications: The present study is limited to medium-sized manufacturing firms. Second, the conclusive model explains only 31.9 percent of the total extended supply chain performance. The study provides an insight to managers that those companies which have implemented TQM and green technology have realized superior performance in comparison to those which have not successfully implemented.\nOriginality/value: The present study is toward understanding Indian green manufacturing practices.\n",
"role": "assistant"
}
] |
Estimating a measure of intellectual capital value to test its determinants
|
[
"Intellectual capital management",
"Intangibles",
"Intellectual capital",
"Long-run value"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This study contributes to intellectual capital (IC) research by innovatively applying a measure from the area of mergers and acquisitions to IC research, which offers new research opportunities regarding IC value. Additionally, determinants of IC value are investigated. The findings on the IC value determinants contribute to IC management, as the development of IC value can be guided by focusing on the significant factors for IC value. This paper responds to the challenge posed by previous IC research to develop more creative quantitative approaches to estimate IC value (Marr et al., 2003; Mouritsen, 2006). These previous studies argue that quantitative measures enhance testing IC-related hypotheses. This study compares three measures to examine which may best indicate IC value: market-to-book (MtB), Tobin's q and LRVTB. The IC value measure with the highest explanatory value in a regression of corporate performance is seen as the best estimator for IC value. This IC value measure is then used to examine IC value determinants.\nLiterature review: IC as strategic advantage for corporate performance\nHypothesis development: Addressing weaknesses of MtB as measure of IC value\nResearch design: Sample of German companies and industry grouping\nResults: Descriptive results for IC value measures\nConclusion: The aims of this study are to identify the IC value measure with the highest explanatory value in a regression of corporate performance out of three estimators and use this measure to examine determinants of IC value. As this study argues that holistic measures capture IC value more comprehensively, this study chose three holistic market-based measures: MtB, Tobin's q and LRVTB. The criticism of MtB as estimator of IC value has initiated the hypotheses to test whether MtB, Tobin's q or LRVTB serve best to estimate IC value. The three IC value measures are estimated and the explanatory power of the three measures is compared in a regression analysis for their association with corporate performance. To test the hypotheses on IC value measures, Vuong's closeness test is applied. The findings identify LRVTB to be the best IC value estimator. This measure from the research area of mergers and acquisitions offers new research opportunities for empirically investigating IC value.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to identify a measure of intellectual capital (IC) value which offers new research opportunities for empirical investigations and to examine the determinants of IC value.\nDesign/methodology/approach: In total, 4,488 firm years of German companies are investigated to compare three measures of IC value: market-to-book, Tobin's q, and long-run value-to-book (LRVTB).\nFindings: LRVTB is observed to be the IC value measure with the highest explanatory value. This measure provides an approach to empirically test previously untested hypotheses on IC value. The results on testing determinants of IC value indicate that IC value is positively related to leverage and motivational payments to employees and negatively associated with company size. In contrast, recognised intangible assets, research and development (R\nResearch limitations/implications: The findings on IC value measures contribute to IC research as they offer a way to estimate IC value for testing IC-related hypotheses. The findings on IC determinants have implications for IC management as the relevant determinants can be considered for IC value creation.\nOriginality/value: This paper responds to the challenge posed by previous IC research to develop more creative quantitative approaches to estimate IC value (Marr et al., 2003; Mouritsen, 2006) in order to test IC-related hypotheses by innovatively applying a measure from mergers and acquisitions research to IC.\n",
"role": "assistant"
}
] |
Factors for successful e-learning: does age matter?
|
[
"Complexity",
"Success factors",
"Technical support",
"Age",
"Authentic learning",
"Organizational e-learning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: There is increasing reliance on the use of e-learning to facilitate learning and development in the workplace, and in Australia, for example, at least 48 per cent of all vocational education and training activity involves the use of e-learning (Flexible Learning Advisory Group, 2013). At the same time, the workforce is ageing and it is argued that many of the training and development practices in organisations have been slow to take into account the needs of older workers (Lazazzara and Bombelli, 2011; Young, 2013). This issue may have implications for e-learning as there are often claims of a gap in technological capability between those who were born prior to the advent of modern technology and those who have always known technology to be pervasive in their world (Prensky, 2001a, b).\nTheoretical background: E-learning has been researched heavily from both an educational/learning perspective and a technology adoption perspective. Whilst an interest from differing disciplines provides an opportunity for breadth and depth of understanding, it also brings a wide range of definitions and theories that have been applied to e-learning research. This study uses the Servage (2005, p. 306) definition of e-learning that refers to \"instructional content or learning experiences delivered or enabled by electronic technology\".\nFactors influencing satisfaction with, and use of, organisational e-learning: E-learning is commonly used to, improve performance, develop skills, provide ease of access, reduce costs and increase levels of motivation (Ali and Magalhaes, 2008; Welsh et al., 2003). While the adoption of e-learning across organisations continues to increase, there has also been reports of significant learner dissatisfaction with some e-learning, which can lead to overall organisational e-learning failure (Guri-Rosenblit, 2005; Sawang et al., 2013; Sun et al., 2008).\nMethod: As the factors for the focus of this study have been identified, and previously validated constructs exist to measure these factors, a quantitative study involving a self-administered online questionnaire was considered most appropriate to test the hypotheses (Sekaran and Bougie, 2013). In the following section, context for the case organisation and the industry in which it operates is presented, along with details relating to participants, procedures and measures used in the study.\nResults: Preliminary data analyses\nDiscussion and implications: The key aim of this study was to explore the impact that age has on the intended future use of e-learning relative to other potential predictors. The findings suggest that three variables are useful predictors of intention for future use of organisational e-learning: low complexity, authenticity and technical support. Furthermore, overall satisfaction with e-learning was found to be an intermediate variable in the relationship between these factors and intentions to use e-learning in the future. However, age was not a significant factor impacting either future use intentions or satisfaction with e-learning. The findings confirm that the critical barriers affecting employees' overall acceptance, satisfaction and future use of e-learning are factors that continuously appear across much of the recent literature, which makes the dissemination of the results important for the rail industry to continue with its research.\nConclusion: E-learning has been and will continue to be adopted by many small and large organisations. This study has reinforced success factors required for organisations to efficiently and effectively adopt and continue to use e-learning as part of their training and development strategies. Importantly, the research has concluded that employee age has no significant impact upon satisfaction and intention to use e-learning in the future. Instead, it was found that perceived complexity, authentic learning and technical support were important factors in determining future intentions to use e-learning systems and that e-learning satisfaction moderates the relationship between each of these factors and future use intentions. Based on these findings, it is suggested that a better understanding of an ageing workforce and consideration of specific issues relating to a blue-collar workforce is important to promoting satisfaction and future use intentions. Further, this study advances the theoretical understanding that there are a number of contextual factors that can be used as predictors of successful employee e-learning.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the factors affecting employees' overall acceptance, satisfaction and future use of e-learning, specifically exploring the impact that age has on the intended future use of e-learning relative to the other potential predictors.\nDesign/methodology/approach: The project developed an online survey and invited employees of one Australian rail organisation to participate. Questions were structured around the factors that affect acceptance and future use of e-learning. Statistical analysis was used.\nFindings: The findings from the study suggest that, despite the often espoused stereotype, age is not a significant factor impacting either future use intentions or satisfaction with e-learning. In contrast, three variables were found to be useful predictors of intention for future use of organisational e-learning; low complexity, authenticity and technical support.\nResearch limitations/implications: The study did not consider other moderating effects related to demographic data other than age, such as educational experience. Further, the case presented is a single organisation and therefore is not necessarily representative of other industries. Future studies should adopt a mixed methods approach.\nPractical implications: This study has emphasised that attention needs to be focussed on factors over which organisations have control when adopting and using e-learning. Employee age should not be seen as an obstacle to e-learning implementation, rather attention needs to turn to effective and user-friendly e-learning interventions along with sufficient technology support.\nOriginality/value: Perceptions within industry and indeed in some literature, suggest that employee age stereotypes still exist in relation to technology uptake. This research has demonstrated that this stereotype is an erroneous assumption and emphasised the importance of other factors.\n",
"role": "assistant"
}
] |
Analysis of employability for the civil aviation maintenance graduates of Turkey
|
[
"Employability",
"MRO",
"Regression analysis",
"Markov analysis",
"Civil aviation",
"Maintenance graduates"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The aviation sector is in the top ranks among the sectors with high value-added, where high technology is most considerably used (Bao and Ding, 2014; Hampson et al., 2012). The activity in the sector creates employment and requires that many branches work simultaneously. The aviation sector, which is in need of skilled labor, is rapidly developing its subordinate industries along with possessing a multi-sectoral structure (Blyton et al., 2010). The maintenance industry reacted to the need for systemic solutions by first focusing on developing awareness-level training programs (Patankar and Taylor, 2008). Certification of maintenance organizations and personnel, education and training and information sharing and administration are challenges of the regulatory reform about civil and military aviation (Le and Lappas, 2015).\nAviation maintenance personnel certification and training developments: When compared with the other sectors, the aviation sector has higher quality standards (Bao and Ding, 2014). Procedures, regulations and documentations are indispensable elements of the sector. Every operation performed must be carried out in accordance with its rule and written procedure. Furthermore, every step of the operation performed has to be recorded. There are certification differences between CASA in Australia and EASA in Europe (Yadav, 2010). However, the Directorate General for Civil Aviation (SHGM) published the regulation on the Aircraft Maintenance Staff License (SHY-66) according to Annex 1 of the ICAO. Turkey must also follow this regulation. It encompasses the requirements and regulations concerning the certifying of maintenance staff members (SHY-66, 2007). Under this regulation, the condition of receiving basic training from the authorized educational institutions is required to become a certifying maintenance staff member. The authority has been divided into four categories as A, B1, B2 and C. A is mechanical technician, B1 is aircraft mechanical maintenance technician, B2 is aircraft avionic maintenance technician and C is base maintenance engineer or technician.\nMethods: Regression analysis of the certifying maintenance personnel\nDiscussion: The aviation sector is in the top ranks among the sectors with high value added where high technology is most considerably used (Bao and Ding, 2014; Hampson et al., 2012). Certification of maintenance organizations and personnel, education and training and information sharing and administration are challenges of the regulatory reform about civil and military aviation (Le and Lappas, 2015). Organization structure - recruitment and training - is one of the main research areas in aircraft maintenance (Shanmugam and Robert, 2015). In recent years, the public and regulators have expressed concerns regarding the effectiveness of repair stations. Expectation is to increasing the service quality of maintenance (Drury et al., 2010).\nConclusion: In the employment analysis performed in this study, the graduates with an associate's degree were considered Category A technicians, whereas the graduates with a bachelor's degree were considered Category B technicians. The very few Category C staff members were disregarded. According to the information which the Directorate General for Civil Aviation provided, it is predicted that Turkey will have a fleet of 750 airplanes in the next decade. It appears that the aircraft MRO sector will also grow parallel with the rate of increase in airway firm fleets, and there will be a need for human resources for maintenance staff at the same rate. Nevertheless, when we considered the growth rates of the fleets of the airways in the past five years, some growth equivalent to the growth expected in the next decade was seen. The increase in employment in the aircraft MRO sector is high as compared with the other sectors. It was also observed that the increases in the quotas for the institutions of higher education regarding technical training on aviation in Turkey between 2008 and 2012 reached higher rates.\n",
"role": "user"
},
{
"content": "Purpose: With every passing day, the center of gravity for air traffic for the Middle East and Turkey is shifting toward becoming a center for airway transport, maintenance, repair, overhaul (MRO) and transfer. The MRO sector is gaining importance because of this increase in Turkey. The purpose of this study is to examine the future employability of the graduates receiving civil aviation maintenance training in Turkey.\nDesign/methodology/approach: The paper uses past seven years' data about maintenance personnel of Turkey. It includes all-level students, recently working personnel and retired personnel of each year. Regression analysis and Markov analysis were used to show the prediction of aviation maintenance personnel.\nFindings: The paper provides quantitative results about employability of maintenance personnel under the real numbers of alumni and working and retired personnel in Turkey. The results show that the number of graduates will be greater than the number of staff members in certifying categories in 2018. Presently, Turkey meets the employment expected, but in the future, aviation maintenance personnel number is rising, and there will be an excess number of graduates left over in Turkey.\nResearch limitations/implications: Because of the chosen region, Turkey, the research results may lack generalizability. Therefore, researchers are encouraged to conduct analysis for other countries which have an important role in aviation.\nPractical implications: The paper includes implications for the government maintenance personnel policy and educational policy for managing all types of maintenance personnel in Turkey.\nOriginality/value: This kind of personnel needs analysis has not been done before, and this is the originality of this paper. It will in turn lead to an analysis of planning for allocation of maintenance staff and alumni. The authors collected all official numbers about maintenance personnel from Turkish Student Selection and Placement Center of Turkey, Turkish Technic and Turkish Civil Aviation Reports.\n",
"role": "assistant"
}
] |
Social enterprise in health organisation and management: hybridity or homogeneity?
|
[
"England",
"Health organization and management",
"Health care",
"Social enterprise",
"Hybridity",
"Health and social care",
"National Health Service",
"Isomorphism"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Over recent decades, policy developments in public sector reform have increasingly emphasised the decline and fragmentation of established welfare bureaucracies. This disaggregation and decentralisation of services has led some to suggest that service provision is becoming increasingly \"hybrid\" as established sectors and boundaries come together. Health and social care has been a key area for such changes. The promotion of competition and choice has led to an increasing emphasis on private and third sector providers entering the healthcare market. These providers have been encouraged on the grounds that they are capable of being more innovative and responsive than their public sector counterparts (Allen, 2009).The English healthcare system has been particularly active in encouraging a diversity of providers. One such organisational model of provision has been social enterprise, which has been encouraged as a more innovative and responsive alternative for service users and healthcare staff. Social enterprises are broadly defined as \"business[es] with primarily social objectives whose surpluses are principally reinvested for that purpose in the business or in the community, rather than being driven by the need to maximise profit for shareholders and owners\" (Department of Trade and Industry, 2002).In England, social enterprises are currently presented as organisational vehicles with the potential for dealing with complex social needs that create social as well as economic capital. This resonates with global trends that have seen governments invest significantly in social enterprise organisations to deliver public services (e.g. Defourny and Nyssens, 2008). In the following New Perspectives essay I want to argue that this increasing interest in social enterprise in health organisation and management reflects wider trends in welfare that draw attention to increasingly hybrid organisational forms and functions. Following this, I also want to challenge this notion of hybridity, arguing that whilst social enterprise increases diversity in our understanding of health organisation and management, its rise and increasing popularity is illustrative of isomorphic processes towards market \"norms\". Social enterprise is symbolic of isomorphic processes pushing healthcare organisations toward greater levels of homogeneity based on market based standardisation and practices.\nBackground: Policy makers are increasingly using social enterprise as a discursive vehicle to address a wide range of social problems (Teasdale, 2011). Nicholls (2010) explains that the organisational legitimacy of social enterprise has been as the result of a dynamic interplay between macro-level institutional structures and micro-level organisations. In the UK, network organisations like the Social Enterprise Coalition (SEC) represent \"paradigm-building\" actors that have been influential in establishing discourses and narratives to develop ideas about social entrepreneurship (Nicholls, 2010, p. 616). They draw attention to how social enterprises are an attractive proposition in providing \"value-added\" services better able to more effectively mobilise \"pro social behaviour\" than for-profit private enterprises (e.g. Allen, 2009).In the English National Health Service, a variety of policy initiatives have encouraged social enterprises with the aim to create patient choice and give greater freedom to staff to be innovative, responsive and improve quality. For example, a Social Enterprise Investment Fund (SEIF) was established to support the development of new enterprises and encourage existing social enterprises to extend into delivery of health care (Millar et al., 2010). In the primary and community health sector, social enterprises were one of the organisational forms that could be considered under the Transforming Community Services programme (Department of Health, 2009). As part of a purchaser-provider split, primary care commissioning organisations (PCTs) were required to transfer their provider services to other organisations by 2012. This policy commitment was highly significant in that NHS employees were to be given a \"Right to Request\" (RtR) to set up social enterprise organisations to deliver community health services (Miller et al., 2012). The RtR programme has been followed in England by a Mutuals Programme to support 20 services to develop into \"public sector mutuals\", and the Right to Provide programme encouraging staff within healthcare trusts to become social enterprises.Existing research in this area has reported some notable benefits in becoming a social enterprise. This included greater innovation in clinical service delivery; the provision of a wider range of services; and efficiency savings from reduced staff absence (Miller et al., 2012; Hall et al. 2012; National Audit Office, 2011). These findings also support wider research about non-profit organisations in healthcare settings suggesting that they achieve better relations with those they serve; enhance quality and deliver more innovative service provision (e.g. Heins et al., 2010; Department of Health, 2009). That said, research also identifies challenges associated with social enterprise entry. Miller and Millar (2011) found only limited interest expressed by NHS staff to develop social enterprises due to a lack of staff support, leadership, organisational support and commissioning support. Commissioners have been highlighted as particularly problematic in often equating social enterprise organisations with being \"not business-like enough\" (Baines et al., 2010). Workforce concerns regarding job security, business skills and the potential loss of public sector branding have also been documented (Department of Health, 2010), as well as the difficulties in measuring anticipated benefits and securing funding from financial institutions and commissioners in a competitive market place.\nSocial enterprise: hybridity or homogeneity?: Social enterprise has been discussed and researched in the health and social care context but as yet the debates surrounding their entry have not been extended to analyse how social enterprises interact with other public and private sectors. It has been documented elsewhere of a \"tension field\" as different sectors and boundaries interact and become increasingly porous and overlapping (Evers and Laville, 2004). Billis (2010), for example, suggests that this hybridity represents an expanding aspect of the complex and overlapping relations between the state, the market and the third sector that is challenging and supplanting the core features, and values, of established welfare institutions. As a result of contracting out, privatisation and performance measurement, the traditional boundaries between market, state and third sector have been breaking down, leading to the emergence of a class of organisational hybrids: complex organisations with opaque accountability structures. Here, Billis (2010) argues that a prominent example of welfare hybridity is social enterprises that fuse third, public and private sector values. Such organisations are indicative of existing third sector organisations in health and social care that are now introducing social enterprise \"trading arms\" as part of their organisation. They are also indicative of changes to health care organisations currently operating within the public sector. For example in England we have seen the introduction of hospital Foundation Trusts that aim to combine public sector with private and third sector approaches (Allen et al., 2012).Social enterprises represent the arrival of new hybrid organisational forms in health organisation and management. Whilst these organisations are associated with greater innovation and responsiveness, they also represent a response to the current pressures faced by healthcare organisations to reform and adopt managerial principles grouped under the umbrella of new public management (NPM). As Brown et al. (2003) suggest, these managerial principles are aimed at changing the public sector in three areas, summarised by Maor (1999) as first, a change from hierarchical to economically based structures, second from regulative to economically based processes and third, from legally based to economically based values. In seeking to achieve these goals, the existing healthcare sector is being required to downsize, devolve managerial responsibility and introduce managerialist methods and practices from the private sector.Social enterprises represent a hybrid organisational form that is symptomatic of this convergence towards NPM ideas and practices. When we start to reflect on debates about the nature and implications of this hybridity, current thinking about hybridity as a concept tends not to extend beyond the \"tension field\" as depicted by Evers and Laville (2004). A critical reflection on these developments in healthcare suggest that social enterprise is indeed symptomatic of this hybridity but perhaps the organisational form is also symptomatic of increasing homogeneity: a convergence and blurring of sectoral boundaries that represents a market based isomorphism currently taking place in health organisation and management. It was DiMaggio and Powell (1983) who most famously described the three processes of isomorphic behaviour. These are coercive isomorphism that stems from political influence and the need for legitimacy, mimetic isomorphism resulting from standard responses to uncertainty, and normative isomorphism associated with professionalisation. DiMaggio and Powell (1983) suggest that the product of coercive, mimetic, and normative isomorphism reflected an increasing homogenisation and convergence of organisational forms fuelling standardisation and professionalisation.It has been documented elsewhere how the appeal of the third sector is based on its ability to bridge government and civil society yet problems remain for these organisations in becoming social enterprises. The dangers of the increasing influence of adjacent sectors, the loss of independence and possible mission creep in which these organisations gain resources, possible influence and the opportunity to deliver more services come at the cost of those fundamental attributes which made it an attractive proposition in the first place - its mission, values and voluntary contribution (Haugh and Kitson, 2007). Kelly (2007) suggests this blurring of sectoral boundaries has usually been regarded either as a cause or as the effect of a process of structural isomorphism in which non-profit organisations have become increasingly bureaucratised and commercial. Mimetic isomorphic forces are also at play as the organisation becomes increasingly professionalised as it is continuously required to adopt business methods that are likely to impact on the rationale and culture of the organisations.The rise of social enterprise in the context of market-based isomorphism also reflects how existing public sector healthcare institutions are susceptible to mimetic, normative, and coercive pressures as much as business and non-profit sectors. Research has tended to view public sector agencies more as playing the role of catalyst to institutionalisation in other organisations, as forces pushing non-profits and business firms toward greater levels of homogeneity. As Frumkin and Galaskiewicz (2004) suggest, these same organisations are also being subject to the same kinds of pressures. This vulnerability of public sector organisations to isomorphic pressures is particularly significant as governments tend to be the principal funder of non-profit service delivery. With increasing amounts of non-profit agency finance flowing from public sector agencies (such as the Social Enterprise Investment Fund in England), the vulnerability of government agencies to institutional pressure may have effects on the scope and character of contracting relations. One effect might well be a narrowing of the range of innovations that government agencies are willing to consider under conditions of uncertainty (Frumkin and Galaskiewicz, 2004). This could in the long run have serious implications for the diversity of non-profit initiatives such as social enterprise and might well be in tension with the drive for social innovation that is commonly associated with these organisational forms.\nConcluding remarks: In DiMaggio and Powell's (1983) classic article, the \"norm\" was the bureaucratic organisation of Max Weber. What we are seeing now is the norms changing towards market forms of organisation. In this New Perspectives essay I have argued that the rise of social enterprise is indicative of increasingly hybrid forms of organisation in healthcare delivery. What I have also argued is that within these hybrid arrangements we are seeing increasing diversity in service provision but we are also seeing an increasing convergence and homogeneity of organisational forms. Social enterprise organisations are symptomatic of the requirement to converge around a market based isomorphic process that is requiring the third sector and the public sector to converge around managerialist methods and practices fuelling standardisation and professionalisation.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to reflect on social enterprise as an organisational form in health organisation and management.\nDesign/methodology/approach: - The paper presents a critique of the underlying assumptions associated with social enterprise in the context of English health and social care.\nFindings: - The rise of social enterprise models of service provision reflects increasingly hybrid organisational forms and functions entering the health and social care market. Whilst at one level this hybridity increases the diversity of service providers promoting innovative and responsive services, the paper argues that further inspection of the assumptions associated with social enterprise reveal an organisational form that is symbolic of isomorphic processes pushing healthcare organisations toward greater levels of homogeneity, based on market-based standardisation and practices. Social enterprise forms part of isomorphic processes moving healthcare organisation and management towards market \"norms\".\nOriginality/value: - In line with the aim of the \"New Perspectives section\", the paper aims to present a provocative perspective about developments in health and social care, as a spur to further debate and research in this area.\n",
"role": "assistant"
}
] |
The role of district leaders for organization social capital
|
[
"Mixed methods",
"Social network analysis",
"School districts",
"District governance",
"District leaders",
"Organization social capital"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In many educational systems around the world, the responsibility for high-quality elementary education is increasingly placed in the hands of school district leaders (superintendents) at central offices. In the Netherlands, where this study takes place, the final responsibility for student achievement and school improvement has long been in the hands of school principals, and even when a central office was established, its role was mostly restricted to providing support services such as personnel administration, finance and logistics to often regionally organized clusters of schools. In that role, district leadership has long been conceptualized as a political-administrative environment (i.e. Delagardelle, 2008; Hofman et al., 2002).\n2. Theoretical perspective: School district governance\n3. Context: Dutch school district leaders are responsible for the districts' strategic direction and educational quality, providing an accountability structure that addresses the needs of the school districts' stakeholders and local communities (Hooge and Honingh, 2014; Claassen et al., 2008). District leaders monitor and evaluate school improvement, and support and challenge improvement processes at all levels of the district organization. District leaders hire the school's managerial staff and make decisions about the school's management alongside the principals (Center on International Education Benchmarking, 2014). Dutch school districts that run two schools or more[1] are managed by one or two district leaders who are professional managers in the sense that they are qualified for this full-time paid job. They are recruited and appointed by the school district supervisory board (a non-executive board), which acts as their employer. School districts usually have a central office with human resource management, financial and educational administrators.\n4. Design and methods: Design\n5. Results: Descriptive findings\n6. Discussion, implications and conclusion: Discussion\n",
"role": "user"
},
{
"content": "Purpose: Although it is assumed that school district governance by districts leaders can impact schools' capacity to improvement and educational quality, there is little systematic evidence to support this claim. The purpose of this paper is to discuss how governance goals and interventions affect school districts' social capital.\nDesign/methodology/approach: The empirical enquiry used quantitative data on district leaders enacting governance as perceived by their school principals. These data were collected among 399 school principals of 23 Dutch school districts in elementary education, using a survey. Social network data on social capital within school districts were collected using a social network survey among educational administrators (i.e. district leaders, central office administrators and school principals). Additionally, examples of the relation between school district social capital and governance at six school districts were described.\nFindings: Results suggest that district leaders can promote the organizational social capital of their school districts through focusing on educational goals. In addition, the findings show that they can reinforce their impact by using interventions varying in coercion level, of which offering support to school principals appears to be \"a golden button\" to make organizational social capital thrive.\nResearch limitations/implications: Limitations to the study are the generalizability of the findings (they can be questioned because \"convenience sampling\" was used) and warrant a longitudinal design to examine how organization social capital develops over time.\nOriginality/value: The study is unique as it addresses the impact district leaders may have on their districts' social capital by focusing on social network approach in the study of school district governance.\n",
"role": "assistant"
}
] |
Performance modeling and optimization for complex repairable system of paint manufacturing unit using a hybrid BFO-PSO algorithm
|
[
"PSO",
"Sensitivity analysis",
"Markov process",
"BFO-PSO",
"Long-run availability",
"Paint manufacturing system"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Nomenclature\n1. Introduction: System availability is always being considered one of the major problems for the industries as it affects directly to the overall industrial performance in terms of maximum utilization of various operating systems. The inaccuracies in the maintenance planning highly influence the system performance and their effects become magnified with increased complexity of the system or the number of the sub-systems/machines used is large. The visual inspection is not enough for the system analyst to predict the perfect system behavior. It is also necessary to analyze all the factors which cause the system failures. These complex industrial systems are the main challenges for system analyst in maintaining the availability of various operating systems for a long period of time without failure. In recent years, evolutionary tools have become popular optimization methods to solve complex engineering problems of industrial systems (Garg and Rani, 2013). It predicts not only the system behavior but provides various availability measures to devise a suitable maintenance policy for better system performance. Therefore, several evolutionary tools have been applied to the complex engineering problems for the purpose of getting maximum performance by utilizing existing operating systems. Different aspects of availability analysis and performance optimization are explored in the literature. Arabi and Jahromi (2013) used the redundancy technique to improve system availability. The availability of the system is optimized considering the redundancy and repair facility allocation. Di Bona, Forcina and Silvestri (2016) applied a new approach based on integrated factor method for reliability allocation using an aerospace system prototype and compared with other traditional methods in order to validate the proposed method. Di Bona, Forcina, Petrillo, De Felice and Silvestri (2016) proposed a new reliability allocation method, i.e. critical flow method (CFM) for complex systems with series-parallel configuration and the proposed method is validated using a real case study of a cooling system through the comparison with conventional methods. (Garg, 2013; Garg and Sharma, 2012) analyzed the system behavior by utilizing the rough and imperfect data of the complex repairable system. They have used the Lambda-Tau technique for behavior analysis of the system whereas Garg and Rani (2013) presented the PSO and IFS technique for reliability analysis of the industrial system. Gupta et al. (2007) used the matrix method to solve the governing differential equations to analyze the system reliability and used Runge-Kutta fourth-order method to solve the same differential equations to verify the solution. Hajeeh (2015) developed the optimization models to study the perfect behavior of operating systems in order to maintain the balance between the cost and performance of the repairable system. Kachitvichyanukul (2012) reported the application of GA, PSO and DE algorithms using the different optimization problems with the same number of function evaluations. Kajal and Tewari (2012) proposed the GA approach for performance optimization in the dairy industry. Kennedy and Eberhart (1995) presented the particle swarm optimization (PSO) technique whose mechanics is inspired by the social behavior of the biological population. Khanduja et al. (2010; Khanduja et al., 2011) developed the performance model in a paper plant and presented the GA technique for the optimization of system performance. Kora and Kalva (2015) proposed the hybrid BFPSO technique to detect the bundle branch block for heart circulatory system. Kumar and Garg (2016) applied the PSO technique to improve the availability of a repairable system in brewery plant. Kumar and Ram (2013) investigated the reliability and sensitivity analysis of a coal handling system of thermal power plant. Kumar et al. (2018) selected the ethanol manufacturing plant to demonstrate the application of PSO algorithm for performance optimization whereas Kumar and Tewari (2017) used the PSO technique to optimize the availability of various sub-systems in a beverage plant. Raju et al. (2018) used a hybrid PSO-BFO algorithm for the optimization of FDM process parameters in order to improve the mechanical and surface quality of complex objects manufactured through the 3D printer. Modgil et al. (2013) developed the performance model using Markov approach for the time-dependent system availability in a shoe industry. Pang (2015) proposed a new computer network technology based Markov model for failure prediction in a manufacturing industry. Rabbani et al. (2018) employed the GA and PSO algorithms to obtain the optimal value of design parameters of a CCHP system. Sharma and Vishwakarma (2014) described the GA approach for the availability optimization of refining system in the sugar industry. Yadav and Zhuang (2014) reported the effectiveness of a reliability allocation approach considering the modified criticality factors. A case example is considered to demonstrate the proposed approach.\n2. System description: The liquid paint is a composite of mainly three constituents, i.e. pigments, binders and solvents (thinners). Some other additional additives are also blended in the solution to get the required properties for specific purposes or applications. Generally, paint is a blend of a finely divided pigment dispersed in a combination of different constitutes. The paint manufacturing system comprises following five sub-systems in series and parallel configuration. The schematic process flow diagram of the paint manufacturing system is shown in Figure 1:\n3. Methodology: 3.1 Performance modeling\n4. Results: 4.1 Sensitivity analysis\n5. Conclusions: In this paper, the sensitivity analysis has been conducted by utilizing the uncertain data of system concerned and traced out the best system performance for a paint manufacturing unit. In the proposed methodology, a Markov process-based performance model has been constructed to analyze the effect of failure and repair rate parameters of various sub-systems, more closely, on system performance. In order to increase the system efficiency and to obtain the optimal combinations of failure and repair rate parameters for various sub-systems, the performance optimization model has been solved through a hybrid BFO-PSO algorithm and their results are compared with basic PSO findings. The following observations can be made from the availability results:\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to optimize the performance for complex repairable system of paint manufacturing unit using a new hybrid bacterial foraging and particle swarm optimization (BFO-PSO) evolutionary algorithm. For this, a performance model is developed with an objective to analyze the system availability.\nDesign/methodology/approach: In this paper, a Markov process-based performance model is put forward for system availability estimation. The differential equations associated with the performance model are developed assuming that the failure and repair rate parameters of each sub-system are constant and follow the exponential distribution. The long-run availability expression for the system has been derived using normalizing condition. This mathematical framework is utilized for developing an optimization model in MATLAB 15 and solved through BFO-PSO and basic particle swarm optimization (PSO) evolutionary algorithms coded in the light of applicability. In this analysis, the optimal input parameters are determined for better system performance.\nFindings: In the present study, the sensitivity analysis for various sub-systems is carried out in a more consistent manner in terms of the effect on system availability. The optimal failure and repair rate parameters are obtained by solving the performance optimization model through the proposed hybrid BFO-PSO algorithm and hence improved system availability. Further, the results obtained through the proposed evolutionary algorithm are compared with the PSO findings in order to verify the solution. It can be clearly observed from the obtained results that the hybrid BFO-PSO algorithm modifies the solution more precisely and consistently.\nResearch limitations/implications: There is no limitation for implementation of proposed methodology in complex systems, and it can, therefore, be used to analyze the behavior of the other repairable systems in higher sensitivity zone.\nOriginality/value: The performance model of the paint manufacturing system is formulated by utilizing the available uncertain data of the used manufacturing unit. Using these data information, which affects the performance of the system are parameterized in the input failure and repair rate parameters for each sub-system. Further, these parameters are varied to find the sensitivity of a sub-system for system availability among the various sub-systems in order to predict the repair priorities for different sub-systems. The findings of the present study show their correspondence with the system experience and highlight the various availability measures for the system analyst in maintenance planning.\n",
"role": "assistant"
}
] |
Why travelers use Airbnb again? An integrative approach to understanding travelers' repurchase intention
|
[
"Repurchase intention",
"Sharing economy",
"Airbnb"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The sharing economy differs from the traditional business model in its dominant reliance on the internet platform, its non-ownership of assets, its ability to access idle resources, its lower prices with more customized products/services and its non-conventional workforce as product/service providers (Belk, 2014). As the most prominent hospitality firm in the sharing economy, Airbnb is an online rental marketplace that provides short-term accommodations to guests as an alternative to hotels (Xie and Mao, 2017). Different from other third-party online travel agent (OTA) platforms such as booking.com that distribute hotel rooms, Airbnb allows individuals, who are distinct from typical business organizations (e.g. hotels), to offer accommodation services to travelers (Tussyadiah, 2016). Hence, Airbnb not only presents a direct challenge to the hotel industry as a disruptive model (Mohlmann, 2015) but also has the potential to induce changes in travel behavior (Tussyadiah and Pesonen, 2015).\n2. Theoretical background: 2.1 The Theory of Planned Behavior\n3. Methodology: 3.1 Data collection\n4. Results: 4.1 Sample profile\n5. Conclusion and implications: 5.1 Conclusions\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to examine the psychological factors that motivate travelers to consider reusing Airbnb.\nDesign/methodology/approach: This study proposes and tests an integrative model that synthesizes the Theory of Planned Behavior (TPB), Prospect Theory (PT) and other Airbnb-relevant constructs (unique experience expectation, familiarity and electronic word of mouth) as the primary determinants of the Airbnb repurchase intention using an structural equation model (SEM) approach.\nFindings: Both attitude and subject norms are significant determinants of repurchase intention, whereas perceived behavioral control is not. In addition, perceived value and risk have only direct significant impacts on attitude and, in turn, indirectly affect repurchase intention. Unique experience expectation, familiarity and electronic word of mouth exert direct and indirect influences on repurchase intention.\nResearch limitations/implications: This study extends the body of knowledge by integrating TPB and PT to investigate consumer repurchase intention in Airbnb, which provides a theoretical baseline and serves as a starting point for exploring the structural relationships of Airbnb and the sharing economy.\nPractical implications: Airbnb should place more emphasis on value packages and authentic/unique travel experience to retain and attract more travelers. More safety/security programs should be put in place and clearly communicated to reduce the perceived risks. In addition, Airbnb may publicize positive word of mouth and introduce and expand familiarity programs to incentivize Airbnb travelers.\nOriginality/value: This study explores the psychological reasons why travelers will re-patronize Airbnb, providing insights into the motives of Airbnb travelers. A clear understanding of Airbnb travelers' repurchase intention will facilitate to develop effective strategies for Airbnb to induce positive repurchase behaviors.\n",
"role": "assistant"
}
] |
A comparison of link and URL citation counting
|
[
"Link analysis",
"Webometrics",
"URL citations",
"Online impact",
"Web sites",
"Search engines",
"Internet"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Link analysis is a broad term used for investigations based primarily upon counting or analysing hyperlinks to or between a set of web sites. For instance, many information science studies have assessed whether counts of links to academic web sites could be used to measure the impact of the owning organisation's research (Thelwall and Harries, 2004), or whether counts of links to journal web sites or individual online articles could be used as an indicator of online impact (Smith, 1999). Hyperlink analysis has also been used for topics as diverse as politics (Adamic and Glance, 2005; Park, 2003), sociology (Rogers, 2005), physics (Barabasi, 2002) and computer science (Henzinger, 2001).Hyperlink investigations have sometimes used web crawlers to visit specific areas of the web (e.g. UK university web sites) but more commonly have used advanced link search queries in the commercial search engines Google, AltaVista, AllTheWeb, Yahoo! or Bing (formerly Live Search) (Bar-Ilan, 2004). Of these, essentially only one is still usable. Google only allows restricted link searches and deliberately reports only a small fraction of the results. Bing/Live Search withdrew its link searches in March 2007 (Seidman, 2007). Yahoo!, AllTheWeb and AltaVista all still work (as of November 2009) but have merged and give very similar results derived from the same crawl database and so effectively only Yahoo! remains. When Microsoft took over Yahoo! in mid 2009 this created the possibility that Yahoo!'s link searches would be phased out to align it with Bing. This would be a serious problem for link analysis because certain types of link analysis would become impossible, particularly those analysing links to sites from anywhere in the web (Thelwall, 2004).Although no general alternatives to link counts have been proposed for academic web sites, two main alternatives have been proposed for online journal articles: URL citations and text searches. An URL citation is a mention of the URL of a web page in another page, whether or not it is accompanied by a hyperlink. In other words an URL citation is an URL visible to a web page visitor but is not necessarily clickable. Searching for web pages mentioning URLs or parts of URLs has also been previously discussed outside academia (Search Engine Promotion Help, 2004) but without naming the concept. URL citations have the advantage that they can be identified using standard queries rather than advanced link search queries and can therefore be calculated by any search engine. Although URL citations are numerous enough to be useful for online academic journal article impact assessments (Kousha and Thelwall, 2006, 2007) no previous study has assessed their prevalence across a range of types of web site, comparing them to link counts. This paper fills this gap so that if link searches are withdrawn from Yahoo! then the implications of switching to URL citation searches will be known. The other alternative to link counts, text searches, are queries for articles using text-based methods such as phrase searches for article titles (Vaughan and Shaw, 2003). This paper focuses only on URL citations, however, as these are, in principle, more general: applicable to any context in which links could be counted. In summary, this study uses sets of URLs and/or domain names from a range of different webometric studies and compares the counts of links to them with their URL citations, as reported by Yahoo!.\nResearch questions: This research is an initial investigation to begin assessing the extent to which URL citations could be used as an alternative to link counts for webometric research. In particular, the following broad questions drive the investigation:1. Are URL citation counts sufficiently numerous in all webometric contexts to be used as a substitute for link counts?2. Are URLs with path information following the domain name rarely used for URL citations in comparison to links?The second research question relates to the need to use different advanced search queries depending upon whether an URL contains a path component, as described below. Note that although there has been much research showing that search engine hit count estimates can be unreliable in many different ways (Bar-Ilan and Peritz, 2004; Mettrop and Nieuwenhuysen, 2001; Thelwall, 2008; Uyar, 2009) and search engines do not index the whole web (Lawrence and Giles, 1999), the current paper is only concerned with the estimates returned by search engines and not the actual number of pages matching the search and available on the public web.\nMethods: The research questions are too broad to be fully answerable with empirical evidence and so the overall approach is to conduct a set of illustrative experiments and to use the results to ground a discussion of the likely answers to the research questions.Each experiment is based upon a set of URLs or domain names used as part of a webometric study at the University of Wolverhampton. This choice was made in preference to an artificial collection of random URLs because it reflects types of URL collections used in webometric research. The use of different sets thus follows the hypothesis that the extent of use of URLs in web sites will vary greatly by web site type or topic, and so it is necessary to test different sets separately. Three additional data sets were added, however, denoted by a star*, to cover additional types of data set of relevance to the study. The following data sets were used:* Bengali - URLs of Bengali web sites mentioning the Mumbai attacks.* BLnews - URLs of web pages linking to a British Library newspaper archive collection.* Chem* - The Dmoz.org Business/Chemicals/Diversified Manufacturers category.* ChemW* - As for Chem but appending www. to the canonical domain name.* EUunis - canonical domain names of European universities identified in the year 2000 (without any initial www.).* EUunisW* - As for EUunis but appending www. to the canonical domain name.* Hindi - URLs of Hindi web sites mentioning the Mumbai attacks.* LifeSci - URLs of pages linking to European life science research groups, as extracted for the NetReAct project.* MySpace - URLs of pages linking to social network site MySpace.* RAI - URLs of pages linking to the Italian news portal www.rainews24.rai.it* Spaces - URLs of pages linking to Microsoft Live Spaces blogs.* Religion - Domain names of blogs mentioning a major religion. 1,000 were randomly selected from an initial set of 59,367.* RuDe - URLs of web sites used by Russians living in Germany.* Urdu - URLs of Urdu web sites mentioning the Mumbai attacks.* ZigZag - URLs of web sites linking to the BBC World Service Trust ZigZagMag online magazine. These are mainly Persian discussing news, politics and culture and the set includes many blogs. 1,000 were randomly selected from 7,849.For each URL in the data sets, site inlink counts were estimated using Yahoo! advanced link searches as follows. If the URL contained only a domain name then the linkdomain search command was used to identify all pages linking to anywhere in the web site (e.g. linkdomain:www.wlv.ac.uk-site:wlv.ac.uk). If the URL also contained file or other path information then a link search command for links only to the individual page was used instead (e.g. link:www.wlv.ac.uk/home.htm-site:wlv.ac.uk).For each URL in the data sets URL citation counts were estimated in Yahoo!! using its API in LexiURL Searcher. For a URL u with web site domain name s the http:// was removed from u and then the search used was \"u\"-site:s. The quotes around the URL ensure that the segments of the URL are matched together and the -site:s part ensures that only web pages outside the hosting web site are included. The former is an innovation for the current study but the latter is standard webometric practice.Note that for some types of URL the URL citation search has a slight mismatch with the link search. For URLs with path information but not a filename, such as www.wlv.ac.uk/computing/ the link search link:http://www.wlv.ac.uk/computing/ -site:wlv.ac.uk matches links only to the exact URL www.wlv.ac.uk/computing/. In contrast, the URL citation search \"www.wlv.ac.uk/computing/\"-site:wlv.ac.uk matches this URL and all URLs starting with http://www.wlv.ac.uk/computing/ such as http://www.wlv.ac.uk/computing/databases/. Most links seem to target web site home pages and so this mismatch is likely to normally have minor repercussions but the issue would be significant for large web sites with deep directory structures, such as dmoz.org.\nResults and discussion: Table I reports the results of the link and URL citation searches, showing that the correlation between the two is highly significant and positive in all cases. The remaining statistics include only results for which one of the two was non-zero. In cases where both URL citations and links are not found, it seems likely that neither exist. Although this data could be left in Table I, it would result in inflating the apparent agreement between URL citations and links for data sets in which there were large numbers of zeros in the results and hence the figures would not fairly reflect the extent of agreement when there was at least a possibility of disagreement. In addition, if both values are zero then the ratio between them is undefined and so the ratio column medians could not be calculated.The YC/YL median column shows that there is a great difference between data sets in the URL citation figures compared with the link count results. Whilst for the majority of data sets there are fewer URL citations than link counts, the reverse is true for EU universities and the chemical companies. In the former case this may be because the universities' domain names are relatively short but also shows that URL citations are very common in academia. URL citations seem also to be reasonably common for business web sites. An inspection of the results suggested that many of the URL citations in the EUunis group were email addresses. This was possible because the university domain names used omitted any initial \"www.\" and only contained the generic domain name ending for each university (e.g. wlv.ac.uk rather than www.wlv.ac.uk). Yahoo!'s algorithm could then match this curtailed domain name to email addresses (e.g. m.thelwall@wlv.ac.uk). An email address is not an URL and therefore \"URL citation\" is a misnomer for data that includes email addresses but an email address is nevertheless arguably a type of citation as it points to the organisation employing the individual owning the address. The lack of email in the EUunisW and ChemW groups is probably the reason for the lower ratios compared to the EUunis and Chem groups respectively.The last three columns of Table I show that in most cases there is a significant difference between the ratio of URL citations to links for URLs depending upon whether the URL has a path. The test is only non-significant in cases where there are less than five values in one of the two groups. When significant, the difference is that there is a higher ration median for URLs without a path than for URLs with a path. It seems therefore that URL citations are relatively less common when the URL has a path. This seems to be a logical explanation because an URL with a path would be long and may not fit well into the text of a web page or may be less useful to copy. Alternatively, however, it may be that Yahoo! has more difficulty to identify URL citations for URLs with a path component. Recall that URL citation searches are more generous than link searches and so the difference in matching algorithms cannot explain the lower number of matches for the former.The main limitation of this investigation is the restriction to 15 case studies. The fact that the results differ widely between cases highlights this factor. In consequence, it is not possible to predict the impact of switching to link analysis to URL citation analysis for any particular future study; although the results suggest that outside academia there will tend to be about a tenth as many URL citations as hyperlinks. Moreover, although all of the correlations between link counts and URL citation counts are positive, the same may not be true in other cases. Finally, this study has not analysed the impact on the outputs of link analyses (e.g. network diagrams, link creation motivation classifications) of a switch to URL citations and so despite the positive correlations between link counts and URL citation counts, they may still produce different findings in practice even when there is enough data for both to be feasible.\nConclusions: The results suggest that URL citations are likely to be less numerous than links in webometric studies of non-academic spaces. The shortfall depends upon context but for most of the data sets included there was less than one URL citation for about every ten links. For these examples the difference may be sufficiently serious - reducing the median to zero in several cases - that there would be too few URL citations for most webometric purposes, except perhaps for identifying the most important web sites or web pages.The shortfall caused by a switch from links to URL citations can be partly redressed in cases involving domain names alone by omitting the initial \"www.\" part of any web site so that the URL citation search also includes email addresses associated with the domain. For academic web sites and the chemical companies this resulted in an increase in the number of citations identified, but this method probably only works for main domains having their own email addresses (e.g. whole universities, large government organisations or businesses) but not for subdomains (e.g. university departments or research groups, blogs, or personal web sites hosted by internet service providers).In conclusion, if Yahoo! and its affiliates withdraw their advanced hyperlink queries so that URL citations would have to be used instead of links for many studies, then this would seriously undermine the power of link analysis, by a factor of about ten, except for academic spaces. This would probably mean that only very large-scale link analyses and academic link analyses could be sustained. This would be a serious threat to webometrics since even recent academic link analyses have tended to focus on the relationships between academic and non-academic web sites. The gap could be partially filled by studies using web crawlers, but these are only useful for analyses of the links from or between specified sets of web sites (Thelwall, 2004), whereas the links to web sites from the rest of the web are often the most interesting.\n",
"role": "user"
},
{
"content": "Purpose: - Link analysis is an established topic within webometrics. It normally uses counts of links between sets of web sites or to sets of web sites. These link counts are derived from web crawlers or commercial search engines with the latter being the only alternative for some investigations. This paper compares link counts with URL citation counts in order to assess whether the latter could be a replacement for the former if the major search engines withdraw their advanced hyperlink search facilities.\nDesign/methodology/approach: - URL citation counts are compared with link counts for a variety of data sets used in previous webometric studies.\nFindings: - The results show a high degree of correlation between the two but with URL citations being much less numerous, at least outside academia and business.\nResearch limitations/implications: - The results cover a small selection of 15 case studies and so the findings are only indicative. Significant differences between results indicate that the difference between link counts and URL citation counts will vary between webometric studies.\nPractical implications: - Should link searches be withdrawn, then link analyses of less well linked non-academic, non-commercial sites would be seriously weakened, although citations based on e-mail addresses could help to make citations more numerous than links for some business and academic contexts.\nOriginality/value: - This is the first systematic study of the difference between link counts and URL citation counts in a variety of contexts and it shows that there are significant differences between the two.\n",
"role": "assistant"
}
] |
Sharing resources of Russian libraries: 10 years of consortia services development
|
[
"Cooperation",
"Library services",
"Document delivery",
"Databases",
"Consortia",
"Resource sharing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The vast area of Russia has given rise to a large number of libraries. As a result, interlibrary communications face several basic problems. Creating a reliable infrastructure for efficient resource sharing was hardly possible before the new information age. Now that libraries have become part of the Internet, they are potentially accessible and searchable from any computer connected to the network. Effective and efficient resource sharing between libraries thus became possible. This article will describe the developments and activities of the Associated Regional Library Consortia (ARLICON)[1], one of the most effective resource sharing networks in Russia.\nThe Russian context: The treasures of Russian libraries are associated by foreign partners with only a few major ones in cities: the National Library of Russia in St. Petersburg, the Russian State Library in Moscow, the Library of Foreign Literature in Moscow, the Library of the Russian Academy of Sciences in St. Petersburg and several others. The collections of many other libraries are hidden, unknown and not accessible.\nProject MARS - the bibliographic database of Russian journals indexed to article level: The birth of this project was initiated by libraries conscious that duplication of cataloguing and processing journals was wasteful. Cooperation was simple - the journals being processed were distributed between libraries, which then shared the bibliographic records created between the involved parties. The choice of journals for which all articles would be indexed consisted of those most frequently used by involved libraries.\nThe electronic document delivery service: This service logically follows from the MARS service; when the record is found, the user often wants the journal article. If it is not available online via the link within the record in the MARS database, the user can ask their library to request a copy.\nILL and EDD: In this section, EDD service refers to documents held and supplied electronically. ILL refers to documents (both journals and books) delivered by conventional means such as post. ILL/EDD refers to both services.\nARLICON services today and tomorrow: The ARLICON portal provides front-end services for all users and some library specialists.\nConclusion: The library landscape of Russian Federation is rapidly changing, following world trends, but conditioned by the specific limits and properties of their country. Consortia technologies implemented by ARLICON and its partners provide visibility and accessibility for many diverse library sources from different regions of Russia. ARLICON has become the starting point for information search and retrieval.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to demonstrate the current situation and history of cooperation and resource sharing in Russian libraries. The resources and services provided cooperatively are described.\nDesign/methodology/approach: A narrative combining description, surveys, analysis and evaluation.\nFindings: Fundamental changes have occurred in Russian libraries as a result of the development of consortia services development, particularly in resource sharing.\nResearch limitations/implications: Digital libraries and databases subscriptions were not within the scope of this study.\nSocial implications: The paper demonstrates the efforts of many libraries cooperatively producing services at a national level.\nOriginality/value: One of the few articles to describe the current state of resource sharing and consortia development in Russia.\n",
"role": "assistant"
}
] |
The fallacy of the multi-API culture: Conceptual and practical benefits of Representational State Transfer (REST)
|
[
"Web applications",
"Information architecture",
"Hypermedia",
"REST",
"Uniform interface",
"Web APIs"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: 1.1. General context\n2. Short history of web architecture: 2.1. Relevance of original Hypertext Transfer Protocol (HTTP) design\n3. The REST model and its implications: 3.1. Definition\n4. A sustainable REST architecture for HTTP servers: 4.1. A generic architecture to support change\n5. Use case: implementing REST at the Cooper-Hewitt Museum: 5.1. The Cooper-Hewitt collection and its collection management database\n6. Conclusions: This paper brought attention to the relevance of REST principles for the LIS discipline. More in particular, we focused on how the REST architectural style can contribute to the longevity of URLs, which are playing a central role within the application of Linked Data principles and the automated usage of documents and their metadata.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to revisit a decade after its conception the Representational State Transfer (REST) architectural style and analyzes its relevance to address current challenges from the Library and Information Science (LIS) discipline.\nDesign/methodology/approach: Conceptual aspects of REST are reviewed and a generic architecture to support REST is presented. The relevance of the architecture is demonstrated with the help of a case study based on the collection registration database of the Cooper-Hewitt National Design Museum.\nFindings: The authors argue that the \"resources and representations\" model of REST is a sustainable way for the management of web resources in a context of constant technological evolutions.\nPractical implications: When making information resources available on the web, a resource-oriented publishing model can avoid the costs associated with the creation of multiple interfaces.\nOriginality/value: This paper re-examines the conceptual merits of REST and translates the architecture into actionable recommendations for institutions that publish resources.\n",
"role": "assistant"
}
] |
Keeping found things found: Challenges and usefulness of personal information management among academicians
|
[
"Perceived usefulness",
"Information fragmentation",
"Challenges of PDIM",
"Engineering faculty",
"Memory load",
"Pakistan University",
"Personal digital information management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Growth of digital contents in the form of websites, databases, repositories and social media expanded the horizons of information landscape. Information explosion, technological advancements and use of digital technology in teaching and learning make the management of faculty's personal information more challenging than ever before. They need to manage their information in a systematic way for future retrieval. The main purpose of managing and organizing personal collection is to retrieve and re-find information in future for personal use (Diekema and Olsen, 2014).\nBackground: People gather their needed information both actively (through web searching) and passively (emails). Albeit, they have to decide either keep or leave that gathered information (Bergman, 2013). Now, semantic Web technologies, information explosion and technological advancement affect the way academicians find and re-find their personal information. For handling overloaded information, PDIM is useful for academicians to find information from various sources such as libraries, internet, friends, observations, surveys and assignments and refind information for potential future use. Teachers can improve their teaching and research productivity if they manage their information properly (Truyen, 2010):\nObjectives of the study: The following research questions were formulated to identify usefulness and challenges of PDIM:\nMethodology: Quantitative research approach was used to identify the usefulness and challenges of PDIM. Previous literature revealed that the quantitative method was used widely for PDIM research. Survey method was opted to meet the objectives of the current study. Population of the study was 608 faculty members working in 24 departments from one of the oldest engineering and technology university in Pakistan. A list of all the faculty members was prepared for sampling purpose. Simple random sampling technique was used to select representative sample. In total, 340 questionnaires were distributed among faculty members. Data were collected through personal visits and with the help of departmental librarians. Next, 257 questionnaires were received, yielding a response rate of 72 per cent. Also, 12 incomplete questionnaires were excluded, and thus, 245 questionnaires were used in final data analysis.\nFindings: Demographic information\nDiscussion and conclusion: The first part of the study is the usefulness of personal information management. Faculty members showed positive behavior in this regard. They professed that PDIM practices improve their teaching and learning skills, along with the effective utilization of resources. They agreed that information management saves their time and energy.\nFuture studies: There is need to explore the relationship among \"keep everything\" and \"information overload\" and between \"information fragmentation\" and \"digital archiving\" challenges. This study recommends conducting further research on information keeping behavior of individuals and relationship among five major challenges of PDIM (predicting future value of information, information keeping or leaving, fragmented information, information overload and archiving digital material). This study also recommends that a survey should be conducted to measure computer self-efficacy, internet self-efficacy, Web self-efficacy and mobile computing self-efficacy of faculty members because these efficacies are significant for PDIM practices. It is recommended to incorporate techniques to predict the future value of information in existed information literacy instruction programs for faculty members to deal with existing challenges in PDIM. It is also suggested to develop PDIM system incorporating semantic technologies to mitigate memory load and other related challenges. This technology established connections among users' different type of personal information such as documents, emails, photos, audio and video files.\n",
"role": "user"
},
{
"content": "Purpose: This study aims to investigate the usefulness and challenges of personal digital information management (PDIM) as perceived by engineering faculty members.\nDesign/methodology/approach: Quantitative research strategy based on survey design was used to collect data through structured questionnaire. Simple random sampling technique was applied to obtain a representative sample from faculty members of University of Engineering and Technology, Lahore.\nFindings: Findings indicate that faculty members understand the usefulness of PDIM practices for their teaching and learning. They opined that these practices improve their performance along with effective utilization of resources. PDIM practices save their time, energy and money. Nevertheless, they have to face challenges in managing their personal information in digital format including technology obsolescence, bulk of incoming emails, prediction of future value of information, information fragmentation and memory load in remembering the location of information. It is also noted that PDIM practices are equally useful for both male and female faculty members. However, female faculty members have to face more PDIM challenges than male faculty members.\nPractical implications: Overall, findings of the study reveal that there are benefits of PDIM and it is useful for faculty members with certain challenges. This study identified five major challenges. All these challenges were from technological and cognitive psychology domain. It is noted that assessing future value of information plays important role in overall process of PDIM.\nOriginality/value: The present study would be valuable addition in literature and it would be helpful in devising information literacy instructions programs in academic libraries. This study would also helpful for improving quality services for teaching as required by Pakistan Engineering Council. Findings of this study would helpful for system engineers, software developers and system librarians in designing and developing personal information management systems in Pakistan.\n",
"role": "assistant"
}
] |
Web structure and influence of the Arab universities of the MENA zone (Middle East and North Africa): Visualization and analysis
|
[
"Social network analysis",
"Arab universities",
"Influence",
"Information visualization",
"Web ranking",
"Webometrics",
"MENA"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The geopolitical zone known as MENA (Middle East and North Africa), occupying so many newspaper front pages last year owing to the \"Arab spring\" in Tunisia, Egypt, Libya and Syria, is at the same time largely ignored by the realm of webometrics, scientometrics, and information visualization. The MENA zone would include all 22 countries whose official language is Arabic, and which also belong to the League of Arab States (www.arableagueonline.org): Saudi Arabia, Algeria, Bahrain, Comoro Islands, Djibouti, Egypt, United Arab Emirates, Iraq, Jordan, Kuwait, Lebanon, Libya, Morocco, Mauritania, Oman, Palestine, Qatar, Syria, Somalia, Sudan, Tunisia and Yemen. The MENA zone takes in countries of two geographic areas - the Middle East and northeastern Africa - that are often studied separately, despite their many cultural, social, economic, political and religious points in common. Precisely these aspects, plus the relations or interactions among the respective countries, governments, organizations and institutions, and their degree of development, become determinant factors in the overall influence of a country, the intellectual weight of its institutions in the regional context, and the predominance of certain universities. Studies of interaction among universities are usually approached through their web links, which serve to gauge the comparative influence of each academic institution.\n2. Materials and methods: 2.1 Source data\n3. Results and discussion: 3.1 Relationship between number of universities and Inlinks by country\n4. Conclusions: We attained our main objective of reflecting the structure and influence of the World Wide Web in the MENA zone, at the geographic level (country) and the academic level (universities). Studies of this type are no longer possible because the search engines no longer provide the commands that allow searches of links or co-links. Hence, this study stands alone in the rink of webometric literature, at the same time a pioneer and the last of its kind. At present, it is only possible to consult URL citations and optionally find mentions in text.\n",
"role": "user"
},
{
"content": "Purpose: This article aims to display the structure and reveal the web influence of institutions in the MENA zone, in geographic terms (country) and academic terms (universities), by means of their links.\nDesign/methodology/approach: Using search engines and webcrawlers designed to gather information about web links, in conjunction with visualization techniques and degree indicators based on social network analysis, the authors achieved their objective and found responses to a series of pertinent research questions.\nFindings: There is no direct relationship between the number of university websites and the number of inlinks. Linking between countries in the MENA zone obeys patterns of vicinity and geopolitics. Arab universities are interlinked following trends governed by territorial proximity. There is a strong endogamic tendency, with universities from a single country citing each other, particularly in the case of Saudi Arabia. The authors present the first ranking of web influence in the MENA zone based on network indicators, namely country and university, and their order is corroborated by comparison with other rankings of a webometric or scientometric nature.\nResearch limitations/implications: Studies of this type cannot be undertaken again, at least not from the web link perspective, as Yahoo!, Google and Bing have since blocked the webcrawlers that attempt to carry out searches of inlinking or co-inlinking between/among sites. Hence, this work can be considered both a pioneer and the last of its kind. The authors do not know if or when it will be possible to again make queries about URLs in webs or, alternatively, in titles.\nOriginality/value: This is the first visual report of the web structure underlying the countries and universities of the MENA zone. It is also the first time that a country and university ranking of this geopolitical zone has been carried out using network indicators based on web links.\n",
"role": "assistant"
}
] |
Exploring a suitable model of destination image: The case of a small-scale recurring sporting event
|
[
"Destination image",
"Sport tourism",
"Tourist satisfaction",
"Attitudinal loyalty",
"Behavioural loyalty",
"Destination loyalty",
"Small-scale sporting event"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Tourism remains one of the largest and fastest-growing economic sectors; bookings in 2017 reached almost $1.6 trillion and tourism accounted for a staggering 10.2 percent of global GDP (gross domestic product) (Now Boarding, 2018). Some cities in Asia, such as Bangkok, Dubai, Singapore, Kuala Lumpur, Tokyo, Seoul, and Hong Kong are highly popular tourism destinations and draw tourists from all over the world (Murray, 2018). Hong Kong attracted about 30m international visitors in 2018, which represented an annual increase of 7 percent, and Bangkok welcomed nearly 24m visitors, which was 5.5 percent greater than achieved in 2017 (Petroff, 2018). The South Korean government has focused on the tourism industry for many years because the industry has unlimited global potential and boosts domestic consumption. In South Korea, tourism was worth $65bn and accounted for 2.5 percent of GDP in 2015 (Park, 2017). Recently, thanks to the huge popularity of Korean pop culture, the number of foreign tourists has continually increased (Bae et al., 2017; Teh and Goh, 2016). In 2018, an estimated 15m international tourists visited South Korea and spent approximately $16bn (Lee, 2019).\nLiterature review and research hypotheses: Reasons behind destination travel\nMethod: Data collection\nResults: The fitting indices of the structural path model results were as follows: kh2/df=2.473, NFI=0.916, CFI=0.900 and RMSEA=0.057. As previous studies have suggested a NFI and CFI of >0.9 (Hair et al., 2006) and a RMSEA of <0.8 (Browne and Cudeck, 1993) model fit was satisfactory.\nDiscussion: Academic implication\nConclusion and limitation: Conclusion\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the structural relationships between destination image, tourist satisfaction, attitudinal loyalty and behavioral loyalty, with an emphasis on the mediating effect of tourist satisfaction on the relation between destination image and loyalty in the context of a small-scale recurring sporting event held in Asia.\nDesign/methodology/approach: Validity and reliability of the measurement scale were proved through a confirmatory factor analysis, Cronbach's a analyses and correlation analyses. A structural equation modeling test with maximum likelihood estimation was conducted to test the relationships among the research variables using 440 participants.\nFindings: The results revealed destination image had a direct influence on tourist satisfaction, attitudinal loyalty and behavioral loyalty, and that tourist satisfaction had a direct influence on attitudinal loyalty and behavioral loyalty. Moreover, tourist satisfaction was found to partially mediate relationships between destination image and attitudinal loyalty, and between destination image and behavioral loyalty.\nPractical implications: First, destination marketers and organizers of a small event should provide tourists with an international or domestic newsletter of the small event. Second, destination marketers should place well-educated employees at popular hotels and restaurants and the event organizers should arrange that trained volunteers be positioned at stadiums, to enable tourists to find the locations of interest, which would help develop a positive image of the destination. Third, marketers and organizers should actively use social media to improve destination images and promote sporting events.\nOriginality/value: The authors offer a new perspective of tourist satisfaction as a mediating effect. Existing studies show tourist satisfaction fully mediates on the relation between destination image and loyalty, but the present study shows tourist satisfaction partially mediates this relation. In this respect, the term \"sporting event\" should be regarded important when attempting to understand tourist psychology and behavior because the level of tourist satisfaction can be affected by a term like \"sporting event\" in the mediating effect context.\n",
"role": "assistant"
}
] |
The impact of word of mouth on service purchase decisions: Examining risk and the interaction of tie strength and involvement
|
[
"Influence",
"Risk",
"Involvement",
"Word of mouth",
"Tie strength",
"Service purchase decision"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Service purchase decisions are ubiquitous and seemingly omnipresent. Think about the last time that you ate at a new restaurant, took your car to a mechanic, got a new babysitter, or sought information on a suitable physician or dentist. Conversely, perhaps you have recently rejected a particular hairstylist or veterinarian. Chances are you received information about that service provider via word of mouth (WOM). It might have come from a close family member, work colleague, or a casual acquaintance. You might have accepted or rejected the WOM advice, however, it almost certainly influenced your service purchase decision, especially if that decision was highly relevant to you. Why did this occur? In terms of the overall WOM influence on you, was there an inter-play between your relationship with the WOM sender and the relevance of the service purchase decision? What is it about WOM that can be so overwhelmingly powerful? Importantly, how can managers harness the power WOM? The aim of the current study is to answer these key questions.\nTheoretical underpinnings and research hypotheses: WOM research (scholarly and practitioner oriented), has been growing exponentially since the mid-1990s. Some studies have sought to understand the effects of interpersonal constructs (e.g. tie strength between a sender and receiver) and non-interpersonal constructs (e.g. perceived risk) on key dependent variables (e.g. WOM influence) (Sweeney et al., 2014; Jun et al., 2011; Stephen and Lehmann, 2009; Bansal and Voyer, 2000). Unknown is the extent to which the strength of the bond between consumers, i.e. tie strength, will interact with service purchase decision involvement to affect WOM influence.\nResearch design, method, and results: Research setting, participants, and measures\nDiscussion and implications: Service marketing is fundamentally about relationships. Consumers can be engaged in a variety of relationships (e.g. with each other, service providers, brands, firms, etc.). Consumer interrelationships serve as conduits for WOM processes. In this research, WOM influence was examined by measuring how it was impacted by both consumers' interrelationships, and by their purchase decision involvement. As expected, finding indicated that both tie strength and involvement positively and directly affected WOM influence. However, more important was the negative moderation effect whereby, tie strength was found to diminish the effect of involvement on WOM influence. This key result makes a significant contribution to WOM theory, and also has important managerial implications.\nLimitations, future research, and conclusion: Limitations\n",
"role": "user"
},
{
"content": "Purpose: Primarily, the purpose of this paper is to investigate the interaction and direct effects of tie strength between sender and receiver of word of mouth (WOM) and the receiver's service purchase decision involvement on WOM influence. A secondary aim is to investigate how a distinctive conceptualization of perceived risk, consisting of two types (outcome risk and psychosocial risk), affects service purchase decision involvement. A conceptual model incorporating these constructs and associated hypotheses is developed and tested.\nDesign/methodology/approach: In a survey of actual service consumers, respondents were asked to recall a recent instance where they had received service purchase information via WOM, and relate their responses to this instance. Established scales were used to measure the constructs. The hypothesized model was tested using structural equation modeling.\nFindings: Principally, findings demonstrate a strong interaction effect between service purchase decision involvement and tie strength. Also, results highlight the complexity of the perceived risk construct, suggesting that it is appropriately modeled as two types: outcome risk, and psychosocial risk.\nResearch limitations/implications: This research has contributed to the service marketing literature by testing a model that predicts WOM influence. Evidence confirmed that the effect of service purchase decision involvement on WOM influence is moderated by tie strength. Additionally, a conceptualization of two different types of risk associated with purchase decisions was suggested, together with empirical confirmation of their hypothesized antecedent effects on service purchase decision involvement. Findings have special implications for the literatures of persuasion, social and interpersonal influence, as well as consumer behavior in general.\nPractical implications: To harness the power of WOM, managers should understand who their target audience is and how consumers are related to each other (tie strength) and to the service purchase decision (service purchase decision involvement). Recommendations are made with specific illustrations of how firms can leverage tie strength under conditions of low service purchase decision involvement to enhance WOM influence.\nOriginality/value: The formidable power of WOM wields substantial influence on consumers, particularly within a service (vs goods) purchase context, typically characterized by higher perceived risk and lower search qualities. The significant interaction between tie strength and service purchase decision involvement is a unique contribution to the service WOM literature.\n",
"role": "assistant"
}
] |
Literature review of managerial levers in primary care
|
[
"Literature review",
"Managerial levers",
"Primary care"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: One of the strategies adopted in response to the recent challenges in healthcare is the shift of the healthcare focus from the hospital level to the local and domicile levels (Dirindin and e Vineis, 2004; Compagni et al., 2010; Creazza et al., 2016), with chronic diseases and diseases that do not require highly intensive treatments being managed at the primary care level versus the hospital level being devoted to acute-phase treatment for short periods only.In this view, the provision of services and the distribution of health technologies toward the primary and community care levels imply the ability to organize \"in-field\" assistance, thus assuring greater appropriateness and timeliness (NHS England, 2019).The provision of healthcare at the primary care level is complex, as services are heterogeneous (including prevention services, hospice care, mental health services and home care) and care programs are customized and fragmented (Matta et al., 2014). Therefore, a multidisciplinary network of professionals and resources is involved, and timely coordination of their activities is crucial. In addition, this context is constantly evolving (Longo et al., 2010), and it proves to be challenging to be formalized into a complete and unambiguous picture.In this complex and articulate setting, managerial levers (i.e. activities and managerial tools allowing the correct functioning of healthcare processes - such as information technology, logistics, human resources, etc. useful to optimize organizations' resources and maximize efficiency and effectiveness (Azzone Bertele, 2009)) could be applied to improve healthcare service provisions. Because the population of patients with long-term conditions being treated also at home is expected to increase worldwide due of the ongoing growth in life expectancy (World Health Organization WHO, 2018), improving the care of these categories of patients represents a priority for the future. This goal could be achieved through the design and management of appropriate healthcare services, assuring the economic and financial sustainability of national health services (NHSs) (Mosca et al., 2017).This complex, articulate and multidisciplinary setting is not yet definitively consolidated, also in the literature, a gap in the analysis of the fragmented body of knowledge on the topic was noticed (Genet et al., 2011; Kringos et al., 2013). Therefore, through a thematic analysis, the aim of the present study was to analyze the existing literature on the managerial levers that have been adopted for the provision of primary and community care services in general and in particular for the specific target population abovementioned. This investigation sought also to unveil those levers scarcely explored but which could potentially be adopted in the future to provide high-quality care and assistance at the primary and community care settings. This research is expected to serve well to advance the level of knowledge on the topic and to pave the way to designing more effective empirical investigations that could contribute to improving the current level of practice in the provision of these healthcare services, as advocated by recent literature (e.g. NHS England, 2019).\n2. Research methodology: A literature review was carried out with an ultimate aim of isolating and exploring the managerial levers contained in a number of topics that were previously addressed in the scientific literature.A structured literature review (SLR) approach was adopted (Petticrew and Robert, 2006), and a thematic analysis was performed. SLR is an evidence-based method that through a structured and repeatable process aims to minimize bias through the identification, evaluation and selection of all existing and relevant studies in response to one or more specific questions. 2.1 Locating studies\n3. Results: 3.1 Search flow\n4. Discussion: It should be noted that the topic of the provision of primary and community care, considered from a managerial point of view (and especially for the patient with long-term conditions treated also at home), has gained interest in recent years, although the scientific evidence does not allow for a complete overview of all aspects related to the abovementioned services. Therefore, the field of research dealing with the management of primary and community care services could be not considered as a specific literature stream, and it has not yet reached a maturity phase, with space still being present for new contributions to provide managerial insights for the improvements of the services while assuring quality and continuity of care.The topic of primary and community care services provided is very complex (Chahed et al., 2009) and often difficult to formalize (Matta et al., 2014), as assistance at the local level requires the involvement of professionals from different organizations who should be coordinated in the best possible way, with an adequate information system, dynamism and flexibility (Batet et al., 2012) to meet the needs of the patient (Abusalem et al., 2013). The present study reveals that the literature addressed this issue from different points of view, thus revealing a fragmented and often incomplete vision. In this sense, health services research needs to pay particular attention to the coordination and sharing of information to improve access, quality and costs in a constrained and rapidly changing environment (Moore et al., 2016). In this view, healthcare decision-makers assume a key role in the planning of services and in optimizing the utilization of the resources, considering also the professionals' experience. This suggests that dedicated focus groups having different competencies (e.g. managerial, technical, clinical) should be considered for inclusion in the phase of decision-making.The results of the present review show that topics already largely explored in the primary and community care services literature at this time were related to information technology/innovation management and policy/regulation. This seems to be because, on the one hand, of the relevance of technological improvements in the sector (e.g. telemedicine), and on the other hand, to the importance of the strategic and regulatory issues for this sector and their implications on policy-making. However, our study unveils that these two main areas are not connected, and this further shows and confirms the lack of a holistic view in the investigated research field.From our literature review, it seems that the themes of logistics/supply chain management, economic evaluations, performance management and customer satisfaction have been less analyzed than others, not because they were considered unattractive or without the possibility of an effective application, but rather because few authors approached these topics with an ability to collect data from the field or test or apply hypotheses and theories. This is probably due to the fragmented and multifaceted nature of this topic, which makes it difficult to find suitable application fields.Finally, the topic of performance in the provision of these services is recognized as a difficult area due to the unclear and complex boundaries of the unit of analysis (especially if compared to the hospital setting; Matta et al., 2014), but research on this topic should not be discouraged (Murillo-Zamorano and Petraglia, 2011). The present literature review revealed additionally the presence of papers that embrace more than one management topic, confirming the multidisciplinary nature of primary and community care services. Indeed the examination of the results revealed that most of the selected papers focused on a specific topic of management and/or a specific process and/or a specific target population, lacking a holistic and integrated view.This lack originates also from the different point of views of the actors involved in the healthcare process (e.g. service provider, patients, human resources) that lead to a fragmented vision of the process, not permitting interventions that may improve the performance of the service in terms of effectiveness and efficacy for each actor involved. This is also exacerbated by the level of analysis: in particular, the micro level, as previously shown, did not allow an overall overview.Also, the papers that hold a macro perspective did not allow a holistic view, probably because of the ease of addressing the issue while assuming the following perspective: data at the macro level are more difficult to collect from an empirical point of view because of the complexity of the studied setting. Therefore, it emerges that today, there is a call for a better coordination and integration of care services, while the execution and the managerial levers are still patchy and left behind in the policy agenda. These considerations emerged also from a methodological perspective: the results showed that the topic of the provision of healthcare services to patients with long-term conditions was addressed using several techniques, and this fact confirmed the complex and various nature of primary care services.In summary, as the area of healthcare services at the local level is becoming more important and the present literature review revealed several underexplored areas, we propose some possible priorities for future research (Figure 3).\n5. Conclusion: In the present study, we analyzed the existing literature on the managerial levers adopted for the provision of primary and community care services. From a theoretical viewpoint, a first contribution of this literature review is represented by the attempt to rationalize the fragmented body of knowledge on the topic of the provision of healthcare services at these levels. As an additional contribution, this allowed for the shedding of some light on the managerial levers already explored in the literature and also contributes to create awareness regarding the need for further investigations that could advance the level of current knowledge in the field, thanks to the trajectories identified for future research.In terms of practical implications, the present study provides practitioners with an indication of the managerial levers that seem to be consolidated in the practice, offering a vast array of examples and suggestions for designing and organizing better systems for providing healthcare services in the primary care settings. The present review may be used to identify and suggest interventions for management in the provision of healthcare at this level. Raising awareness could have implications also in terms of policy-making. In fact, this study could stimulate policymakers to look at the provision of healthcare services in the primary and community care settings with a renewed interest and to promote interventions also in those areas that seem to be neglected but relevant in terms of cost reduction and quality of the provided services.Importantly, it should be noted that this study has some limitations, common to the practice of conducting literature reviews; these include the selection of keywords in a diverse and multifaceted environment, which still needs to reach its maturity, the process of the selection of studies, the criteria for the inclusion of papers, the choice of relevant literature, the methods of analysis, the interpretation of results and the clarity of reporting, which could all be affected by some bias due to the personal views and backgrounds of the researchers. However, the adoption of a systematic approach is intended to mitigate these potential issues (Denyer et al., 2003).Despite limitations, we believe that consideration of the produced evidence and identified research trajectories could advance the level of knowledge in the area and stimulate further investigations (Park and Martin, 2017), offering opportunities for additional academic research.\n",
"role": "user"
},
{
"content": "Purpose: This study aims to analyze the managerial levers previously considered in literature in the setting of the provision of primary care and community services (in particular for patients with long-term conditions being treated also at home) as well as those scarcely explored that could potentially be adopted in the future.\nDesign/methodology/approach: This study was a structured literature review. The authors retrieved papers, published from 2005-2020, from electronic databases (i.e. ABI/INFORM Complete, Jstor, PubMed and Scopus). Each selected paper was assigned to a framework category, and a thematic analysis was performed.\nFindings: Topics scarcely explored in literature were related to logistics/supply chain, economic evaluations, performance management and customer satisfaction. Some papers embraced more than one management topic, confirming the multidisciplinary nature of territorial healthcare services. The majority of research, however, focused on only one aspect of primary care services, and a lack of an integrated view regarding the provision of those services emerged.\nOriginality/value: This study represents a first attempt to rationalize the fragmented body of knowledge on the topic of the provision of primary and community care services. This study enabled some light to be shed on the managerial levers already explored previously in literature and also identifies a number of trajectories for future research.\n",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.