title
stringlengths 6
301
| keywords
listlengths 0
23
| messages
listlengths 2
2
|
|---|---|---|
Development elasticity of religious assets: a Turkish experience
|
[
"Economic development",
"Religion",
"Industrialization"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: This paper investigates development elasticity of religious assets for a cross-section of all major Turkish cities (provinces, Il in Turkish) and towns (Ilce in Turkish), and small towns (Belde in Turkish) with municipal authorities. Villages are excluded from the sample since data are not available for them. Investigating development elasticities of religious assets at the aggregate city level is not a trivial endeavor since scholars from different disciplines are trying to understand the relationship between economic development and various aspects of religion, especially after the September 11 attacks. Religious assets in this paper are measured or operationalized by the percentage of religious buildings in a particular city in total number of all buildings in that city. Religious buildings include all buildings in use belonging to different religions such as mosques, masjid, tomb or turbehs, etc. for Muslims, churches for Christians, synagogues for Jews, etc. A masjit is a Muslim place of worship for which an imam is not necessarily appointed by the central government. One more difference between mosques and masjids is the fact that mosques are the property of religious affair department of central government or foundations even though land of the mosque is donated and mosque is built by individuals or private institutions, and foundations, whereas masjids can also be owned by private individuals. Mosques, masjits, and turbehs are usually built on a land that is usually donated by individuals or religious foundations. Local governments (municipalities) also designate a piece of land for mosque, masjit, and turbeh buildings. And land in the city centers is more expensive. Mosques, Masjits, and turbehs are usually built by private donations, religious foundations (Vakif in Turkish), and/or local governments. There was a fierce political debate when a local government with religiously conservative mayor in Istanbul, decided to built a mosque in a public park in Goztepe in 2005, a district in Istanbul. The opposition party took the decision to the administrative court. And there was a huge public debate on TV, popular newspapers, etc. over this issue (Simsek et al., 2006). There are special mosques that were built by the central government for multi-purposes (such as tourism) like Kocatape Mosque in Ankara. However, mosques that were built by the central government are not in high number.\n2. Literature review: This paper investigates the elasticity of religious assets with respect to economic development since there is a popular understanding in Turkey, and in many other circles in different countries in this matter, that economic development reduces the need for religious services and needs. Therefore, the most related literature to this current paper is the literature about development and its effects on various aspects of religion.\n3. Data and descriptive statistics: The domain of the empirical study is the cross-section of the Turkish cities. Provinces (il), towns (ilce), and small towns (belde) are used in the study. There are 81 provinces, 850 towns, and 2,267 small towns in Turkey. Villages are excluded from the study due to non-availability of the data. The building inventory is prepared/recorded/calculated by TSI for the year 2000. In other words, this building inventory data shows the year of 2000 building stocks.\n4. The model and results of regressions: The first model to estimate:\n5. Robustness check: In this section, for robustness check, I use a different data for religiousness/religiosity of the cities, namely percentage share of the leading religiously conservative party in the 2002 general parliamentary election. Justice and Development Party (AKP in Turkish abbreviation) was the leading political party gaining the 34.43 percent of the votes. AKP, defines itself as religiously conservative party with religious part more emphasized, was a new party. However, it had two important preceding political parties, Welfare Party (Refah Partisi) and Virtue Party (Fazilet Partisi), both of which were closed by the CC of Turkey for being accused of radical/fundamentalist parties which, as judged by the CC, were trying to change the regime from republic to Islamic rule. It is argued that votes of parties do not exactly show the religiosity of the cities since there are different mergers, concerns, and negotiations during the election process. However, AKP is a continuation of Welfare Party closed by the CC in 1998 and Virtue Party which was established after Welfare Party, which was also closed by the CC in 2001. After these closings, AKP was running for the elections in 2002 and became the leading party out of the elections. After Welfare Party and Virtue Party were closed, AKP was trying to attract also moderate Muslim along with some radical/fundamentalist religious voters as well. There was one more party, Felicity Party (Saadet Partisi), also established after the closing of Welfare Party and Virtue Party, which defines itself ultra fundamentalist.\n6. Conclusion: This paper investigates empirically the relationship between the religious assets (mosques, masjits, and turbehs, etc.) and economic development for a cross-section of Turkish cities. It is observed that there is a non-linear relationship between the religious assets and industrialization. Initially as industrialization increases, the religious assets also increase. As industrialization increases more, the religious assets are decreasing, confirming the hypothesis of modernization and secularization.\n",
"role": "user"
},
{
"content": "Purpose: The aim of this paper is to investigate the development elasticities of religious assets namely, the relationship between the stock of religious assets (religious buildings like mosques, mescits, turbehs, etc.) and economic development.\nDesign/methodology/approach: Development elasticities of religious assets with respect to development is empirically investigated for a cross-section of all Turkish cities with municipal authorities. Two different regression analyses use religious building in total number of building and percentage share of religiously conservatives parties vote as the dependent variable. Independent variables include various development measures.\nFindings: It is found that the stock of religious assets is negatively related to the development. Also, it is found that economic development and the stock of religious assets have a non-linear relationship. Religious assets increase with industrialization first, however, as the industrialization increases more, they decrease. Coastal towns have smaller religious assets. Mosques/masjits and schools are complements rather than substitutes as they statistically significantly affect each other positively. For robustness check, a different measure of religiosity is used in empirical research. Vote shares of religiously conservative parties are smaller in developed regions of the country. In other words, religiously conservative parties attracted the poor or less developed parts of the country.\nPractical implications: Economic development and religious assets are related to each other in a developing country case study. Increased economic development can cause religious assets to decline in this particular case study.\nOriginality/value: This paper uses a novel data set to study the relationship between development and religion in a developing country.\n",
"role": "assistant"
}
] |
Investigating agent-based inter-organizational systems and business network performance: Lessons learned from the logistics sector
|
[
"Coordination",
"Agent-based technology",
"Business networks performance",
"Cross-case analysis",
"Inter-organizational systems",
"Logistics information systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Nowadays, a business can no longer survive working as a self-contained organization that collaborates only with a handful of partners (Ghoshal and Bartlett, 1990; van Heck and Vervest, 2007; Tatarynowicz et al., 2015). Participating in a number of business networks can offer many opportunities, such as extended market reach, potential partnerships, and so on. In this networked era, inter-organizational systems (IOS) play an important role as technological vehicles that foster coordination activities (Kumar and Van Dissel, 1996; Bala and Venkatesh, 2007; Zhao and Xia, 2014).\n2. Theoretical conceptual model: In answering the main research question, this study synthesizes theoretical concepts from the literature on the smart business network (van Heck and Vervest, 2007; Busquets, 2010; Pau, 2013), coordination theory (Malone, 1987; Williamson, 2002; Olson et al., 2012), IOS (Kumar and Van Dissel, 1996; Bala and Venkatesh, 2007; Zhao and Xia, 2014), multi-agent systems (Wooldridge, 2009; Lesser and Corkill, 2014; Zambonelli et al., 2015), and coordination performance concepts (Marschak and Radner, 1972; Dove, 1999; Carley and Gasser, 1999).\n3. Research method: Figure 2 portrays an overview of this study's research design. To answer the research question, the positivist cross-case analysis method is used due to its relevance and its suitability for studying contemporary phenomenon within real-life contexts in which the researchers have no control over the research objects behavior (Yin, 2009). Regarding the choice of empirical cases, this study analyzes ABIOS applications in the warehousing (the Kiva system case), freight forwarding (the Kuehne and Nagel Greece case), and intermodal transportation (the SeaRail case) sectors. The cases comply with these criteria: the ABIOS has already been implemented in real-life situations; the cases represent the perspectives of different actors that work together within a specific business network, in this case logistics; and there are sufficient data sources to analyze the cases.\n4. Case studies: 4.1 ABIOS in the warehousing sector: the Kiva system\n5. Analysis: Table V provides an overview of the case studies framed according to the predefined conceptual model, specifically P1a-P2d2 (see Section 2). The content analysis was applied to examine and assess the coherence between the empirical findings and the conceptual model's propositions. Analyzing the impact of ABIOS on stimulating the emergence of decentralized coordination structures (P1a), coordination structure alterations were significant in the Kiva system case. Decentralization of decision-making authority onto the subordinate working agents was evident. With the Kiva system, misalignments between the decision-making authority and the task execution responsibility are dismantled so each agent can operate autonomously.\n6. Conclusion: 6.1 Research findings\n",
"role": "user"
},
{
"content": "Purpose: From the theoretical perspectives of both multi-agent systems and smart business networks, empirical studies analyzing agent-based inter-organizational systems (ABIOS) in a real-life business setting are rare. The purpose of this paper is to investigate the impact of ABIOS on the performance of business networks.\nDesign/methodology/approach: This study presents a theoretical conceptual model portraying the influence of ABIOS on clients' coordination structure and information architecture; and the impact of those structural alterations on business network performance in terms of the coordination, agility, and informational performances. To validate the model, a cross-case analysis was conducted in three logistics cases, namely, warehousing, freight forwarding, and intermodal transportation.\nFindings: The application of ABIOS requires adjustments to the information architecture or the coordination structure, or both. Subsequently, those structural adjustments will stimulate improvements in the coordination, agility, and informational performances.\nResearch limitations/implications: The assessment of the clients' performance improvement is done at the company level not at an aggregate network level. Moreover, the study only covers cases from the logistics sector.\nPractical implications: This study explains the structural consequences of ABIOS applications. The adoption of an inter-organizational system is a strategic decision that requires support from multi-stakeholders. While the applications of ABIOS can offer performance improvement opportunities, adjustments must be made to the existing coordination structure or the information architecture, or both.\nOriginality/value: This study contributes to the smart business network literature and the ABIOS literature by presenting a validated conceptual model explaining the interplay among ABIOS, the coordination structure, informational structure, and business network performance, namely, the coordination, agility, and informational performances.\n",
"role": "assistant"
}
] |
Additive manufacturing in the mechanical engineering and medical industries spare parts supply chain
|
[
"Customer satisfaction",
"Inventory management",
"Additive manufacturing",
"Business performance",
"Supply chains"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: 1.1 Background\n2. Literature review and conceptual framework: The literature review is divided into three parts. In Section 2.1, literature about AM including definitions, overview and history, and key benefits it offers. In Section 2.2, AM adoption within the spare part supply chain will be explored and key benefits and challenges will be highlighted. In Section 2.3, the research conceptual framework will be introduced and the four main research hypotheses will be presented. Finally, Section 2.4 will summarise the existent research gap.\n3. Research methodology: 3.1 Research approach\n4. Analysis and results: As shown in Table III, data were collected from 69 respondents for this study. The respondents worked for firms operating in two industrial segments; industrial and commercial machinery and computer equipment (66.67 per cent); and measuring, analysing and controlling instruments, photographic, medical and optical instruments (33.33 per cent). The majority of respondents (82.61 per cent) worked for small organisations of 99 employees or below. The respondents were self-identified as, and categorised into, three groups: customers (21.74 per cent), suppliers (34.78 per cent), and customers and suppliers (43.48 per cent).\n5. Conclusions and recommendations: This chapter provides insights about key research conclusions related to potential impact of AM adoption on CSy to price and lead time and on overall CS. This part also provides views about the potential impact of AM adoption on SR and on overall inventory management. Furthermore, three main recommendations are developed based on the findings of this research. This is followed by a summary of the research theoretical and practical contribution. In addition, a number of research limitations and suggestions for future research are presented. Finally, a \"lessons learned\" section is added that could simplify future survey research in this field.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to approximate the impact that additive manufacturing (AM) will have on firm inventory performance (IP) and customer satisfaction (CS) when it is applied within the spare parts (SP) supply chain of manufacturing organisations. This research also explores the influence of customer sensitivity (CSy) to price and delivery lead time and supply risk (SR) within those approximations.\nDesign/methodology/approach: An online survey was used to collect the primary data for this research. Data were collected from 69 respondents working for organisations in two industrial segments within the UK manufacturing sector: \"Industrial and Commercial Machinery and Computer Equipment\" and \"Measuring, Analysing and Controlling Instruments, Photographic, Medical and Optical Instruments\". The respondents worked for entities that were categorised in three groups: customers, suppliers, and entities that were both customers and suppliers. The groups that were self-identified as \"customers\" or \"suppliers\" answered 20 survey items each and the group that was identified as both \"customers\" and \"suppliers\" answered 40 survey items.\nFindings: The results revealed that AM was considered a suitable vehicle for the fulfilment of SP demand. However, AM appeared to make no material difference to CS; the scenario used improved delivery time of SP but increased price. Also, AM was thought to improve IP through less reliance on buffer stock to manage SR and spikes in demand and less carrying of SP at risk of obsolescence.\nResearch limitations/implications: The respondents worked for entities within two manufacturing industry segments within the UK and the insights garnered may not be indicative of similar organisations competing in other manufacturing industry segments within the UK or in other countries. In addition, approximately 82 per cent of the surveyed respondents worked for small organisations with fewer than 100 employees and the results may differ for larger organisations. Further limitations were the relatively small sample size and lack of open-ended questions used in the survey. Larger sample size and the usage of open-ended survey questions may lead to more reliable and valuable responses and feedback.\nPractical implications: The findings from this research are considered to be of interest to practitioners contemplating adoption of AM and to developers of AM wishing to increase market share due to the positive reaction of entities within the industrial and commercial machinery and computer equipment, and measuring, analysing and controlling instrumentation industrial segments. This research raises awareness to the possible risks and rewards - from a range of perspectives, of AM to practitioners considering its adoption in the spare parts supply chain (SPSC).\nOriginality/value: The paper takes a novel perspective on AM in SPSCs by illuminating the supplier and buyer perspective based on empirical data. This research provides new insights about the appreciation of the use of AM in SPSCs of mostly small sized manufacturing companies located in the UK. This paper also gives new insights about the willingness/conditions of manufacturing companies in the UK to adopt AM for the provision of SP. The originality of this research is twofold: it broached the applicability of AM in the supply chains of the two targeted industrial segments, and as far as the authors are aware, the influence of CSy (e.g. to price or lead time) and SR on SPSC players' attitude to AM is yet to be considered. Finally, this research adopted a systems theory lens and considered system-wide impact of AM introduction.\n",
"role": "assistant"
}
] |
Monopoly behavior of China's manufacturing enterprises: Rent-seeking or innovation-oriented
|
[
"Rent-seeking",
"Monopoly power",
"Total factor productivity",
"Productivity paradox"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Research questions: Monopoly enterprises (monopolies) are the common concern of all sectors of society as their every single move may have a huge impact on the industry, the real economy, and even the entire society. With the control of existing resources, monopolies can choose between the behaviors of rent-seeking and innovation. In choosing rent-seeking, they can win a bigger portion of a \"cake\" in the zero-sum game, whereas choosing innovation gets them an entire new \"cake\" in the field. Although both rent-seeking (cutting the \"cake\") and innovation (producing the \"cake\") can achieve the profit-seeking purposes, their impacts on social welfare are diametrically opposite. The total welfare losses caused by rent-seeking to society or groups have been denounced by academic circles (Tullock, 1967; Krueger, 1974; Posner, 1975; Cowling and Mueller, 1978; Dechenaux et al., 2015)[1]. The driving role of innovation in economic growth is not only supported by neoclassical economic growth theory (Solow, 1956), but also further confirmed by the theory of endogenous economic growth in the fields of product category innovation (Judd, 1985; Grossman and Helpman, 1991) and product quality innovation (Grossman and Helpman, 1991; Aghion and Howitt, 1992). Certainly, rent-seeking and innovation may also occur in non-monopoly enterprises, but this is far less likely to happen than in monopoly enterprises[2]. In terms of rent-seeking, entities can be determined as rent-seeking monopolies directly according to the definition of rent-seeking, that is, the non-productive profit-seeking activity for securing monopolistic profits. In terms of innovation, according to Chen and Schwartz (2013), monopolistic firms are more motivated to innovate than non-monopoly firms, and based on the expanded quality innovation model of Acemoglu and Cao (2015), the higher the degree of monopoly, the more R&D investment in monopoly firms. In terms of financial constraints, rent-seeking non-productive inputs and innovative R&D investments require large amounts of capital, and the monopoly profits accumulated by monopoly enterprises in the early stage can directly provide financial support for rent-seeking and innovation. Therefore, the basic logic of this paper is as follows: the enterprise has an ex ante monopoly position, such that to maintain a continuous flow of monopoly profits, it will adopt innovative or rent-seeking methods, but followed by different \"additional effects\" (high TFP or low TFP). Here, the \"additional effect\" characterized by TFP (Total Factor Productivity) refers to the increase in output caused by factors other than traditional production factors like capital and labor, thus reflecting the production efficiency form factor input to product output.\n2. Research design: The data used in this paper are the fusion of the Chinese Industrial Enterprises Database and the China Economic Census Database, with a total sample size of 4,446,309. In the absence of ambiguity, this paper directly refers to the fusion data as the Chinese Industrial Enterprises Database. The processing software in this paper is STATA14.2, and all key data are operated with double precision. Furthermore, this paper also carried out two basic data collation procedures, namely, enterprise ID identification and price adjustment. First, enterprise ID identification adopts the cross-identification method, that is, the enterprises with the same name (in the same province) are identified as the same enterprise, and those with the same organization code are identified as the same enterprise. This method is more accurate[5] than the sequential identification method[6] of Brandt et al. (2012), because the latter identifies the same company with different organizational codes as multiple enterprises. Second, as for price adjustment, this paper uses the regional price index[7] to deflate the nominal price, and the base period is set at 2010.\n3. Analysis of the measurement results: In the quantitative regression, the interpreted variable is the TFP logarithm (lntfp), and the explaining variable contains the monopolistic power logarithm (lnmarkup) and the cross-terms (1.state#c.lnmarkup, 1.export#c.lnmarkup). Given that TFP may adversely affect the monopoly power and produce endogenous problems, this paper uses the instrumental variable (IV) method to solve it. The IVs used here are the monopolistic power log lag 1 (L1.lnmarkup), the state-owned equity vs monopoly power cross-term lag 1 (L1.1#c.lnmarkup), and the export vs monopoly power cross-term lag 1 (L1. 1.export#c.lnmarkup).\n4. Robustness testing and further analysis: 4.1 Robustness test\n5. Conclusions and implications: This paper studies the monopolistic behavior of Chinese manufacturing enterprises by measuring corporate TFP and corporate monopoly power. The study found that, on the whole, China's manufacturing monopolies are innovation-oriented monopolies and that rent-seeking behavior is not serious. However, there are certain differences between diversified types of monopolies. Differentiated by the shareholding structure, the state-owned equity monopolies are more inclined to innovate than their non-state-owned counterparts. Differentiated by product market, export monopolies are more inclined to seek rent than their domestic sales counterparts. Finally, differentiated by corporate scale, small monopolies are more inclined to seek rent than their large-scale counterparts. In addition, among large monopolies, state-owned equity export enterprises are most inclined to innovate, whereas among small monopolies, state-owned equity domestic sales enterprises are most inclined to rent-seeking. Thus, the following revelations are drawn.\n",
"role": "user"
},
{
"content": "Purpose: With the increase of state capital, corporate total factor productivity (TFP) has a tendency to jump up at first and then slowly decrease. Generally, no significant \"productivity paradox\" can be observed in China's manufacturing industry. With the increase of export density, corporate TFP also shows a trend of initial jump growth and subsequent slow decline. This paper aims to discuss these issues.\nDesign/methodology/approach: Using the 1996-2013 China Industrial Enterprise Database, this paper studies the monopolistic behavior of Chinese manufacturing enterprises through the measurement of TFP and corporate monopoly power.\nFindings: Results show that China's manufacturing monopoly enterprises are generally innovation-oriented rather than rent-seeking. However, there are certain differences between diversified types of monopoly enterprises: the ones with state capital are more inclined to innovate than those without, whereas the ones with export delivery value are more inclined to seek rent than those without.\nOriginality/value: Therefore, the government should implement differentiated policies for diversified types of monopoly enterprises, and do so in a targeted manner fully reflecting the containment of rent-seeking and the encouragement of innovation.\n",
"role": "assistant"
}
] |
The mediating role of job satisfaction in the relationship between job training satisfaction and turnover intentions
|
[
"Job satisfaction",
"Turnover intention",
"Training",
"Training evaluation",
"Job training satisfaction"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Turnover intention (TI) refers to the conscious and deliberate willingness to leave the organization and has been one of the major challenges organizations faced. In the literature, TI has been shown to be highly correlated with actual turnover (Han and Jekel, 2011; Shahnawaz and Jafri, 2009). Employees' actual turnover is deemed as a bad sign for the organization as it involves considerably visible and hidden cost, for example, it results in loss of tacit knowledge that the employees might have developed during their association with the organization (Droege and Hoobler, 2003), recruitment and training of new employees, decreased organizational efficiency (Alexander et al., 1994), etc. In 2008, a survey of 600 organizations indicated that 81 percent of them viewed employee retention as a critical business priority (Kreitner and Kinicki, 2010). An abundance of studies have been devoted to understanding and predicting employees' TI (Han and Jekel, 2011; Shahnawaz and Jafri, 2009) and the relationship between job satisfaction (JS) and turnover is negative (Kreitner and Kinicki, 2010).\nHypotheses: JS, which is defined as an affective or emotional response toward various facets of one's job, reflects the extent to which an individual likes his or her job (Kreitner and Kinicki, 2010). JS is one of key work attitudes. Attitudes toward the job should be related to behaviors on the job, and the most central of which is performance on the job (Judge et al., 2001; Robbins and Judge, 2013). Research has indicated that there exists a positive relationship between job attitudes and job performance (Judge et al., 2001; Ostroff, 1992; Ryan et al., 1996; Harter et al., 2002). Many researchers have further reported the negative relationship between job performance and TI (Allen and Griffeth, 1999; Hom and Griffeth, 1995; Zimmerman and Darnold, 2009) and the relationship between JS and turnover is found to be negative (Kreitner and Kinicki, 2010). Low-JS may be a symptom of an employee's intention to quit. Managers are able to reduce employee turnover by increasing employee JS, which was supported by a meta-analysis of 67 studies covering 24,556 people (Kreitner and Kinicki, 2010).\nMethodology: A survey, which refers to a process of collecting data by asking questions and recording people's answers (Whitley, 1996), was conducted to test the hypotheses. It took place in a classroom setting where the questionnaires were distributed to 150 employees who work for different companies and received training in the continuing education center of Feng Chia University in the middle region of the Taiwan. All trainees, part of the sample, enrolled in an 18-week course entitled \"foundation business course.\" Prior to any data collection, the questionnaire was analyzed to establish the validity and internal reliability. The method of direct administration to a group was used to acquire data from all members of the sample. Specifically, the researcher went to the classroom to deliver a verbal explanation about the study and also distributed the questionnaire to the participants. The participants, selected through convenience sampling, were asked to complete the questionnaire.\nResults: The questionnaires were randomly distributed and completed by 150 participants and 115 questionnaires were useable. The Cronbach's a which was found to be 0.88 for the JTS (11 items), 0.88 for the JS (27 items) and 0.94 for the TI scale (4 items).\nDiscussion: Theoretical and research implications\nConclusion: In this study, the relationship among JTS, JS and TI was examined. A significantly negative relationship between JTS and TI is found and mediated by JS. The JTS can be used to predict TI, which could serve as guidance when designing the training programs. The finding also suggests that JS is positively related to JTS but is negatively related to TI, which is in accordance with what have been reported in the earlier studies.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the relationship among job training satisfaction (JTS), job satisfaction (JS) and turnover intention (TI) and moreover, the role of JS in the relationship between the JTS and TI.\nDesign/methodology/approach: A survey was used in this study for a sample size of 115. The principal component analysis was carried out to examine the factors in JTS and JS. The regression and mediation analyses were applied to assess the relationship among the JTS, JS and TI.\nFindings: A significantly negative relationship between JTS and TI was found and mediated by JS. The paper also suggests that JS is positively related to JTS but is negatively related to TI, which is in accordance with what have been reported in the earlier studies.\nPractical implications: This study reinforced the significance of job training in human resource management practices. The employees' satisfaction with job training that would lead to higher JS and lower TIs was emphasized. Therefore, it becomes important for HRD professionals to consider both the delivery of the training to the employees and the employees' satisfaction with the training approach when designing the training program.\nOriginality/value: While many different variables have been studied in the relationship with TI, the JTS is rarely discussed. This paper attempts to examine the relationship among JTS, JS and TI with addition of new insights.\n",
"role": "assistant"
}
] |
Moderating effects of service separation on customer relationships with service firms: A social-exchange perspective
|
[
"Customer loyalty",
"Affective commitment",
"Service separation",
"Relational trust",
"Social benefits"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: During the last few decades, scholars have identified customer relationships to be one of the most important elements of marketing strategy (Ashley et al., 2011; Berry, 1983; Giovanis et al., 2015; Gummesson, 2002; Morgan and Hunt, 1994; Palmatier et al., 2006; Verma et al., 2016). The abundance of literature devoted to relationship marketing (RM) (e.g. Akerlund, 2005; Bowden et al., 2015; Dwyer et al., 1987; Halinen and Tahtinen, 2002; Kingshott and Pecotich, 2007; Morgan and Hunt, 1994) reveals the importance of nurturing customer relationships to help improve customer satisfaction and long-term profitability (Aurier and N'Goala, 2010; Gronroos, 1994; Sheth and Parvatiyar, 1995). However, the emergence and growing use of mobile and online technological platforms by service firms to engage with the customer have also changed the underlying dynamics in building service relationships with them (Ramaseshan et al., 2015), due to the accompanying service separation (i.e. Keh and Pang, 2010).\nConceptual background and hypotheses: Service separation and the customer relationship\nMethodology: Research setting and design\nData analysis and results: We used the well-established two-step process recommended by Anderson and Gerbing (1988) to estimate the measurement and structural models, using a multi-group structural equation modeling approach with AMOS 24. The initial measurement model for the two groups (separated and unseparated) showed a relatively poor fit (kh2=865.18, df=328, kh2/df=2.64, RMSEA=0.043, SRMR=0.039, NFI=0.92, CFI=0.95, TLI=0.94) based on the criteria suggested for model fit with sample sizes exceeding 350 (Hair et al., 2010). Hence, we used modification indices to correlate four pairs of error terms with each other and the reduced measurement model shows a better fit (kh2=610.84, df=310, kh2/df=1.97, RMSEA=0.042, SRMR=0.032, NFI=0.95, CFI=0.97, TLI=0.96). Table I shows all the scale items and their standardized parameter estimates (factor loadings) for the two groups.\nConceptual contribution: By revisiting the RM literature, we are now able to better understand how service separation associated with the proliferation of new mobile and online self-service potentially impacts the relational capabilities of the service organization. Amid the increased use of SSTs, and the inherently limited personal interactions they engender, this study advances both Keh and Pang's (2010) and Hartley and Green's (2017) call to further examine the effects of service separation. We test our model in the context of retail banking customers in Saudi Arabia.\nManagerial implications: Our findings show that bank managers need to be aware of the consequences that service separation emanating from new and emergent technologies potentially has upon their relational building efforts. This is critical in building customer loyalty considering that a recent banking report indicates that over 50 percent of customers globally are unsure whether or not they will stay with their banks in the next six months despite the improvement in overall positive experiences with their banks (World Retail Banking Report, 2013, p. 5). Our findings suggest that such discontent is not a function of a mismatch between the providers' relational efforts and the mode of service delivery but rather related to the overall service being experience. In other words, banks need to broaden their marketing strategies to incorporate customer-centric approach that caters to both emotional and social needs of customers.\nLimitations and future research: This study has a few limitations that future research may address. First, we use retail banking in Saudi Arabia as our research setting because its relatively stronger reliance on relationship building and social benefits helped us test the moderating role of service separation on customer relationships. However, this could potentially limit the generalizability of our findings because retail banking interactions tend to be largely need based, whereas many forms of service relationships require ongoing interactions, either due to the chronological needs of the relationship and/or some legally binding contract that compels the customer to constantly interact. For example, services, such as higher education, telecommunications, insurance and many professional services require protracted engagements in the relationship for either party (customer and firm) to fulfill each's obligations to one another. Therefore, future research could test our model with other services and cultural settings. Such studies would help find out if RM strategies can be universally applied to separated services, irrespective of the type of service and cultural environment. Second, in this research, we focused on the relational approach to building customer loyalty by studying the moderating impact of service separation on the mediating roles of trust and affective commitment. However, there are many other variables that could reflect the success of the firm's relational efforts, including purchase intentions, word-of-mouth, share-of-wallet, convenience, satisfaction and customer reactions to service failure. Hence, future research could extend our model using these variables as customer outcomes.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the impact of service separation on the ability of service firms to build and maintain customer relationships, by exploring the differences in the strength of interrelationships among key relational constructs between separated and unseparated service delivery modes.\nDesign/methodology/approach: A field survey was conducted with retail banking customers in Saudi Arabia (n=592) using a structured self-administered questionnaire consisting of well-established scales. Data were analyzed using AMOS 24.\nFindings: Service separation negatively moderates (weakens) the ability of the service firms to leverage their social benefits, relational trust and affective commitment to increase customer loyalty and to strengthen overall customer relationships.\nResearch limitations/implications: This study uses retail banking customers in Saudi Arabia to test the impact of service separation in their relationship with the bank; hence, its findings may not be generalizable to other types of services and cultural settings.\nPractical implications: Service firms using online and mobile technologies should be aware that trust and commitment remain key to building customer loyalty. Hence, the trade-off between the benefits of these technologies and their negative impact on customer relationship needs to be factored into managerial decision making.\nOriginality/value: The paper highlights the importance of maintaining face-to-face interactions with service customers to create robust relationships that yield loyalty, despite the growing popularity of online and mobile technologies.\n",
"role": "assistant"
}
] |
What is meant by adaptability in buildings?
|
[
"Change",
"Stakeholders",
"Flexibility",
"Communications",
"Briefing",
"Clients"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Adaptability has long been considered to be a desirable characteristic in the built environment. For instance, in the 1950s, Lynch (1958) discussed adaptability in the context of urban design and planning, and in the 1960s, Weeks (1965) made the case for more adaptable hospital buildings. In the decades since, adaptability has been promoted as a design strategy in a wide range of building types, including offices, housing and health care. Leaman et al. (1998) suggested that the word \"adaptability\" is \"[...] now commonplace in the vocabulary of briefing, building design and building management\". But what do practitioners in the building industry mean when they talk about adaptability? To what extent is there a shared understanding of adaptability across the industry? Such questions are pertinent, because if the literature is anything to go by, there is still considerable uncertainty concerning the meaning of adaptability in the context of the built environment.\n2. Background: The existing literature on adaptability in buildings is very diverse, encompassing a variety of disciplines, including architecture, engineering, facilities management (FM) and planning and a wide range of building types. This diversity has given rise to a broad range of perspectives on adaptability, but it also means that adaptability remains very much a misconstrued concept: the word means (or signifies) different things to different people. Friedman (2002, p. 1) suggested that \"Misconceptions about adaptability are the outcome of the term's many definitions and interpretations\", and this was certainly evident when reviewing the literature. \"Adaptability\" has become a \"buzzword\" (Carthey et al., 2011, p. 89) - frequently used, but much misunderstood.\n3. Research methods: The aim of this study was to explore what industry practitioners mean when they talk about adaptability in buildings. The authors adopted an inductive approach in which knowledge and meaning are socially constructed and situated in practice (Berger and Luckmann, 1966). Reflecting this approach, the authors undertook unstructured face-to-face qualitative interviews with practitioners from a range of professional disciplines in the construction industry (Table IV). Industry practitioners were defined as people who are engaged in the building development process and influence how buildings are designed and constructed (Olander, 2007). Gaining a multidisciplinary perspective on adaptability was important, because professional background and situated practices can influence how one views the world (Fischler, 1995).\n4. Discussion of findings: Coding of the interview transcripts revealed a range of perspectives, but three themes emerged that relate primarily to the \"what\", rather than the \"why\" or \"how\", of adaptability (Figure 1), namely:\n5. Conclusions: Despite suggestions in the literature that the word \"adaptability\" is commonly used in the construction industry, there is very little understanding about how practitioners' comprehend adaptability or the terminology that they use when talking about the ability of buildings to accommodate change. This study has therefore addressed a gap in the existing literature by foregrounding the voices of industry practitioners and exploring their (sometimes very different) interpretations of adaptability in buildings. Although many of the examples reflect the UK bias of the interviewees and that around half of them were architects, we believe that the evidence supports five conclusions that could have international ramifications. The first three relate specifically to the themes discussed above (terminology, meaning and briefing), whereas the other two conclusions are drawn from the findings of our research.\n",
"role": "user"
},
{
"content": "Purpose: Despite being a common term in the literature, there is little agreement about what the word \"adaptability\" means in the context of the built environment and very little evidence regarding practitioners' understanding of adaptability. This paper aims to examine what practitioners in the building industry mean when they talk about \"adaptability\".\nDesign/methodology/approach: This study adopted a qualitative approach, involving 82 unstructured face-to-face interviews with practitioners from a range of professional disciplines in the construction industry, including architects, engineers, facilities managers, property agents and planners. The interview transcripts were coded inductively to identify themes in the qualitative data.\nFindings: The interview data revealed a wide range of perspectives on adaptability, particularly regarding terminology, the meanings practitioners associate with adaptability and the way in which these meanings are communicated to others in the industry. The applied meaning of adaptability varied depending on context.\nPractical implications: Conflicting language, and different interpretations of adaptability, is a potential barrier to the development of adaptable buildings. A clearer articulation of the meaning of adaptability (particularly by clients) during briefing and design could give rise to a more appropriate level of adaptability in the built environment.\nOriginality/value: This study has addressed a gap in the existing literature by foregrounding the voices of industry practitioners and exploring their (sometimes very different) interpretations of adaptability in buildings.\n",
"role": "assistant"
}
] |
Development of public sector information management systems: challenges and promising practices
|
[
"Information systems",
"Public organizations",
"Challenges of information management system",
"Information management system",
"Key project management areas",
"Practices of information management system"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The chaos in development of information systems brings in tons of complexity if project planning is improper. Development of huge projects creates a lot of challenges for managers. The challenges would normally pledge allegiance with the factors that drive projects toward failure. Resultantly, projects are not completed according to agreed-upon deadlines, cost and budget constraints. The common scenarios can be observed by peeking a little into information systems development industry (Mandl-Striegnitz and Lichter, 1996). Pakistani software industry, in recent past, has witnessed a number of projects aiming at general population. These projects have come up from different organizations and government as e-government initiatives. Like other software projects, these projects also face various challenges during the development.\n2. Literature review: E-government is the use of information and communication technology (ICT) to deliver better services to citizens, businesses and government employees (Arfeen and Kamal, 2014). Advancements in technology, economics and society have forced governments to think ahead and plan to benefit from ICT. We are on the brink of era where governments do not have room for paper work, time for manual transactions and resources to pay for conventional long routines. Therefore, governments show keen interest toward the implementation of ICT procedures, but often services provided over past few years have not been up to the mark (Soufi and Maguire, 2007; van Velsen et al., 2009). These services are either not easily accessible due to lack of internet services in remote areas or are too complex to be used by common no-so-technical people (Furuholt and Saebo, 2017). There are many reasons of these deficiencies; lack of user involvement in designing of public service processes, lack of infrastructure and lack of proper guidance and training are some of the reasons (Folstad,H et al., 2004; Simonofski et al., 2018).\n3. Methodology: To study the contemporary challenges encountered during the development of public sector information management systems and to find the practices which are adopted to cope with such challenges, case study methodology has been opted as it is very suitable approach in such scenarios (Runeson and Host, 2009). Research shows that managers and practitioners have to face various types of challenges that are complex, difficult to comprehend and evolve over the time (Gregorya et al., 2016). We have also dealt with a complex situation involving people and sentiments; therefore, analytical research paradigm has been used which has proven sufficient to undertake such studies (Runeson and Host, 2009) and explain critical phenomenon.\n4. Findings from the study: Challenges, as described by the managers, belong to different categories and are presented in 18 different subthemes categorized under six key project areas as given in the PMBOK guide. Let us now discuss key project areas one by one along with responses from the managers.\n5. Conclusion: This case study helps in investigating the challenges that are confronted during the development of information systems for government sector. The practices that are used to deal with the challenges are also explored. The best practices have been presented that can help project managers in formulating a plan to ensure success while working with traditional government departments. Moreover, this study provides theoretical assistance to the current challenges and practices for the public IMS, supported by our most recent research and interview outcomes. Furthermore, this study has found how managers paved their way to success in difficult times and believe that their practices should be followed by other practitioners. The challenges confronted by managers in this project are not different from those challenges which can be experienced in many other similar projects. The way management has led this project to success is surely worth working for.\n",
"role": "user"
},
{
"content": "Purpose: More and more governmental organizations are switching to information systems to enhance their operations and reduce cost but the development of these systems involves a lot of challenges. This paper aims to find out the challenges that project managers have to face during the development of such systems and the practices they can adopt to address these challenges.\nDesign/methodology/approach: To identify the challenges, data have been gathered across six key project management areas. The six targeted key areas are project integration management, project scope management, project time management, project cost management, project human resource management and project communications management. The authors have coordinated with 11 managerial-level IT professionals using semi-structured interviews and have gone through their communication archives.\nFindings: Findings prove that practices such as focusing on cost reduction, informal congregations, trainings and frequent communication between vendor and client help in addressing the challenges. Furthermore, learning from management experiences of the managers can assist managers in similar role to create a pattern of success while working with governmental projects.\nOriginality/value: The authors peek into the development life cycle of a public sector project named as prison information management system. The project has encountered numerous challenges and has been accomplished in significantly delayed time than designated.\n",
"role": "assistant"
}
] |
Ambidextrous IT capabilities and business process performance: an empirical analysis
|
[
"Business process management",
"Ambidextrous BPM",
"Business process IT capabilities",
"Explorative BPM",
"Exploitative BPM",
"Business process management capabilities"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Since the 1970s, organization theorists (e.g. Duncan, 1976) have been attracted by the notion of organizational ambidexterity. Indeed, a very long time ago Abernathy (1978) suggested that a firm's ability to compete over time was rooted not only in its ability to increase its efficiency but also in its ability to be efficient and innovative at the same time. Afterwards, a consistent series of research across different disciplines in organizational adaptation appeared (e.g. March, 1991). The need for dual organizational capabilities has been raised amid a period of florid managerial activity and institutional pressures involving business process management (BPM) and control (e.g. Adler, 1993; Hackman and Wageman, 1995; Hammer and Stanton, 1999). In this regard, Davenport (1993) defines the business process as \"the specific ordering of work activities across time and space with a beginning, an end and clearly identified inputs and outputs.\"\n2. Literature backbone: 2.1 Ambidextrous IT capabilities\n3. Hypothesis development: Organizations are usually considered as a set of interlinked BP which meaningfully influence organizational strategy and performance (Benner and Tushman, 2003). BP are the vehicle through which business value is created from technological IT resources (Melville et al., 2004) such as functional systems, enterprise systems or BPM platforms. In times of highly dynamic business environments, IT flexibility is important to rapidly adapt BP to fluctuating customer demands and implement innovative technologies (Lu and Ramamurthy, 2011; Chen et al., 2014; Kumar and Stylianou, 2014). Based on Heckmann (2015), this activity is defined as explorative BPIT capabilities with the aim of identifying and implementing the usage of innovative and new IT resources to support a business process in order to achieve better performance. Firms can do this by integrating new IT resources into the IT infrastructure already in action (Joachim et al., 2013). Some mechanisms should be developed in order to allow the mixing and matching of task implementations to adapt quickly to changing requirements (Schilling, 2000), for example complementing the core ERP system with individual spreadsheet solutions (Alter, 2014). Higher levels of explorative BPIT capabilities indicate a focus on identifying and adopting new and innovative IT resources that have the potential to support BP in the future. Explorative BPIT capabilities can be seen as digital options, which are particularly crucial in dynamic environments (Sambamurthy et al., 2003). In order to adopt new IT resources continuously and to integrate them with the company's IT infrastructure, a modular architecture is required (Ling et al., 2009). This can be implemented by using service-oriented architectures (Joachim et al., 2013) or platform approaches (Tiwana et al., 2010). The constant ability to integrate the latest value-adding IT components provides the potential to innovate BP regularly (Gebauer and Schober, 2006). Thus, we propose the following:\n4. Data and methodology: 4.1 The context of analysis\n5. Results: We tested the hypotheses using OLS regression analysis, and we present the results in Table I. Model 1 represents the effect of the control variables only on BPP. Model 2, on the other hand, is implemented to test the impact of the two independent variables (explorative and exploitative BPIT capabilities). Finally, in Model 3 the interaction terms are proposed to test the moderator effect of BPM capabilities. Model 1 has an R2 of 0.17 (adjusted R2 is 0.12) and an F-value of 3.01 (p<0.05). Model 2 has an R2 of 0.56 (adjusted R2 is 0.48) and an F-value of 4.67 (p<0.001). Model 3 has an R2 of 0.66 (adjusted R2 is 0.59) and an F-value of 7.22 (p<0.001).\n6. Discussion: 6.1 Main implications\n",
"role": "user"
},
{
"content": "Purpose: In several studies, it has been found that organizational performance is affected by ambidextrous IT capabilities. Nevertheless, business processes are essential to the value generation conversion of IT investment into performance. In the literature, this focus on the impact of IT capabilities at the business process level is still under investigated. So, the purpose of this paper is to test the effects of explorative and exploitative business process IT capabilities on business process performances (BPP) and the positive moderator role of business process management (BPM) capabilities.\nDesign/methodology/approach: This analysis has been done through a quantitative study in the Italian hotel industry. An OLS regression analysis has been carried out on a sample of 404 firms.\nFindings: The study identifies distinct effects related to exploration and exploitation and finds a moderating effect of BPM capabilities, explaining their positive impact on BPP.\nOriginality/value: The main purpose of the paper is to contribute to the area of business process management by demonstrating the importance of both explorative and exploitative IT capabilities for a business process as well as the managerial capabilities at the process level. Furthermore, this focus at the process level allows us to add original insights into research on ambidexterity by expanding existing works.\n",
"role": "assistant"
}
] |
Mediation analysis in partial least squares path modeling: Helping researchers discuss more sophisticated models
|
[
"Structural equation modelling",
"Partial least squares",
"Bootstrapping",
"Indirect effects",
"Mediation analysis",
"Multiple mediation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Partial least squares (PLS) is a variance-based structural equation modeling (SEM) technique that has become very popular in management and social sciences in recent years. Current discussions about PLS emphasis its capability to model both composites and factors (Henseler et al., 2016) and its prediction orientation (Shmueli et al., 2016). In addition to these reasons, PLS is a useful tool for testing hypotheses especially in complex path models in an explorative manner (Chin, 2010; Wold, 1980). Nevertheless, with complex path models, it is much easier to overlook the occurrence of effects that do not directly manifest their influence (cf. Hair et al., 2012; Nitzl, in press). In a naive manner, researchers focus only on direct relationships and ignore mediating effects completely. This focus can heavily bias the interpretation of the results when a variable has no direct effect because its effect is mediated by another variable. In the worst case, researchers assume that a variable is not relevant for answering their research question at all.\n2. The mediating effect and Baron and Kenny's procedure and beyond: The core characteristic of a mediating effect (i.e. indirect effect or mediation) is that it involves a third variable that plays an intermediate role in the relationship between the independent and dependent variables. Technically speaking, the effect of the independent variable X on the dependent variable Y is mediated by a third variable, M, called the mediating variable or mediator (see Figure 1). Figure 1(a) shows the total effect c of the causal relationship between variables X and Y, and Figure 1(b) shows a mediated effect in which X exerts an indirect effect axb through M on Y. Thus, when we formulate mediation hypotheses, we focus on how an independent variable (X) affects a dependent variable (Y) by an intervening variable (M) (Baron and Kenny, 1986). The researcher's aim in mediation analysis is chiefly explanation because the main subject of mediation is to understand the development of processes (Henseler et al., 2016; Iacobucci et al., 2007). However, mediation analysis could also play an important role in prediction (Shmueli et al., 2016).\n3. Advanced procedure for mediation analysis in PLS: As shown, PLS researchers have to start by testing the indirect effect axb when analyzing mediating effects. The indirect effect can also be formulated as the difference between the total and direct effect:(1)\n4. Additional aspects for assessing mediation models fit and strength in PLS: Before the background that mediation analysis mainly deals with explanation, a discussion of the use of goodness-of-fit indices is appropriate. The goodness-of-fit of a model is the ability of a PLS path model to reproduce the data. Iacobucci et al. (2007) emphasize that a good fit is required before interpreting mediation analysis in a structural model in the context of CB-SEM, which is in line with the general suggestion of Henseler et al. (2016) for PLS that it should also become customary for PLS to determine the model fit. According to Wood et al. (2008b), many authors inferred full mediation when a model excluding direct effect c' exhibited a better fit than a model including both direct c' and the indirect effects axb.\n5. Handling multiple mediations: PLS is regularly characterized by complex path models (Hair et al., 2012; Nitzl, in press). There may be multiple relationships between one or more independent variables, one or more mediator variables, or one or more dependent variables (for general SEM examples, see Wood et al., 2008a). For instance, a complementary mediation variable (M1) may mitigate the independent variable (X) to a dependent variable (Y), and at the same time, a competitive mediation variable (M2) may also exist. From a naive perspective, someone can assume that the independent variable is not relevant because there is no relevant total effect c. However, when one of the mediator variables has a strong influence in a certain situation, the independent variable also wins in terms of relevance. Such areas can become very challenging, for example, when using a PLS path model to analyze which process improves or hinders the influence of the external pressure to work on performance. However, when more than one mediating effect is present, the abovementioned differentiation between direct and indirect effects for detecting mediation relationships remains applicable, and the above recommendations remain unchanged (Hayes, 2009).\n6. Conclusion: PLS applications must routinely account for mediating effects and apply state-of-the-art procedures. For this reason, we propose an alternative procedure for mediation analysis in PLS. Several articles using PLS applied at least some form of mediation. Although a few PLS studies already used a modern approach to test mediation, no study has yet presented a systematic overview and guideline of how to perform and classify a mediation analysis in a PLS path model context. PLS researchers are keenly interested in testing mediational hypotheses. However, they have used (if any) Baron and Kenny's (1986) approach, which is often criticized. Therefore, we have systematically transferred the recent findings from different research areas on mediation analysis to PLS. We summarize the findings more comprehensively and also evaluate their applicability for PLS, resulting in adjusted recommendations for mediation analysis in the context of PLS.\n",
"role": "user"
},
{
"content": "Purpose: Indirect or mediated effects constitute a type of relationship between constructs that often occurs in partial least squares (PLS) path modeling. Over the past few years, the methods for testing mediation have become more sophisticated. However, many researchers continue to use outdated methods to test mediating effects in PLS, which can lead to erroneous results. One reason for the use of outdated methods or even the lack of their use altogether is that no systematic tutorials on PLS exist that draw on the newest statistical findings. The paper aims to discuss these issues.\nDesign/methodology/approach: This study illustrates the state-of-the-art use of mediation analysis in the context of PLS-structural equation modeling (SEM).\nFindings: This study facilitates the adoption of modern procedures in PLS-SEM by challenging the conventional approach to mediation analysis and providing more accurate alternatives. In addition, the authors propose a decision tree and classification of mediation effects.\nOriginality/value: The recommended approach offers a wide range of testing options (e.g. multiple mediators) that go beyond simple mediation analysis alternatives, helping researchers discuss their studies in a more accurate way.\n",
"role": "assistant"
}
] |
Human resource management in merger and acquisition planning
|
[
"Communication",
"Due diligence",
"Planning",
"Integration",
"Human resource",
"Merger/acquisition"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Traditionally, leaders have thought that the success of mergers and acquisitions (M&As) came in the aftermath of the signing of the agreement. However, the literature offers examples of failures that may be due to the implications of human resources (HR) in the early stages of the operation (Bagdadli et al., 2014).\nTheoretical background: M&A processes involve a modification of the organizational structure of the company, creating a need to reorganize HR and modify traditional work procedures. In practice, however, little attention has been paid to the role of HR in this process, even though HR are one of the main actors in implementing the changes envisaged (Marks and Mirvis, 2000).\nMethodology: In this research, the case study methodology has been used for three reasons: the investigated phenomenon is exploratory and descriptive; the literature is not conclusive, and it is necessary to perform an in-depth and qualitative analysis of the subject; and primary information is available owing to the collaboration of the company studied, which has allowed interviews with people involved in the process and a stay of one of the researchers within the organization. The steps followed in the case study are shown in the following list.\nCase study findings: The parent organization is immersed in a development process with the aim of becoming world leaders in the industry by the year 2020. The reasons that justify the M&A are mainly strategic: having the size for competing in the global market, acquiring the advantages of vertical integration, entering or consolidating its position in a country, specifically in Spain, and reducing competition and influencing future developments in the industry.\nDiscussion and conclusions: This study, regarding how to effectively manage HR, focuses on the planning stage of an M&A operation to achieve success. Therefore, the literature is reviewed to identify the most relevant factors in HRM in the M&A processes, which are contrasted in a case study. According to Guest (1997), \"it is essential to measure HRM outcomes to understand how HRM impacts on performance\". In that sense, his model identifies six steps: HRM strategy; HRM practices; HRM outcomes (commitment, quality and flexibility); behaviour outcomes (motivation, cooperation, involvement and organizational citizenship); performance outcomes (high productivity, quality and innovation; low absence, labour turnover, conflict, customer complaints and labour turnover); and financial outcomes (profits and ROI).\n",
"role": "user"
},
{
"content": "Purpose: Many of the papers that analyse human resource management (HRM) in merger and acquisition (M&A) processes focus on the last two stages (integration and implementation). Then, the purpose of this paper is to propose an HRM model for the first stage of the process (planning) with four key factors: the due diligence process, the integration plan, the communication plan and the learning plan.\nDesign/methodology/approach: From the theoretical model, the authors analyse a case study of a multinational company that operates in the mechanical engineering sector.\nFindings: The results show the special importance of HRM in the success of the M&A process. And, the main actions implemented in human resources (HRs) contributing to the success of this process are identified.\nResearch limitations/implications: Subsequent investigations could conduct similar analyses for the rest of the stages of the M&A process (integration and implementation), with the objective of presenting a complete HRM model in M&A processes.\nPractical implications: The case study allows researchers to learn from professionals and business leaders while also offering a theoretical model that can help managers make decisions and improve the management of these processes.\nOriginality/value: The main contribution of this study has been to observe how HRs are managed in the planning stage of M&As.\n",
"role": "assistant"
}
] |
Relationship between perceived justice and identification: The mediating role of organizational images
|
[
"Perceived justice",
"Organizational identification",
"Organizational images",
"Social identity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Influential studies on organizational identification have drawn inspiration from social identity theory (Ashforth, 2016; Ashforth and Mael, 1989; Au and Marks, 2012; Bergami and Bagozzi, 2000; Joensson, 2008), which provides multiple ways of explaining how the identification process occurs. In particular, identity scholars have developed two main lines of research over the last two decades. On the one hand, we have the cognitive tradition, which adopts a cognitive perspective and primarily considers intergroup factors (e.g. Dutton et al., 1994; Hogg et al., 2004), while on the other hand, we have the group-value model (Lind and Tyler, 1988) and the subsequent group engagement model (GEM) (Tyler and Blader, 2003), which adopt a relational perspective that mainly considers intragroup factors (Blader and Tyler, 2015). The former perspective is directly focused on explaining what identification is, why it is important, how the process occurs and how different foci of identification are integrated (Ashforth et al., 2008), while the latter perspective considers organizational identification to be a specific mediator between justice judgments and work-related outcomes. Although both traditions have enlarged the scope of their focus over time, including other constructs (e.g. GEM has been extended to include economic outcomes and extrarole behaviors, Blader and Tyler, 2009), these two research traditions have never fully been combined, even though they study the same core construct and share theoretical foundations.\nSocial identity theory and organizational identification: Social identity is defined as \"that part of an individual's self-concept which derives from his knowledge of his membership in a social group (or groups) together with the value and emotional significance attached to that membership\" (Tajfel, 1978, p. 68). It derives from intergroup social comparisons that seek to confirm in-group-favoring evaluative distinctiveness (Hogg et al., 2004). As a result, a person perceives him- or herself not only in terms of personal traits that differentiate him or her from other individuals but also in terms of the characteristics that he or she shares with other members. This self-categorization process allows and supports individuals to develop over time affective and evaluative bonds with the group and its members (Bergami and Bagozzi, 2000; Haslam, 2004). The first significant attempt to connect the social identity theory to work organizations is ascribed to Ashforth and Mael (1989), whose contribution generated a cognitive tradition that we present below.\nMethod: Setting and procedures\nResults: Measurement assessment\nDiscussion: This paper aimed to bridge two perspectives on organizational identification that we labeled cognitive (tested in Model 1) and relational (Model 2). When both models were analyzed separately, all goodness-of-fit measures were satisfactory, and all coefficients exerted significant effects on organizational identification. These results provide support for the validity of both models and their theoretical assumptions. Comparing the two models, we might note that a slightly greater variance is explained by the cognitive model, wherein positive perceived identity shows the strongest relationship with identification. All coefficients are also significant in the relational model. Among the three antecedents, interactional justice exerts the strongest influence on organizational identification, followed by procedural justice; distributive justice is the weakest antecedent, confirming the theoretical assumptions of the GEM, wherein the identity hypothesis prevails over the distributive resource hypothesis (Tyler and Blader, 2003).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to contribute to our understanding of the antecedents of organizational identification. Specifically, this paper aims to integrate two perspectives developed within the social identity domain, labeled \"cognitive\" and \"relational,\" by comparing and reconciling their relationship organizational identification.\nDesign/methodology/approach: This study uses a survey method and a structured questionnaire to collect data from people working in a call center. The hypotheses were tested on a sample of 743 employees by using structural equation models and Hayes' (2017) bootstrapping procedure.\nFindings: The results provide evidence for a mediational model in which the attractiveness of organizational images (cognitive representations) mediates the relationship between perceived justice (relational judgments) and organizational identification.\nResearch limitations/implications: The data were obtained from a single source in a cross-sectional design, which may inflate common method variance. To address threats to validity, the authors employed several procedures, the results of which revealed that no parameters corresponding to the hypotheses changed in sign or significance, thus suggesting that the presence of method bias, if any, was nonconsequential.\nPractical implications: Not only does perceived justice relate to the sense of belonging to an organization, but it also contributes to shaping the long-term cognitive representations of the company. In particular, both HR and line managers should be aware that in this respect, the interactional dimension of justice shows the strongest effect.\nOriginality/value: Building on and enlarging the scope of the extant literature, the findings contribute to our knowledge of how relational judgments shape cognitive images about the company, influencing, in turn, the individual-organization relationship.\n",
"role": "assistant"
}
] |
Dynamics of obesity in Finland
|
[
"Causality",
"Cointegration",
"Finland",
"Obesity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Obesity rates in Finland have more than doubled in the past three decades. Health Statistics of OECD (2010) show that the percentage of self reported adult obese people rose substantially from 6.6 per cent in 1978 to 15.9 per cent in 2009. In regards to comparable obesity data provided by the Health Statistics of OECD (2010), Finland ranked 17th out of 30 OECD countries, with 14.3 per cent in 2005. In the same period, Japan had the lowest estimate of obesity (3.9 per cent) whereas the highest rate was recorded in the USA (34.3 per cent). Koskinen et al. (2006) states that in the early 2000s, one in five adults in Finland was obese. Lahti-Koski (2001) estimates the trends and determinants of adult obesity in Finland using the survey data running from 1972 to 1997. The study of Lahti-Koski associates obesity with several socio-economic factors such as gender, ethnicity, marital status, dietary intake, physical activity, education, alcohol consumption and smoking. The results conclude that the mean BMI increased for both genders during the estimation period.\n2 A brief literature review: Obesity is a complex phenomenon but economics is at the heart of the obesity epidemic. The literature on obesity is large and covers many disciplines, but there is little agreement about causes. The fundamental reason for being overweight is basically related with overeating, in other words, calorie intake exceeds calorie use. The literature identifies several socio-economic factors for a positive balance of calories. Finkelstein and Strombotne (2010) points out that economic forces have made it easier and cheaper to consume high-energy, tasty, affordable foods and have allowed us to be increasingly sedentary at work, at home, and in between. Ruhm (2010) indicates that not only economic forces but biological factors are also to blame for obesity. Moreover, medical advances have lowered the health costs that results from excess weight and have decreased the motivation to diet and exercise. As discussed in Popkin (1999), rising urbanization is associated with increased opportunities for eating and reduced opportunities for physical activity. Moreover, people in rural areas typically have higher levels of physical activity due to the focus on agricultural work. Phillipson (2001) identifies technological change as another responsible factor for the obesity epidemic, largely because of its effect at reducing energy expenditure in the workplace. Sedentary technological change has lowered both the real price of food and physical expenditures of calories per hour worked in both market and household production. As a result, the cost of consuming calories decreased and the cost of expending calories increased, thus contributing to rise in obesity in two ways. Unemployment is another contributor to the obesity epidemic as, for people out of work, the energy expenditure or the amount of physical activity is substantially reduced leading to an imbalance of calories. The increase in obesity has given rise to demands for public intervention, mainly in the form of education and health programs, to reduce obesity through better diet and more exercise. It is not plausible to assume that individuals would make rational choices in the free market economy against the costs of obesity since the market forces promoting food consumption are extremely powerful to influence an individual's decision adversely. Consequently, rising public health expenditures decreases significantly the risk of obesity and this reduction brings about more economic benefits in terms of higher level labour force participation and an increase in income taxes, a decrease in absenteeism, etc. in the long-run.\n3 Model and econometric methodology: Following the empirical literature on obesity, this study adopts the following long-run relationship between obesity, public health expenditures, unemployment, urbanization, alcohol consumption and calorie intake in double linear logarithmic form as: Equation 1 where o\n4 Results: Annual data over the period 1978-2007 were used to estimate equations (2) and (3) by the ARDL cointegration procedure of Pesaran et al. (2001). Variable definition and sources of data are as follows. Data are collected from four different sources, namely: food balance sheets of the Food and Agriculture Organization (FAO) of the United Nations, Health Statistics (HS) of the OECD, main economic indicators (MEI) of the OECD and World Development Indicators (WDI) of the World Bank. All variables are expressed in natural logarithm.\n5 Conclusions: This study has attempted to identify the causes of increasing obesity prevalence amongst the Finnish adults using time series cointegration procedure. The results illustrate that there exists a long-run association between obesity, public health expenditures, unemployment, urbanization, alcohol consumption and calorie intake. The econometric results reveal that the main factor of obesity in Finland is urbanization, as it provides an increasing availability of fast- and convenience foods with high-energy density, and sedentary jobs with little incentive to exercise. Alcohol consumption seems to be another major contributor to obesity, whereas the contribution of unemployment to obesity is relatively minimal. The results also demonstrate that public health expenditures appear to have a substantial impact to alleviate the obesity epidemic. As for the Granger causality tests, there exists a long-run causality running from all of the explanatory factors to obesity. In the short-run, there is a unilateral causality from alcohol consumption to obesity which requires public policy intervention to combat it.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this research is to study empirically the dynamics of obesity in Finland and provide empirical evidence of temporal causality between obesity, health expenditure, unemployment, urbanization, alcohol consumption and calorie intake.\nDesign/methodology/approach: The paper employs bounds testing cointegration procedure and augmented causality tests.\nFindings: The econometric results reveal that the leading cause of obesity is urbanization which is followed by the increasing availability of fast and convenience foods with high energy density, and sedentary jobs, with little incentive to exercise.\nPractical implications: The conclusion drawn from this study could be useful for policy makers and practitioners in international health organizations.\nOriginality/value: The paper provides first time time-series evidence on the dynamics of obesity in Finland, using the framework of cointegration and causality tests.\n",
"role": "assistant"
}
] |
Listed property trusts and downside systematic risk sensitivity
|
[
"Financial risk",
"Financial management",
"Property",
"Real estate",
"Australia"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Listed property trusts (LPTs) are some of the most successful indirect property investments in Australia. As at 31 December 2006, the total market capitalisation of Australian LPTs was approximately AUD$135 billion, representing over 12 per cent of the global listed real estate market (ASX, 2006a, b). Following the increasing acceptance of LPTs or REITs as an investment vehicle in other countries such as UK, factors affecting the level of risk, particularly systematic risk, in LPTs have received considerable research attention.Previous real estate studies have generally accepted that the characteristics of LPTs have direct implications for the level of systematic risk. Such studies have employed the capital asset pricing model (CAPM) to quantify the systematic risk of a LPT (or REIT). However, using CAPM has several drawbacks, such as the assumptions that return distributions are normally distributed and utility functions are irrelevant for investors. More importantly, these assumptions have been rejected by extensive studies. Therefore, it is not surprising that there is little empirical evidence supporting CAPM in finance and real estate literature (Chen et al., 1998; Conover et al., 2000; Fama and French, 1992, 2004; Malkiel and Xu, 1997).In order to address these limitations, alternative measures for capturing the market-related risk of a LPT must be employed. A viable alternative is to use downside systematic risk since it has been shown that downside systematic risk is a more intuitively appealing risk measure compared to conventional risk measures (Bawa and Linderberg, 1977; Hogan and Warren, 1974). Reasons for acceptance of downside systematic risk as an efficient measure of risk are that it:* does not require an assumption about the return distribution of an asset;* is more consistent with the investor's expected utility function; and* is a risk measure that only focuses on the downside and combines information provided by variance and skewness into one measure (Estrada, 2002; Nawrocki, 1999).Recent studies have also highlighted the ability of downside systematic risk to explain return variations and offer some empirical support for downside systematic risk (Ang et al., 2006; Post and Vilet, 2004).The questions surrounding downside systematic risk in LPTs (or REITs) have been the subject of relatively little research. Therefore, the aim of this paper is to address this gap in the literature by examining downside systematic risk in Australian LPTs. This study also examines variations in the sensitivities of systematic risk and downside systematic risk in relation to LPT's financial conditions as well as its management structure.The reminder of the paper is structured as follows. Section 2 reviews the connection between financial determinants, management structure and systematic risk. The importance of downside systematic risk is also reviewed in Section 2. The data and methodology are described in Section 3. Section 4 discusses the results. Section 5 presents the conclusions.\n2. Literature review: This section contains two parts, with the first part focusing on a review of the determinants of systematic risk in real estate literature. The second part reviews the importance and advantages of lower partial moment-CAPM (LPM-CAPM) in estimating market-related risk.2.1 Systematic risk and its determinants\n3. Data and methodology: 3.1 Data\n4. Results and discussion: 4.1 Comparison between beta and downside beta\n5. Conclusions: In recent years, the use of CAPM and beta have been widely criticised in the literature. In this study, LPM-CAPM and downside systematic risk have been employed in analysing Australian LPTs and three important results are found. First, consistent with the previous empirical evidence, it is confirmed that downside systematic risk and traditional systematic risk in LPTs are distinguishable. Second, although turnover and property asset reveal some explanatory power for betas in Australian LPTs, it is also evident that their explanatory powers have diminished in recent years. Thus, little evidence is found for the link between systematic risk and financial variables and management structure. More importantly, there is little evidence to suggest that these variables could explain betas in Australian LPTs in the same manner as in US REITs. Finally, a strong relationship between downside systematic risk and leverage, management structure and market capitalisation was identified. However, a declining trend in the explanatory power of size of LPTs in recent years was also observed.The important practical implications from this study are that more rigid interpretation by investors and real estate analysts is required with reference to the higher returns from highly leveraged and internally managed LPTs (or REITs). There is little evidence that these high returns appear to be a form of compensation for greater systematic risk, although these returns could be rewarding downside systematic risk in line with the different traits for both downside and traditional systematic risks. Considering the limitations surrounding the relatively small number of LPTs in Australia, future research should extend this study by employing data from different markets in order to ascertain the implications of the financial variables for downside systematic risk. The additional variables should also be considered for the purpose of increasing the explanation power of the models.\n",
"role": "user"
},
{
"content": "Purpose: - This paper aims to identify and examine the determinants of downside systematic risk in Australian listed property trusts (LPTs).\nDesign/methodology/approach: - Capital asset pricing model (CAPM) and lower partial moment-CAPM (LPM-CAPM) are employed to compute both systematic risk and downside systematic risk. The methodology of Patel and Olsen and Chaudhry et al. is adopted to examine the determinants of systematic risk and downside systematic risk.\nFindings: - The results confirm that systematic risk and downside systematic risk can be individually identified. There is little evidence to support the existence of linkages between systematic risk in Australian LPTs and financial/management structure determinants. On the other hand, downside systematic risk is directly related to the leverage/management structure of a LPT. The results are also robust after controlling for the LPTs' investment characteristics and varying target rates of return.\nPractical implications: - Investors and real estate analysts should conscious with the higher returns from high leverage and internally managed LPTs. Although there is no evidence that these higher returns are related to higher systematic risk, there could be the compensation for higher downside systematic risk.\nOriginality/value: - This study provides invaluable insights into the management of real estate risk in Australian LPTs with implications for REITs in other countries. Unlike previous studies of systematic risk in REITs or LPTs, this is the first study to assess downside systematic risk and explore the determinants of downside systematic risk in LPTs.\n",
"role": "assistant"
}
] |
Impacts of stratospheric aerosol geoengineering strategy on Caribbean coral reefs
|
[
"Caribbean Sea",
"Coral bleaching",
"Coral growth",
"Coral recruitment",
"Hurricanes",
"Stratospheric aerosol geoengineering"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Recent global warming has serious effects on coral reefs, as they are the most sensitive ecosystems on the planet to climate change, and results in widespread bleaching and mass mortality events (Baker et al., 2008). Although coral reefs make up only 0.2 per cent in area of the marine environment, they are among the most biodiverse ecosystems in the ocean, estimated to harbor around one-third of all described marine species (Crabbe, 2009). Coral reefs support the livelihoods of millions of people, especially those engaged in marine fisheries activities, and they also provide some important chemical compounds for many of the world's most prevalent and dangerous illnesses and diseases; e.g., the Caribbean Sea squirt can be used in the treatment of ovarian cancer (Miththapala, 2006). The growth and subsistence of corals depend on many environmental and climatic factors, including temperature, irradiance, hurricanes, calcium carbonate saturation, sedimentation and nutrients (Crabbe, 2009). These factors influence the key physiological processes of photosynthesis and calcification, as well as coral survival, and as a result, scleractinian coral reefs occur only in select areas of the world's oceans. Owing to recent global warming, serious degradation has been observed on many coral reefs worldwide, and the coral cover in the Caribbean has declined in some areas to ~10 per cent in the early 2000s (Isabelle et al., 2013; Schutte et al., 2010). Some reefs can self-recover, while others need help from artificial restoration, and some are unable to undergo restoration because the substrate or the environment is not suitable for coral growth.\n2. Study region and climate modeling: In this study, we will concentrate on Caribbean coral reefs, which are severely threatened by climate-induced ocean warming (Hoegh-Guldberg, 1999). There are about 26,000 km2 of coral reefs in the Caribbean region, approximately 7 per cent of the shallow reefs of the world (Burke et al., 2011). These reefs provide numerous benefits to nearby human communities, e.g. shoreline protection from the hurricanes. However, the Caribbean Sea is generally regarded as the reef region with the lowest resilience (Gardner et al., 2003). Caribbean coral reefs have experienced unprecedented changes in the past 40 years. The coral cover has fallen sharply, from about 50 per cent in the 1970s to 10 per cent in the first decade of the 2000s (Gardner et al., 2003). It has serious consequences for reef biodiversity, ecosystem functioning and related environmental services.\n3. Coral bleaching: The predominant source of nutrition for corals comes in the form of photosynthetic products produced by the zooxanthellae. Under unusually high sea temperatures, coral bleaching occurs when corals lose their zooxanthellae. Coral bleaching may result in coral mortality, catastrophic loss of coral cover and loss of critical habitat for associated reef fishes and other biota (Eakin et al., 2010).\n4. Hurricane impacts and coral recruitment: Hurricanes and tropical storms in the Caribbean Sea can cause considerable damage to coral reefs with great destruction of corals. Across the Caribbean Sea, coral cover is reduced by 17 per cent, on average, in the year following a hurricane impact. The frequency of hurricanes is higher in the north and east of the Caribbean Sea than in the south and west, so coral cover in the north and east of the Caribbean Sea impacted by hurricanes has declined at a significantly faster rate than that in the south and west. After hurricane impacts, corals show no evidence of recovery to a pre-storm state for at least eight years, which is roughly equivalent to the average return period of hurricanes in the most hurricane-prone parts of the Caribbean during 1951-2001 (Gardner et al., 2005). Moreover, there was a significant negative correlation (r = 0.72, p < 0.01) between recruitment estimates and storm severity. Intermediate storm severity resulted in variable levels of recruitment of non-branching corals, while the severest storms resulted in significantly (p < 0.002, Student's t-test) lower recruitment estimates (Crabbe, 2016).\n5. Coral growth rate: Reef-building scleractinian corals maintain a symbiotic relationship with photosynthetic zooxanthellae, which limit them to the photic zone. Therefore, the growth of coral reefs depends largely on the amount of light available for photosynthesis. Corals can grow from the surface to depths where there is between 1 and 10 per cent of the surface irradiance (Chalker et al., 1988).\n6. Discussions and conclusions: Coral reefs provide an environment in which one-third of all marine fish species and tens of thousands of other species are found, and from which 6 million tons of fish are caught annually. Present and future increases in sea temperature are likely to have severe effects on the world's coral reefs within 50 years. Stratospheric aerosol geoengineering has low costs and can rapidly mitigate climate change, so it is viewed as one of the most promising geoengineering approaches to be implemented in the future. In this study, we consider a stratospheric aerosol geoengineering in 2020-2069 with daily injections of SO2 at a rate of 5 Tg SO2 per year and concentrate its impacts on Caribbean coral reefs, which are currently threatened by climate-induced ocean warming. Stratospheric aerosol geoengineering could significantly mitigate coral bleaching. Under geoengineering implementation, coral bleaching in the Caribbean would not occur, except for some small regions near the southern coastline or east of Florida, while under the RCP4.5 scenario, coral bleaching will occur in most parts of the northern Caribbean Sea. At the same time, any changes in downward solar irradiation, sea level rise and the change of sea temperature variation in the Caribbean Sea caused by geoengineering implementation should have very little impacts on coral growth. For the impact related to severe Category 5 hurricanes, although geoengineering could prolong the return period of hurricanes during 2020-2069, if compared with the RCP4.5 scenario, it may not be enough for corals to recover after hurricane impacts. Therefore, stratospheric aerosol geoengineering with 5 Tg SO2 injection per year may not be enough to ensure full mitigation of climate change for corals in the Caribbean Sea. In addition, stratospheric aerosol geoengineering cannot avert the continued absorption of increasing anthropogenic CO2 emissions by the global ocean, which leads to rising acidity and to decreases in coral calcification and growth.\n",
"role": "user"
},
{
"content": "Purpose: Currently, negotiation on global carbon emissions reduction is very difficult owing to lack of international willingness. In response, geoengineering (climate engineering) strategies are proposed to artificially cool the planet. Meanwhile, as the harbor around one-third of all described marine species, coral reefs are the most sensitive ecosystem on the planet to climate change. However, until now, there is no quantitative assessment on the impacts of geoengineering on coral reefs. This study aims to model the impacts of stratospheric aerosol geoengineering on coral reefs.\nDesign/methodology/approach: The HadGEM2-ES climate model is used to model and evaluate the impacts of stratospheric aerosol geoengineering on coral reefs.\nFindings: This study shows that (1) stratospheric aerosol geoengineering could significantly mitigate future coral bleaching throughout the Caribbean Sea; (2) Changes in downward solar irradiation, sea level rise and sea surface temperature caused by geoengineering implementation should have very little impacts on coral reefs; (3) Although geoengineering would prolong the return period of future hurricanes, this may still be too short to ensure coral recruitment and survival after hurricane damage.\nOriginality/value: This is the first time internationally to quantitatively assess the impacts of geoengineering on coral reefs.\n",
"role": "assistant"
}
] |
Now for the science bit: implementing community-based learning in chemistry
|
[
"Students",
"Communities",
"Chemistry",
"Self development",
"Ireland"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper presents a case study on implementation of community-based learning (CBL) projects in the School of Chemical and Pharmaceutical Sciences in Dublin Institute of Technology (DIT). Two projects were piloted in 2007. One involved students preparing suitable posters and demonstrations and interacting with secondary school students to help develop their understanding of chemistry topics and, in the other, undergraduates performed chemical risk assessments and safety audits in schools or small businesses. Both projects have been continued and the provision has extended since and there are now three CBL projects and two community-based research (CBR) projects offered by the school.\nStudent learning from community engagement: background in DIT and terminology used: This paper focuses on CBL initiatives implemented to promote student learning from community engagement on undergraduate chemistry courses at Dublin Institute of Technology (DIT). Some CBR initiatives will also be discussed but they are not examined in the same detail. To ensure clarity, the interpretations of the terms employed to classify the learning and teaching activities reported in this paper are now provided. The original definition of \"community-based learning\" adopted in 2006, when these projects were developed, was taken from Bringle and Hatcher (1996, p. 122):[...] a credit-bearing educational experience in which students participate in an organised service activity that meets identified community needs and reflect on the service activity in such a way as to gain further understanding of the course content, a broader appreciation of the discipline and an enhanced sense of civic responsibility.Bringle and Hatcher were actually defining \"service-learning\" in this case and this is a term that originated in the USA. There has been some debate over the most appropriate words to use to categorise student-community engagement in Ireland. Boland and McIlrath (2007) discuss the localisation process involved in more detail and, as an interim measure, have proposed using \"pedagogies for civic engagement\" to label practices that involve student-community engagement. A similar deliberation occurred locally in DIT. In 2001, some academic staff in the Faculty of Tourism and Food were the first to implement this approach and they initially used the term \"service-learning\" (Kerins, 2010). By 2005, DIT was successful in obtaining funding for three years to extend these practices across all disciplines (Ahern, 2005) and, at this point, the initiative was named the Community Learning Programme (CLP). The successor to this project has been funded for another three years and was launched in September 2008. It was renamed the programme for Students Learning With Communities (SLWC). This change of title has been found by the authors to facilitate a more ready recognition by prospective community partners of what is involved. One of the aims of the SLWC programme remains the initiation of effective community engagement across DIT but development and consolidation of existing projects and national and international collaboration are also important objectives reflecting that the project has become more established and embedded (Cooke, 2009).Currently, the authors use the terms CBL (community-based learning) and CBR (community-based research) to describe the learning and teaching strategies used but we clearly identify to all partners - students, fellow staff and community partners- that these activities are part of the SLWC programme in DIT. It may seem that unnecessary detail is being provided in relation to labels and terms but it is often the case that misunderstandings arise due to confusion between the type of projects described here and volunteering activities or student work placements. Bringle and Hatcher (1996, p. 226) have observed that \"creating a common understanding of what constitutes service-learning at a particular institution will pay dividends later\". In this regard, an institutional level definition is now provided by SLWC which incorporates CBL and CBR[1]:Students Learning With Communities involves DIT staff and/or students working with community partners (local groups, not-for-profit organisations, etc.) to develop real-life projects. Learning comes alive for the students as they work on these projects with real clients, applying their specialist subject skills, and receiving course credits for their work. The community becomes part of the teaching process and benefits from the students' work. These projects give all participants the opportunity to engage in critical thinking and to develop their social awareness. The Programme for Students Learning With Communities ultimately aims to energise participants to work for social change.\nStudents learning with communities/community-based learning in chemistry: When the first two CBL projects were piloted in the School of Chemical and Pharmaceutical Sciences at DIT in the academic year 2006/2007, not much information on similar initiatives in chemistry in Europe was available. The authors were aware of a Chemistry Science Shop at the University of Groningen (Mulder, 2008) but the projects there essentially involved CBR, which was not our initial focus. Some service-learning projects were reported in the USA (Kesner and Eyring, 1999; Wiegand and Strait, 2000; Hatcher-Skeers and Aragon, 2002; Draper, 2004; Esson et al., 2005; Middlecamp et al., 2006) however.The expected advantages of implementing CBL projects were that students would see how their discipline knowledge is applied in practice and that staff and students would be involved in working with the community as a core element of their work and studies instead of having to find extra time for these activities (Esson et al., 2005). Hatcher-Skeers and Aragon (2002) observe that service-learning provides an excellent means of achieving active learning. Active learning and a focus on what the student, not the teacher, is doing are central elements of teaching that leads to good quality learning (Biggs, 1999).\nInitial implementation: As CBL was already in place in the DIT Faculty of Tourism and Food, this made it easier on an institutional level to introduce it elsewhere. The approach taken in developing pilot projects recommended by the CLP staff was to review all modules already in place to determine whether they could be modified so that one of the learning and teaching activities could be applied to the benefit of the community and also to find out if there were any existing CBL projects which had not had the term applied to them. In this case, we identified two suitable activities that required some modification. One difference between the implementation of these projects and those previously introduced in the Faculty of Tourism and Food was that, in our case, they were not elective and all students in a particular group were required to take the modules that included CBL activities. This decision was based on resource and operational considerations. However, the CBR projects introduced since then are options available to be selected by final year undergraduate project students. Another difference was that we introduced components of modules (either 40 per cent or 60 per cent) instead of entire modules involving CBL. This was for pragmatic reasons relating to the learning hours available and the existing curriculum. Table I summarises details of these module components, including associated learning outcomes and assessments.This method that we used initially provided a means to implement modules with a strong CBL emphasis quickly and thus was the preferred approach during the pilot phase. This approach did not strictly follow the recommended model of first formally asking potential community partners to identify their requirements at both an institutional and course level (Bringle and Hatcher, 1996). In the case of the Year 2 project described in Table I which has been labelled Activity A, it had already been established informally that the community partners (teachers and students in local secondary schools) had an existing need which could be met by suitably designed teaching and learning activities. For the Year 4 project labelled Activity B described in Table I however, the existence of a genuine community need has not been as clearly established.Additional details on Year 2 community-based learning project (Activity A)\nCritical assessment of the implementation of community-based learning activities: Several components have been analysed; assessment grades, student reflective accounts and evaluation questionnaires and a self-assessment, and each will be dealt with in turn. In addition, implications of the main outcomes for the adoption of CBL by other institutions are discussed, as are potential obstacles to implementing a CBL approach to chemistry. The CBR initiatives developed were not included in this evaluation. One reason is because the sample size is very small and another is that a suitable assessment mechanism that requires reflective analysis from these undergraduate research students has not yet been implemented. Therefore, a means by which a minor component of their current project assessment can be modified to allow this will be introduced as soon as possible.Student assessment grades for community-based learning activities\nSummary and conclusion: The purpose of this paper is to present a case study on the implementation since early 2007 of community-based learning (CBL) activities in a chemistry department in Ireland (the School of Chemical and Pharmaceutical Sciences at Dublin Institute of Technology). The terminology associated with student learning from community engagement at a local, national and international level and the importance of clear communication in that regard is addressed initially. A critical assessment of the implementation of CBL in an individual school in a third-level institution that did not have a well-established culture of applying this teaching and learning strategy at the time is then presented. The approach used from initial implementation is to identify suitable activities already in place that could be modified and to ensure that the learning outcomes and assessments are in alignment with the modified teaching and learning activities. The positive impact of the evolution of the institutional culture as a result of the activities of the CLP and its successor, the SLWC programme, is also evident from this case study (e.g. coordination of two cross disciplinary CBR projects, provision of training on student reflection and introduction to potential community partners).In relation to the impact on student learning, analysis of assessment grades, reflective statements and evaluation forms allow the identification of several themes which include the value of the active learning and application of knowledge in authentic situations that takes place and the development and appreciation of key skills (communication, information literacy, teamwork) that occurs. In addition, the importance of providing an \"induction\" to CBL for students and an awareness of potential difficulties that can arise from time management and group conflict issues is identified. The role of personal development in student-community engagement is also evaluated and the critical contribution of effective reflection is recognised. One significant omission up until this year was providing an explicit prompt to students to consider civic engagement when reflecting.The self-assessment (Shumer, 2000) undertaken identifies current strengths (\"We consider SLWC activities important in improving teaching and learning\", \"The purposes of our SLWC programme are clear to everyone involved\" and \"The purpose of our initiative is clearly linked to meaningful activities and learning objectives\"). In addition, priority areas for action have been established. These are the level of student reflection that occurs and confirmation that genuine community needs are being met.Implications for the adoption of CBL by other institutions are then drawn together and potential difficulties and benefits are discussed. Specific obstacles that may be encountered when implementing a CBL approach to teaching and learning chemistry are also addressed and the main concerns are that it is difficult to find many examples of chemistry CBL projects, subject-associated health and safety and legal considerations and a lack of familiarity among staff and students with the type of critical reflection required. This paper also allows dissemination of some innovative practices, as CBL is not well established in the discipline concerned, chemistry, and it has also not often been applied to the benefit of small local businesses.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to contribute to the understanding of student learning from community engagement by critically assessing the implementation of this pedagogical approach in the context of teaching and learning chemistry and also evaluating the role of personal development in student-community engagement.\nDesign/methodology/approach: - A case study on the implementation since 2007 of community-based learning (also called service-learning) projects in an academic department in Ireland is presented. Analysis of assessment grades, student reflective accounts and evaluation questionnaires informs this work, as does a recently completed self-assessment of our activities using Shumer's Self-Assessment for Service-Learning.\nFindings: - A marked improvement in student engagement and confidence, and their appreciation of how their subject is applied in real-world situations, is reported. Some difficulties arise however, in relation to the level of critical thinking and self-awareness evident in reflective writing assignments, as the students were generally not experienced in reflection on personal experiences. The extent to which it has been demonstrated that genuine community needs are being met is also an issue in some instances.\nOriginality/value: - The paper describes some innovative practices, as CBL is not well established in the discipline concerned, chemistry, and it has also not often been applied to the benefit of small local businesses. A critical analysis of the process followed when CBL activities were implemented in an institution without an existing, well-established culture of applying this pedagogical approach is provided, as this is a perspective not often explored in the literature.\n",
"role": "assistant"
}
] |
An analysis of heterogeneity in inflation expectations across cities in India
|
[
"India",
"Inflation expectations heterogeneity",
"Nightlights"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Inflation expectations are an important marker for monetary policy makers in any economy. With inflation targeting being the explicit goal for many central banks around the world, including India, tracking and analyzing inflation expectations is of primary importance. As pointed out by Rajan (2014), anchoring inflation expectations is essential for achieving inflation targeting. Inflation expectations anchoring are feasible through a monetary policy provided the policy makers are aware of how inflation expectations among the general public are formed to begin with.\n2. The model: In trying to understand why there is heterogeneity in inflation expectations across agents (across cities), we have to start with the question \"how inflation expectations are formed to begin with?\". As pointed out by Coibion and Gorodnichenko (2015a, b), \"[...]. How those expectations are formed, and how best to model this process, remains an open question. From the simple automatons of adaptive expectations to the all-knowing agents of modern full-information rational expectations models, macroeconomists have considered a wide variety of frameworks to model the expectations formation process, yielding radically different results for macroeconomic dynamics and policy implications.\" The authors further point out that in recent times, rational expectations models with information friction can account for otherwise puzzling empirical observations compared to full-information rational expectations models.\n3. Results: We organize our analysis of inflation expectations heterogeneity at the city-level across India in three parts.\n4. Conclusion: The Inflation IESH, conducted quarterly by the RBI, indicates that there is considerable disparity in inflation expectations across cities in India. Why do different cities have divergent expectations about inflation despite coming under a central monetary policy umbrella? This work investigates the factors that might contribute towards heterogeneity in inflation expectations across Indian cities. Using a SUR model for 12 cities (divided on a zonal basis), our results indicate that the presence of information friction plays a significant role in explaining disparity in inflation expectations across cities, whereas the effect of macro-level variables like interest rate, exchange rate, EPU and oil prices, vary from city to city, thereby accentuating inflation expectations dispersion. As India has adopted inflation targeting as its explicit monetary policy target, inflation expectations anchoring is a necessary pre-requisite to attain the end objective. In this backdrop, the presence of considerable heterogeneity in inflation expectations across cities due to information friction, implies that the RBI should enhance its communication with the general public to attain effectiveness of its central monetary objective. As pointed out by Montes and Nicolay (2015), the communication by the central bank acts as a guide to inflation expectations formation. A better understanding of inflation expectations heterogeneity requires access to micro-level data of the survey respondents such that the RBI can target specific groups (gender-wise, age-wise, profession-wise, etc.) for effective communication.\n",
"role": "user"
},
{
"content": "Purpose: The Inflation Expectations Survey of Households, conducted by the Reserve Bank of India (RBI), indicates that there is considerable disparity in inflation expectations across cities in India. The purpose of this paper is to investigate why different cities exhibit heterogeneous inflation expectations despite coming under a central monetary policy umbrella.\nDesign/methodology/approach: First, the correspondence between city-level inflation expectations and city-specific economic characteristics is mapped. Second, how the disagreement in inflation expectations across cities, measured by dispersion, behaves over the business cycle is investigated. Finally, using seemingly unrelated regression technique, the economic factors that play a role in explaining inflation expectations heterogeneity across cities are estimated.\nFindings: Cities with higher economic activity and cost of living have higher inflation expectations. Disagreement across cities regarding inflation expectations rise with an increase in output gap and inflation. Information friction plays an important role in explaining the disparity in inflation expectations across cities, and the effects of macro-level factors vary across cities, thereby accentuating expectations dispersion.\nResearch limitations/implications: Monetary policy-related communication by the RBI (toward the general public) should increase in order to address information friction, which, in turn, would temper down the extent of inflation expectations heterogeneity across cities in India.\nOriginality/value: This is a novel application of the data from the monetary policy perspective. Heterogeneity in inflation expectations across cities or regions is an unexplored area. The use of nightlights as a proxy for city-level economic activity in India (in absence of data on city-level income) is another original contribution.\n",
"role": "assistant"
}
] |
Honest EPS: a measure of GAAP earnings relative to pro forma earnings
|
[
"Earnings",
"Quality audit",
"Factor analysis",
"Stocks and shares"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Since 1987, there has been a significant increase in the number of firms disclosing pro forma earnings (also known as street earnings) in their press release in place of or more prominently than their earnings calculated according to generally accepted accounting principles (GAAP). The definition of earnings is a controversial issue. Earnings according to GAAP are filed with the S.E.C., and are the official US standard for reporting quarterly earnings. Pro forma earnings are unofficial quarterly earnings issued by the company itself. The S.E.C. recently compelled companies issuing press releases that highlight their pro forma quarterly earnings to also highlight their earnings according to GAAP. Pro forma earnings, a measure of operating earnings, is used most notably by Wall Street analysts as the basis for a company's actual quarterly earnings to be compared against an analyst's estimate of its quarterly earnings. This comparison is the criterion for determining whether the company met the \"sell-side\" analysts' or \"street\" consensus quarterly earnings estimate, an important driver of the stock's performance. Investment banking considerations may in part explain why \"sell-side\" analysts accept a non-standard pro forma number supplied by the company since the \"sell-side\" bankers expect to participate in lucrative investment banking opportunities with the same company.Pro forma earnings deviate from GAAP earnings primarily by excluding special, one-time, or non-recurring charges from expenses. GAAP has a method of excluding the extraordinary items from earnings, but it has strict criteria for classifying a charge as extraordinary. In contrast, pro forma reporting is purported to increase the relevance of earnings by evaluating the company as an ongoing concern, but it actually lowers the transparency of earnings in a number of ways. First, since there is no authoritative guidance on the items that companies can reasonably exclude, the firm's managers have considerable room in reporting pro forma earnings. Second, security analysts, investors and the press are increasingly relying on pro forma earnings to interpret corporate information, so the firm managers are under the pressure to make earnings look strong. Pro forma earnings provide managers an opportunity to manage company earnings. Third, due to the ad hoc nature of pro forma earnings, the pro-forma numbers are not comparable across firms or even for the same firm over time. Therefore, the information contained in the pro forma earnings lowers transparency at best and is misleading at worst, especially for less sophisticated investors.In this paper, we provide empirical evidence that a profitable strategy can be formed by taking advantage of the information contained in the difference between GAAP earnings and pro forma earnings. More specifically, we look at the difference between GAAP quarterly earnings as reported by Compustat and pro forma earnings as proxied by quarterly earnings reported by I/B/E/S. We define Honest earnings per share (EPS) as the four quarter difference between GAAP earnings per share before extraordinary items and the pro forma earnings per share, normalized by stock price. We find that the stocks of firms with higher Honest EPS consistently outperform those of firms with lower Honest EPS. Further robustness tests confirm this result when we control for other risk factors, such as size and book-to-market ratio.The investment intuition of Honest EPS is that GAAP earnings and pro forma earnings may be expected to diverge for one quarter, but should again converge over time (four quarters). Companies that consistently inflate their pro forma earnings relative to their GAAP earnings signal a low earnings quality and increase the risk of a subsequent negative earnings surprise. Honest EPS is a signal for identifying either abuse or excessive conservatism as it relates to pro forma reporting. While pro forma earnings drive prices in the short-term (less than one year), eventually the price will adjust toward true economic earnings reflected by proper GAAP accounting. We expect that a company's stock price will have downward pressure if it inflates its pro forma earnings and upward pressure if it is overly conservative with its pro forma earnings relative to GAAP earnings. The idea behind the tests in this paper is that GAAP, with its strict standards on reporting, is a better measure of economic profits and additionally lends itself to greater comparability across stocks. Pro forma earnings have no formal standards and thus are highly subject to manipulation, inadvertent or not.There is considerable research that studies GAAP vs Street earnings. Ciccone (2002) finds that firms with GAAP losses report significantly higher street earnings. Additionally, he finds that firms with volatile GAAP earnings tend to report higher, smoother street earnings. His results suggest that some managers attempt to make financial performance look healthier by reporting inflated street earnings. Bradshaw et al. (2000) points out that the move to street definition of EPS has been accompanied by a dramatic increase in the proportion of corporate expenses that are classified as excluded items. The outcome has been a large and growing gap between the GAAP EPS that is reported in firms' financial statements and the street EPS that is tracked by analysts and priced by investors. Frederickson and Miller (2002) argue that pro forma earnings disclosures affect less sophisticated and more sophisticated investors differently. Their results indicate that non-professional investors who see an earnings announcement that contains both pro forma and GAAP disclosures assess a higher stock price than do non-professionals who see an earnings announcement that contains only GAAP disclosure, whereas analysts' stock price adjustments are not affected.In contrast to prior literature, we provide a simple but powerful profitable trading strategy based on the difference between GAAP and street EPS. Our results reveal that, based on the largest 3,000 stocks, a small yet consistent return can be earned by buying the best and selling the worst \"Honest EPS\" stocks. The abnormal return is significant even after we control for firm size and book-to-market ratios.\nSample: GAAP Quarterly reported earnings are obtained from Compustat, pro forma actual quarterly earnings are obtained from I/B/E/S, while returns are calculated from Compustat. All common stocks are included in the sample after excluding American Depository Receipts (ADRs), limited partnerships, Real Estate Investment Trusts (REITs), mutual funds, Canadian companies, Exchange Trade Funds (ETFs), and subsidiaries of companies.\nHonest EPS variable measurement: We calculate Honest EPS as Equation 1 GAAP earnings in equation (1) are the lagged four-quarter trailing sum of Compustat's primary earnings, excluding extraordinary items, based on GAAP. To prevent a forward-looking bias, we used the \"As First Reported\" (AFR) quarterly earnings, which were available at the time of the report. Using the AFR quarterly earnings mitigates the forward-looking bias of earnings restatements in current (restated) Compustat numbers. When an AFR value is not present, indicating that there has been no restatement, we use the current (restated) Compustat quarterly earnings, since in this case it is the first reported number[1].Pro forma earnings in equation (1) are the lagged four-quarter trailing sum of pro forma earnings number proxied by I/B/E/S actual earnings. The difference between GAAP EPS and the pro forma EPS is normalized by the price at the end of the month when the portfolios are formed. We assume there is a two-month lag between the fiscal quarter ending date and month end date when portfolios are formed to reflect the dissemination of information. If a company's fiscal quarter provided in Compustat for gathering GAAP earnings does not line up with the fiscal quarter available in IBES for gathering pro forma earnings or contains missing data, the stock's observation is omitted[2]. We have utilized a quarterly lag data concept from Charter Oak Investment Systems' Investment Workstation software that correctly matches fiscal to calendar quarters and have assumed a two-month reporting lag in quarterly data. We have found that Charter Oak's lag concept greatly increases the integrity of our data on this seemingly simple function of lagging and addresses an area where most research falls short.\nBuy and hold abnormal return: We calculate the one-, three-, six- and, twelve-month ahead buy-and-hold abnormal returns (BHAR) of firm i as: Equation 2 where Ri,t is the return of firm i in month t, p is the portfolio benchmark to which firm i is assigned, and m=1, 3, 6, 12, respectively, depending on the length of the period under study. We rank stocks into deciles based on the Honest EPS measure each month, then ten portfolios are formed. Buy-and-hold returns for each of the portfolios are calculated and abnormal returns are obtained when we subtract the benchmark returns from the portfolio returns. The buy-and-hold abnormal return (BHAR) calculation period starts two months after the end of the fiscal quarter when the Honest EPS is ranked.Table I shows the summary statistics of GAAP, Street and Honest EPS. Figure 1 shows the Honest EPS by year. Interestingly, there is strong trend that the difference between GAAP and street earnings has become progressively larger during recent years. This finding is consistent with the findings in Ciccone (2002) that use annual observations.\nEmpirical evidence: Table II reports the buy-and-hold abnormal returns of the Honest EPS strategy for the top 1,000 and 3,000 market capitalization stocks. Stocks are ranked based on their Honest EPS measure and allocated into deciles for each month, and a hedge portfolio is formed that is long in the most positive Honest EPS stocks and short in the most negative Honest EPS stocks. The average of deciles return is reported over time and an equally weighted average return is used as the benchmark. As predicted, the decile spread for the top 1,000 universe, next one-, three-, six- and twelve-month returns are 0.47, 1.21, 2.21 and 3.75 percent, respectively, and the corresponding t-statistics are 2.10, 3.60, 5.51, and 7.16.Figure 2 shows the annualized returns of the Honest EPS strategy for the top 1,000 and 3,000 stocks. When we annualize returns, we assume that the portfolio, which is long-top and short-bottom Honest EPS stocks, is rebalanced every month, throughout the year. For the top 1,000 stocks, the Honest EPS strategy generates significant and positive returns 12 out of 16 years, and the performance is especially strong for recent years.Table III displays the correlations between Honest EPS and other commonly used factor anomalies including analyst revision factor, momentum, free cash flow, dividend yield, capital expenditure, and changes in working capital. Analyst revision factor is the monthly difference in the increase and decrease in analyst earnings estimates normalized by the number of earnings estimates provided by IBES. Momentum is the total return from twelve months ago to two months ago provided by CRSP. Dividend yield is the indicated dividend yield by the company provided by Compustat. Free cash flow is the net operating income less capital expenditures normalized by price provided by Compustat. Capital expenditure is the four quarter capital expenditures normalized by depreciation provided by Compustat. Change in working capital is the four quarter change in accruals for working capital items not including cash provided by Compustat. The results indicate that Honest EPS is a highly unique factor that is positively correlated with earning revisions, momentum, and free cash flow factors, but negatively correlated with dividend yield and accruals factors.In order to provide additional evidence about the performance of Honest EPS factor, we also report the results from regressing abnormal returns on Honest EPS. Table IV reports the coefficients and t-statistics of the regressions. Two types of regressions are performed. The first regression is a pooled regression of abnormal returns on Honest EPS. If the Honest EPS measures were serially correlated, the estimation result based on pooled regression would be biased. In order to resolve this problem, we also follow the methodology suggested in Fame and MacBeth (1973). Therefore in the second regression, a cross section regression of abnormal return on the Honest EPS measure is done each month and then pooled cross sectional estimates are used to calculate the mean and t-statistic. Table IV reports the coefficients of both regressions. For all the regressions, the coefficients on Honest EPS are positive. For pooled regressions, all coefficients are significant at 1 percent level.\nConclusion: Our interest in researching the factor we coined as \"Honest EPS\" was the acceleration of companies reporting \"actual\" quarterly earnings that were grossly disconnected from the quarterly earnings based on the US standards or GAAP. It seems incongruous that Wall Street analysts are content to compare their earnings estimate for a company to a number that is subjectively determined by the company they cover. This number referred to as pro forma earnings and recorded by First Call and I/B/E/S as the \"actual\" earnings number and eliminates \"special charges\" or \"write-offs\" that are treated by GAAP as an ordinary expense. This method has become an all too easy way for companies to inflate their earnings. The investment community, both Wall Street \"sell side\" research and the \"buy side\" investment managers, by accepting the pro forma earnings as the \"actual\" earnings gives these inflated earnings the credibility required for the practice to continue.We have shown that investors can profit by selling or shorting stocks that measure low in Honest EPS and by buying or going long stocks that measure high in Honest EPS. We also believe that as long as the market continues to accept this practice of reporting pro forma results as \"actual\" the behavior will continue and Honest EPS will not only persist as an anomalous factor but will strengthen in its signal.\n",
"role": "user"
},
{
"content": "Purpose: - This study empirically examines whether there is an anomalous factor return to a stock's Honest EPS.\nDesign/methodology/approach: - Honest EPS captures any large discrepancy between a company's primary earnings per share and the company's Pro-forma earnings per share, normalized by price. Pro-forma earnings issued in a company's press release have become a pejorative word to investors and the financial media alike.\nFindings: - Honest EPS is rigorously tested for the top 1,000 largest market capitalization stocks and the broader top 3,000 largest market capitalization stocks from 1987 through 2002. It was rebalanced monthly, quarterly, semi-annually, and annually. We show a portfolio long the highest or top decile Honest EPS and short the most negative or bottom decile in which Honest EPS produced significant excess return with a commensurate high information ratio.\nPractical implications: - Investors can use this simple formula for Honest EPS to identify companies that are abusing pro-forma quarterly reporting to inflate their earnings relative to primary earnings reported to the S.E.C. according to generally accepted accounting principles.\nOriginality/value: - An investor can choose to punish the worst Honest EPS companies by short selling their stock, reward the best Honest EPS companies by buying their stock long, or add this as a risk to be aware of in their stock selection.\n",
"role": "assistant"
}
] |
The concept of neutrality: a new approach
|
[
"Libraries",
"Librarians",
"Library and information science",
"Phenomenography",
"Value",
"Conceptions",
"Dimensions of variation",
"Neutral",
"Neutrality"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction and scope: Neutrality is a core - yet controversial - value within library and information science (LIS), and the literature on the subject constitutes a theoretical or opinionated discussion of the merits or demerits of its application. This debate casts neutrality as the property of a dominant positivist conception of LIS (Radford, 1992), a shield for the \"status quo\" in collection development (Iverson, 2008) and a democratic value that eliminates bias (Hart, 2016). To date, these ranging presentations have neither been mapped nor related. As a result, neutrality has become an elusive concept, arguably poorly understood.\nResearch objectives: The study reported here had five objectives:\nLiterature review: This review charts the varied understandings of neutrality in existing debate and forms a backdrop for the following phenomenographic study. First, the conceptual origin of neutrality is outlined. Second, a favourable conception of neutrality in LIS is examined. Third, it is argued that critical literature breaks down into three distinct conceptions. For clarity, these four broad conceptions - one favourable and three critical - are then summarised.\nMethodology: phenomenography: Phenomenography was chosen as the research approach for this study; Marton describes it as \"a research method for mapping the qualitatively different ways people experience, perceive and understand [...] the world\" (Marton, 1986, p. 31), and a means of mapping experiential \"variation\" in an \"outcome space\" (Marton and Booth, 1997, p. 112). Primarily used in education, it is also used in other settings (Bowden, 2000), as epitomised by Theman's research into political power (Marton, 1986, p. 39). It is, therefore, ideal for investigating concepts like neutrality.\nResults: Four qualitatively different conceptions of neutrality emerged from the interviews: the \"core value\" Conception, the \"subservient\" conception, the \"ambivalent\" conception and the \"hidden values\" conception. These were logically related via three dimensions of variation: the perception of neutrality as a positive or negative value, the perception of neutrality as a coherent value and the personal, institutional or societal level of engagement at which neutrality is conceived.\nDiscussion: This section relates the research results to findings to the literature, and addresses the final two objectives, \"to compare categories and literature, looking for similarities and disparities\" and \"to consider normative implications\". The word \"compare\" is preferred to \"triangulate\" due to ambiguity in literature in which triangulation is often associated with \"validating evidence\" (Ritchie, 2003, p. 46). As phenomenography charts \"second order\" conceptions, an unloaded \"comparison\" describes the process of bringing together literature and categorisations, and structural similarities and differences between categories and the literature are discussed. Crucially - rather than correlating \"individual voices\" - the purpose of this section is to link the structure of categories to literature; in other words, it examines which categories are evident in the literature, and which are not.\nContribution to practice: The polarised nature of the \"neutrality debate\" in literature indicates a need for need for further discussion across the profession to raise awareness of differing conceptions of neutrality. This lack of awareness has resulted in all-encompassing conclusions about neutrality's worth from small pockets of professional experience in specific contexts. The beginning of a non-judgmental and inclusive debate would be a first step to a more thorough understanding of how neutrality can both help and hinder the aims of the LIS profession.\nConclusion and contribution: a change of approach: This investigation has identified four distinct conceptions of neutrality spanning three dimensions of variation: the importance of neutrality as a positive or negative, its conceptual coherence and its level of contextual engagement. Categorisations were then compared to literature. Three categories - \"core value\", \"subservient\" and \"hidden values\" - bore structural similarities with arguments identified in literature. Crucially, however, the \"ambivalent\" conception - characterised by an appreciation of neutrality as a slippery and elusive concept - was largely missing. Typically, writers argue \"for or against\" neutrality rather than appreciating the nuances in which neutrality may be beneficial or harmful. This difference was buttressed by noting variation within participants; all participants conceived of neutrality in at least two different ways as the emphasis of questioning changed. This variation suggests that a largely one-dimensional literary debate - construing neutrality as an \"all or nothing\" concept - overlooks its complexity within the profession.\nFuture research: reconciling viewpoints: Future research could develop an ethical framework that provides normative guidance on negotiating the line between neutrality and non-neutrality. Burgess (2016), in an exception to the \"for or against\" character of the neutrality debate, argues that a virtue ethics framework unifies neutrality and social responsibility. As reflected in the \"ambivalent\" conception, Burgess (2016) acknowledges ranging professional contexts that \"complicate the goal of ethical unity\" (p. 163). This variety manifests itself through differing \"cultural perspectives\" that grind against a \"homogenous\" ethical understanding (Burgess, 2016, p. 164), providing an \"opportunity to break out of the discourse of competition between [...] values and instead focus on ways to reconcile the two under a broader understanding of LIS ethics\" (p. 169).\n",
"role": "user"
},
{
"content": "Purpose: Neutrality is a much debated value in library and information science (LIS). The \"neutrality debate\" is characterised by opinionated discussions in contrasting contexts. The purpose of this paper is to fill a gap in the literature by bringing these conceptions together holistically, with potential to deepen understanding of LIS neutrality.\nDesign/methodology/approach: First, a literature review identified conceptions of neutrality reported in the LIS literature. Second, seven phenomenographic interviews with LIS professionals were conducted across three professional sectors. To maximise variation, each sector comprised at least one interview with a professional of five or fewer years' experience and one with ten or more years' experience. Third, conceptions from the literature and interviews were compared for similarities and disparities.\nFindings: In four conceptions, each were found in the literature and interviews. In the literature, these were labelled: \"favourable\", \"tacit value\", \"social institutions\" and \"value-laden profession\", whilst in interviews they were labelled: \"core value\", \"subservient\", \"ambivalent\", and \"hidden values\". The study's main finding notes the \"ambivalent\" conception in interviews is not captured by a largely polarised literature, which oversimplifies neutrality's complexity. To accommodate this complexity, it is suggested that future research should look to reconcile perceptions from either side of the \"neutral non-neutral divide\" through an inclusive normative framework.\nOriginality/value: This study's value lies in its descriptive methodology, which brings LIS neutrality together in a holistic framework. This framework brings a contextual awareness to LIS neutrality lacking in previous research. This awareness has the potential to change the tone of the LIS neutrality debate.\n",
"role": "assistant"
}
] |
Threshold capability development in intensive mode business units
|
[
"Management",
"Accounting",
"Threshold concepts",
"Business",
"Intensive mode",
"Threshold capabilities"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The theory of threshold concepts was first proposed by Meyer and Land (2003), and asserts that every discipline has concepts that are transformative ontologically and epistemologically. Since this time the theory has been explored in discipline areas such as mathematics, engineering, gender studies and business. The theory of threshold concepts was expanded by merging it with capability theory, which proposes that in higher education students should develop capabilities to address unseen problems (Bowden, 2004). According to Baillie et al. (2013) threshold capabilities represent the ability to respond to previously unseen problems. These capabilities are thresholds themselves and usually also require understanding of other threshold concepts. Like threshold concepts, threshold capabilities are transformative and critical to future learning and practice (Baillie et al., 2013). Therefore they are of paramount importance when developing curricula. As more intensive modes of teaching are used in postgraduate business education it is critical those charged with curriculum design are aware of the mode's impact on student attainment of threshold concepts and capabilities.\n2. Theoretical framework: According to Meyer and Land (2005), most disciplines have threshold concepts that act as \"conceptual gateways\" or \"portals\" to once inaccessible and potentially \"troublesome\" ways of thinking. These threshold concepts are critical to future learning and working in the discipline and are usually troublesome for students in one of several ways such as being foreign, complex, counter-intuitive or requiring unfamiliar language (Perkins, 2006).\n3. Context: This paper reports a study of students' experiences of threshold capability development in two graduate business units: an accounting unit taught in an intensive mode and the same accounting unit taught in a flexible (more traditional) mode, and a unit on strategic management taught in intensive mode. These units were part of the MBA programs at two Australian universities, and are classified under the Australian Qualification Framework as a level nine qualification, where:\n4. Method: The method used to examine student experiences of threshold concepts and development of threshold capabilities with IMT was adapted from that developed by Male and Baillie (2011) and involved two phases.\n5. Findings and discussion: In this section we explore the key findings from the two units included in the study, looking at the types of threshold concepts and capabilities identified by students and the factors that hindered and supported them in attaining these. Included in the findings are the views of the unit coordinators about important learning outcomes and threshold concepts for students.\n6. Implications and conclusions: A number of key implications and conclusions emerge from this research project. It further demonstrates the utility of the threshold concept and threshold capabilities framework in understanding student experiences of business education. Students undertaking the strategic management and accounting units in this study were able to identify a number of threshold concepts that were potentially troublesome, and upon which they could become stuck. Similarly, there was some evidence of the development of threshold capabilities in these cohorts. Further research could be undertaken across a range of discipline areas to further understand the utility of these frameworks to the design, delivery and assessment of the business curriculum.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore: student perceptions of threshold concepts and capabilities in postgraduate business education, and the potential impacts of intensive modes of teaching on student understanding of threshold concepts and development of threshold capabilities.\nDesign/methodology/approach: The student experience of learning was studied in two business units: strategic management, and accounting. The method involved two phases. In the first, students and unit coordinators identified and justified potential threshold concepts and capabilities. In the second, themes were rationalized.\nFindings: Significantly more so in intensive mode, the opportunity to ask questions was reported by student participants to support their development of the nominated threshold capabilities. This and other factors reported by students to support their learning in intensive mode are consistent with supporting students to traverse the liminal space within the limited time available in intensive mode.\nResearch limitations/implications: Respondents from future cohorts will address the small participant numbers. Studies in only two units are reported. Studies in other disciplines are presented elsewhere.\nPractical implications: The findings will be important to educators using intensive mode teaching in business, and researchers working within the framework.\nOriginality/value: This is the first study to explore the potential impacts of intensive modes of teaching on student understanding of threshold concepts and development of threshold capabilities.\n",
"role": "assistant"
}
] |
The politics of the executive, legislative veto players and foreign debt
|
[
"The executive",
"Legislative veto players",
"Foreign debt",
"Curvilinear effect",
"Empirical analysis"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: During the 1980s and early 1990s, many developing countries were plagued by astronomical foreign debt and the possibility of default. The debt crisis could have shattered the entire infrastructure of international financial and political systems (Gilpin, 1987, p. 317). As a result, academics and policy makers alike have had great interest in explaining and understanding why and how the debt crisis occurred. Although economists were the first to pay attention to the causes of the crisis by examining various negative macroeconomic developments, including a worldwide recession, reduced export sales and a sharp increase in interest rates, they tended to overlook political dimensions (e.g. Cline, 1995; Manzocchi, 1997). Several political economists have started to fill the gap by exploring the effect that such political factors as state autonomy, political instability and democratic governance had upon the crisis (e.g. Kaufman and Stallings, 1989). However, prior analyses have neglected the importance of institutional settings in analyzing how foreign debt is determined, especially variations in the policy-making power configuration between the executive and the legislature.\n1. Brief literature review: The existing literature can be categorized into three groups. The first group of studies provides macroeconomic explanations. They link economic development, economic growth, exports, budget deficit, government consumption and IMF participation to the rise and fall of foreign debt (e.g. Wiesner, 1985; Kaufman and Stallings, 1989; Cline, 1995; Manzocchi, 1997; Potter, 2000; Vreeland, 2007). For example, quantitative studies such as Cline (1995) and Aggarwal (1996) explore the determinants of sovereign default in times of economic crisis. The second group pays attention to the type of regime (e.g. Olson, 1993; Oatley, 2010; Dasgupta and Ziblatt, 2016). It contends that democratic and autocratic countries have different political discount rates, which influence the borrowing and investment decisions of the government in a different way. For example, Easterly (2002) argues that because autocrats wish to stay longer in power, they are more likely to continue to borrow from the future in order to entertain/bribe their key supporters. By underscoring institutional constraints on the top executive as the democratic advantage, the third group further refines the second group (e.g. Cox and Saiegh, 2018). North and Weingast (1989) are the first to conceptualize that constitutional checks and balances play an important role in resolving commitment problems for the repayment of public debt. As government commitment and credibility increases, interest rates on public bonds should fall and reduce the need for debt financing. But several researchers later dispute the democratic advantage argument. For example, Biglaiser and Staats (2012) demonstrate that credit-rating agency bond raters do not care about the regime type itself but are rather interested in the stability of a certain government system, whether it is democratic or not. Along the same line, Yu's (2016) cross-sectional, time-series analysis for the years 1970-2010 finds that political instability increases the likelihood of default.\n2. The executive, legislative constraints and foreign debt: Why and how foreign debt changes for individual countries may require multiple causal explanations if the uniqueness of each case were to be scrutinized. To illustrate, Nicolae Ceausescu, the dictator of Romania from 1965 until December 1989, decreased debt by refusing to allow new loans; in contrast, however, Argentina had an explosion of debt when the central government was forced to take over the loans of the provinces. Although each of these cases should increase our knowledge about a particular debt problem, it pays little attention to finding a general pattern that reflects the nature of the ebb and flow of foreign debt across a large number of countries over several time periods. In this study we argue that, ceteris paribus, the political constraints which legislative veto players impose on the executive are an important predictor for foreign debt.\n3. Statistical model building and operationalization: To explore the hypothesized linear and curvilinear effects of legislative constraints on foreign debt empirically, two statistical models are considered:(1)\n4. Results from testing: Table III shows the multivariate regression results. The results for the inverse linear equation are displayed under odd model numbers 1 and 3, while those for the curvilinear equation are under even model numbers 2 and 4. Models 1 and 2 depict the results using the fixed-effects regression, but Models 3 and 4 show the results using the maximum-likelihood random-effects regression. A one-tailed test for each variable is employed because all the hypotheses are directional. Following the common practice in studies of international political economy, 0.01, 0.05 and 0.10 are used as significance levels (e.g. Jensen, 2003; see also Li, 2005). Because the estimated autocorrelation coefficient, r, across the models is close to zero, no serious serial correlation of the first order is suspected.\n5. Conclusion: How domestic institutional settings affect foreign debt is an important issue to academics and policy makers alike. However, there has been no existing longitudinal empirical study that explores the determinants of foreign debt from the perspective of the policy-making power dynamics between the executive and the legislature, a relatively understudied area for political scientists. We have conceived of foreign debt as a product of an interaction between re-election-minded executives and congressional legislators, assuming that both pursue distributive or particularistic politics to stay in office longer. Specifically, we have underlined that legislative constraints stemming from the number of legislative veto players, their policy preferences, and their internal coherence should produce a curvilinear influence on the fluctuation of foreign borrowing in developing countries. We have found that in highly constrained developing countries in which legislative constraints outweigh the executive's veto power, the rate of change in foreign debt increases because multiple legislative veto players are likely to increase spending for their particular constituents. In highly unconstrained developing countries in which the executive is free from legislative opposition, the rate of change in foreign borrowing increases because the executive is likely to initiate spending legislation targeted at his/her political supporter. However, in developing countries where the policy-making power is equally distributed between the executive and the legislature, the rate of change in indebtedness is likely to be low because the power parity hinders each institution from fueling the impulse to particularism.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine a curvilinear effect of legislative constraints on foreign debt.\nDesign/methodology/approach: A cross-sectional, time-series data analysis of 68 developing countries during the period from 1981 to 1999 was performed.\nFindings: Foreign borrowing is most likely to increase at both low and high levels of legislative constraints, while it is most likely to decrease at moderate levels.\nOriginality/value: The paper is a first-cut empirical analysis of a curvilinear relationship between legislative constraints and foreign debt.\n",
"role": "assistant"
}
] |
Automatic replenishment of perishables in grocery retailing: The value of utilizing remaining shelf life information
|
[
"Information sharing",
"Simulation",
"Shelf life",
"Replenishment"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Perishables are of major importance for grocery retailers and account for 25 percent of the total sales and more than 35 percent of the growth in the European grocery market (Nielsen, 2016). Compared to other food products, the main difference is the shorter shelf life (less than 30 days) (Van Donselaar et al., 2006). Food waste is often reported to be higher for these perishables products (Kaipia et al., 2013; Mena et al., 2014) and the short shelf life complicates the inventory management practice (Karaesmen et al., 2011). Non-perishable products are typically managed through an automatic replenishment system, which generates order proposals to the stores based on point-of-sales, safety stock (SS) levels, batch sizes and delivering times (Potter and Disney, 2010). These systems have shown to improve performance of grocery retailer; however, they do not function satisfactorily in its original form for perishable products (Van Donselaar et al., 2006).\n2. Background: In perishable, inventory management demand is modeled either as deterministic or stochastic, and the shelf life is either considered as fixed or random. A fixed shelf life refers to a known deterministic life time, whereas random shelf life refers to a known or unknown probabilistic life time (Janssen et al., 2016). In this study, focus is placed on modeling products with a fixed shelf life under stochastic demand. Within grocery retailing, this includes products such as: dairy, meat, cold cuts, or other perishable products with a predetermined expiration date. Findings of how to manage these products with a fixed shelf life may later be extended to include products with a random shelf life.\n3. Development of modified EWA policy: As observed by Van Donselaar and Broekmeulen (2012), the order quantity proposed by the EWA policy may be too high in some cases due to how SSs are included in the policy. Generally, for products with a short shelf life, the EWA policy will place orders earlier than a base stock policy to account for the products that outdate. Consequently, if orders are placed earlier and the same SS levels are kept (as if a base stock policy was used), we risk having too many products with a limited RSL on inventory. This will result not only in a very high service level but also increase the risk of products outdating. As the predetermined shelf life of the product increases (say, above 15 days), the risk that products outdate on the shelf may decrease. And, as the number of products which outdate approaches zero, the EWA policy becomes equal to a normal base stock policy (Broekmeulen and van Donselaar, 2009). Therefore, to make a better balance and not just add to the order if there is a high number of products outdating, it is proposed to slightly modify the EWA policy.\n4. Research design: To investigate the performance of the EWA and EWASS policies in a more realistic setting than current literature, the analysis in this study is based on a simulation model of a Norwegian grocery retailer. To estimate the impact of utilizing RSL information through these policies, the simulation model mirrors the structure and settings (delivery times, service levels, review periods, etc.) and uses POS data from one year to reflect the sales pattern. Using simulation allows us to test various scenarios and investigate the causality between variables (Croom, 2009).\n5. Numerical results: To evaluate the applicability and performance of the EWA and EWASS policy, three main scenarios with different replenishment policies between the stores and the warehouse were established: baseline scenario, EWA policy and EWASS policy. The simulations were run for one year (plus four months warm-up period for the forecasting procedure) and for each scenario, the shelf life of the product was gradually increased (with 1 day) from 4 to 20 days. These limits were made because the total lead time through the supply chain is at least 2x38 h and products with a shelf life less than four days would have expired before they reached the stores. Additionally, no changes were observed with a shelf life above 20 days.\n6. Discussion and conclusion: The purpose of this paper is to investigate the impact of utilizing RSL information from grocery retail stores for the replenishment of perishables. A discrete event simulation model has been developed involving one warehouse, one product, 232 stores, mixed FIFO and LIFO depletion, differentiated service levels, batch size constraints and ordering frequencies. Two age-based replenishment policies were evaluated based on one year POS data from one of Norway's largest grocery retailers.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the impact of sharing and utilizing remaining shelf life (RSL) information from grocery stores by the use of age-based replenishment policies for perishables.\nDesign/methodology/approach: The performance is evaluated through a discrete event simulation model, which mirrors a part of one of Norway's largest grocery retailer and uses their POS data to reflect a realistic demand pattern of 232 stores for one year.\nFindings: The findings indicate that a current age-based replenishment policy (EWA policy) provides a significant improvement of 17.7 percent increase in availability for perishables with a shelf life between 4 and 11 days, but suffers from high inventory levels and only reduces waste by 3.4 percent compared to a base stock policy. A proposed adjustment to the EWA policy, EWASS, provides a more balanced performance in the conducted study with a reduction of 10.7 percent waste and 10.3 percent increase in availability by keeping the same average inventory level.\nPractical implications: Sharing and utilizing RSL information for replenishment of perishables with a predetermined shelf life between 6 and 11 days can be beneficial, and could enable the replenishment processes to be automated. However, for products with longer shelf life, the benefits slowly diminish.\nOriginality/value: The study proposes a new age-based replenishment policy which in the conducted study showed a more balanced performance improvement, in both waste and availability, compared with previous replenishment policies.\n",
"role": "assistant"
}
] |
Collection building trend among the institutes of higher learning in India: a preferential race between print and electronic resources
|
[
"India",
"Higher education",
"Collection building",
"E-documents",
"Library budget",
"Print documents"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Libraries are known as the jugular veins of the academic and research institutions. These sub-institutions supply the necessary oxygen to the other parts of the institution for sustenance. No institution of whatsoever size can sustain in the absence of a good library. S. Radhakirshanan (Bhatt, 2010), the first Vice-President and the second President of India, has rightly said that library is the heart of any educational institution.\nEarlier studies: Drawing comparison between the procurement and the subscription of electronic and print resources in academic libraries has been studied by researchers all across the world. The earlier studies concerning to collection development give us a better idea, as how collection development practices among libraries changed gradually but steadily among the libraries over a period of time. Some of the earlier works undertaken in the field of library collection development and fund spending patterns have been reviewed hereunder to give a better idea about the subject under study. In a study undertaken at Dresel University's W.W. Hagerty Library on the fund-spending pattern of the university library, Montgomery and King (2002) found that there is a constant shift towards the procurement of electronic resources. The authors observed that after considering all the cost, subscription of online journals is more cost-effective. Round the clock availability and accessibility of online journals is considered as a major advantage, whereby library professionals no more route journals among information seekers. Keeping in view the demand for both the types of resources, libraries are somewhat impelled to make allocations for the procurement of both the types of resources. In a similar study undertaken on expenditure involved in the procurement of resources and extending services in a conventional library to that of digital libraries, Connaway and Lawrence (2003) found that labour and space requirement are more in paper than in a digital library. The authors even pointed out that the practice of maintaining both paper and digital libraries simultaneously incurs an extra expenditure.\nObjective of the study: This study aims to analyse the collection building trend across the libraries of leading Institutes of Management, Engineering and Technology and Universities in India and analyze the percentage of budget spent on the procurement of electronic and print resources in the libraries associated with academic institutes across India.\nMaterials and methods: The present study is based on the ranking list released by the Ministry of Human Resource & Development, Government of India, on 4 April 2016. Accordingly, from the ranking list, 20 leading institutes each were chosen from the Institutes of Management, Institutes of Engineering and Technology and Universities, and the data were retrieved from the official website of the Ministry of Human Resource & Development, Government of India, which can be accessed at www.nirfindia.org/Ranking (Ministry of HRD, GoI, 2016). The identified institutes were classified into three different groups as per their ranking list. Given the top rankings of the institutes under study, presumably these institutes meet the set standards and norms and so should we hold true about their libraries. The analysis has been undertaken on the data retrieved for the period 2012-2013, 2013-2014 and 2014-2015.\nScope: The present study is limited to the geographical confines of India, the country which has over 1.25 billion population, with 65 per cent population below 35 years of age, having more than 750 institutions of higher learning spread across the length and breadth of the country. Because the study reflects the trend among institutions of higher learning in India, as such the outcome of the study can be generalized, especially among the countries which belong to the same League of Nations in terms of development.\nResults: The retrieved data have been tabulated as per the objective of the study, and to perform the simple mathematical expressions such as addition, subtraction, division or drawing percentage, the data were entered in MS Excel. Percentage at most of the places has been drawn up to two decimal places and has not been rounded off, and as such may show slight variation while computing figures for 100 per cent (Table I).\nDiscussion: From the overall data analysis, it emerges that the library document procurement preferences in the institutions of higher learning and research in India have changed considerably. Even though there is no uniform trend which we may say emerged from the overall analysis, the preference varies considerably from institution to institution, mostly depending upon the information need of that particular institute. This gets corroborated by the fact that most of the Institutes of Engineering and Technology and Institutes of Management have spent major portion of their library budgets on the procurement of e-resources, but there are some Institutes which have given more preference to print resources. Similarly, a good number of universities have either shown preference towards procurement of print resources or electronic resources. Moreover, the document procurement preference among Institutes of Engineering and Technology and Management is more towards e-resources; however, print is still the preferred source of information among the university libraries. The procurement preference among institutions under study varies considerably from institution to institution. Although there is no specific reason for this considerable difference in the procurement preferences, the factors which play their part in deciding the procurement preferences are age of institution, infrastructure, richness and resourcefulness of the institutional library in terms of print or electronic documents, access to internet facility, skillfulness of human resource towards the exploitation of e-resources and many more.\nConclusion: There is no denial in the fact that both print and electronic documents have equal importance, but both do not enjoy the same level of readership. Given the readership priorities of library clientele and some advantages which e-documents have over print, the institutional libraries are somewhat impelled to procure more electronic documents. Both print and electronic documents enjoy readership of their own kind, but that does not mean preference of form of information source consulted or not consulted by a specific group of people does not make them anyway more or less informed. As a fact of matter, it is only the form of document which varies and not the content, which altogether is same in both the forms of documents.\n",
"role": "user"
},
{
"content": "Purpose: The present study aims to assess the resource procurement preferences among the leading academic libraries of the institutes of higher learning across India. The study analyses the amount and percentage of budget spent by the institutions under study during the period 2012-2013, 2013-2014 and 2014-2015 on the procurement of print and electronic resources in their libraries. The study also discusses about concerns with regard to the subscription of e-resources and the reasons for weaning interest among the Indian academicians towards the print resources.\nDesign/methodology/approach: By and large, the scope of the study is limited to institutes of higher learning across India, but reflects a great deal of relevance to the global higher education practices. The study has been undertaken on the 20 leading institutes each from Universities, Institutes of Management and Institutes of Technology in India, based on the ranking list released by the Ministry of Human Resource Development, Government of India, on 4 April 2016.\nFindings: The electronic documents are the dominant forms of present day library procurements in India, with the print struggling to retain its readership. The Institutes of Engineering and Technology and Management have almost similar procurement preferences, spending almost three-fourth of their library budgets on the procurement of e-documents. On average, the University libraries in India have spent 43.54 per cent of their budget on the procurement of print resources and 56.46 per cent of their budget on the procurement of electronic resources during the period of study. Similarly, on average, each Institute of Engineering and Technology has spent 24.04 per cent of their library budget on the procurement of print documents and 75.96 per cent towards the procurement of electronic resources, whereas 27.51 per cent of the library budget by the Institutes of Management in India is being spent on the procurement of print resources and 72.49 per cent towards the procurement of electronic resources.\nOriginality/value: The study is original and is first of its kind, undertaken on three different category of institutions, based on the national ranking of the country.\n",
"role": "assistant"
}
] |
Using put option contracts in supply chains to manage demand and supply uncertainty
|
[
"Supply chain management",
"Demand uncertainty",
"Put option contracts",
"Yield uncertainty"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Nomenclature\n1. Introduction: The real option was first designed to amplify fortune or mitigate loss in the capital investment decision making under uncertainty (Lander and Pinches, 1998). Over the last 40 years, the real option has drawn constant attention from scholars in economic field. Different theoretical types of real options have been developed in the early literature, including growth options (Myers, 1977), options to abandon (Bonini, 1977), options to defer (Tourinho, 1979) and scale options (Trigeorgis and Mason, 1987), etc. Meanwhile, various option-type decision-making frameworks have been developed to model and value these real options and derive optimal investment strategies, such as continuous-time models (Black and Scholes, 1973), finite-difference schemes (Brennan and Schwartz, 1978), and trinomial models (Kamrad (1995) and binomial models (Lander and Pinches, 1998). Lander and Pinches (1998) stated that most types of the early models are complex and require substantial mathematical techniques to solve, which are often violated in the application of real options. Later studies endeavor to propose more practical frameworks/models and solutions, for example, the papers of Cassimon et al. (2011), Zhang et al. (2014), Nigro et al. (2014) and Cassimon et al. (2016). Due to its success in investment market, many flexibility contracts derivated from the real option have been widely applied by practitioners and scholars to manage supply chain risks in the last two decades (Barnes-Schuster et al., 2002; Nagali et al., 2008; Chen and Shen, 2012).\n2. Related literature: As suggested in the previous section, issues of randomness involved in production process and real option contracts in supply chain are the essential research topics this paper touches. We provide a brief review of the related literature below. Then, we refer to a tabular to show the position of this work.\n3. Model formulation: The sequence of events goes as follows:\n4. Centralized model--the benchmark case: To obtain a benchmark for decentralized models, we first consider the centralized case where the manufacturer and the supplier are assumed to belong to the same company. Let QI denote the centralized supply chain's decision on how many products (components) to produce, then the expected profit of the centralized system is:(1)\n5. Decentralized models with put options: In the decentralized case, the manufacturer and the supplier are primarily concerned with optimizing their own objectives. In this section, we develop models to derive and analyze the manufacturer's optimal ordering policy and the supplier's optimal production policy with put options.\n6. Effect of put options: Lemma 2 and Lemma 3 suggest that there is a unique Nash equilibrium between the order quantity of the manufacturer and the production input of the supplier under the decentralized system with put options. Based on the research findings, in this section, we will investigate the effect of put options on the decision making and profit of supply chain management by comparing with the case of without put options.\n7. Supply chain coordination: Now we discuss the supply chain coordination. In this section, we focus on two questions: Can the supply chain with random yield and stochastic demand in the presence of a spot market be coordinated under put option contracts? If not, how can the supply chain coordination be achieved?\n8. Conclusions and future research: The paper studies on the application of put option contracts in supply chains where the supplier is subject to random yield and the manufacturer faces a stochastic market demand in the presence of a spot market. We take into consideration that both the manufacturer and the supplier can buy or sell the components on the spot market. Table I shows that this is the first study to apply put option contracts in supply chain management with random yield. By developing supply chain models incorporating put option contracts and random yield, we derive several management insights as follows:\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to value put option contracts in hedging the risks in a supply chain consisting of a component supplier with random yield and a manufacturer facing stochastic demand for end products.\nDesign/methodology/approach: This paper adopts stochastic inventory theory, game theory, optimization theory and algorithm and MATLAB numerical simulation to investigate the manufacturer's ordering and the supplier's production strategies, and to study the coordination and optimization strategies in the context of random yield and demand.\nFindings: The authors find that put options can not only facilitate the manufacturer's order but also the supplier's production, that is, the manufacturer and the supplier can effectively manage their involved risks and earn more expected profits by adopting put options. Further, the authors find that the single put option contract fails to coordinate such a supply chain. However, when combined with a protocol, it is able to coordinate the supply chain.\nOriginality/value: This paper is the first effort to study the intersection of put option contracts and random yield in the presence of a spot market. From a new perspective, the authors explore the supply chain coordination. The authors propose a mechanism to coordinate the supply chain under put option contracts.\n",
"role": "assistant"
}
] |
Disclosure of governance information by small and medium-sized companies
|
[
"Small enterprises",
"Governance",
"Alternative investment markets"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: An overwhelming majority of investors are willing to pay a premium for companies with high corporate governance standards (Berghe and Levran, 2003). For major investors, such as the pension and hedge funds[1], transparency is increasingly becoming the primary corporate governance concern (Hebb, 2004). This study concentrates on small and medium-sized companies (SMEs) listed on the alternative investment market (AIM). In 1995, AIM was established by the London Stock Exchange (LSE) in order to provide a simplified regulatory environment specially designed for the needs of smaller companies. For example, there is no trading record required, no minimum shares to be in public hands and no minimum market capitalisation. Instead, all prospective companies are required to have a nominated advisor, whose duty is to ensure that the companies are suitable for AIM and ready to be admitted to a public market (London Stock Exchange, 2004). AIM allows SMEs to operate in a capital market setting. AIM enables smaller companies to enjoy access to different sources of finance, and if they intend to grow large enough, to be able to eventually move on to the main market (i.e. LSE). AIM-listed companies are often small and fast growing companies that enjoy status and credibility through their listing. AIM has become a globally unique market in providing SMEs with access to capital markets; currently even foreign SMEs can join the market (Shaver, 2006). However, evidence shows that money channelled into SMEs is falling because institutional investors are losing interest in them (Bartram, 2003).Given that AIM-listed companies endeavour to access external funds in highly competitive markets, transparency of governance mechanisms is significant for gaining and maintaining investor interest. The existing literature offers no evidence on the extent of compliance with governance regulatory requirements by SMEs. Hence, this paper concentrates on the disclosure of governance information by AIM-listed companies, in order to examine the extent of compliance with the governance regulatory codes and requirements. The paper also aims to unravel those corporate governance and firm structure-specific attributes popular among those AIM-listed companies that comply with the governance regulatory codes and requirements.\n2. Corporate governance and small and medium-sized companies: Since the recent clamour for better corporate governance in companies, the reviews of compliance with the Cadbury Code have revealed some gaps in compliance, particularly in the case of small companies (Mallin and Ow-Yong, 1998).The difficulties SMEs may encounter in compliance have been recognised by the Cadbury Report (1992) and the Combined Code (Financial Reporting Council, 2003). The Cadbury Code therefore recommends that smaller listed companies which cannot comply with some requirements of the code should instead give their reasons for non-compliance (Cadbury Report, 1992: para. 3.15). The Combined Code gives special consideration to smaller companies in the area of board composition, and the compositions of the audit and remuneration committees (para.s A.3.2, B.2.1 and C.3.1).AIM companies are required to prepare annual reports in accordance with AIM rules, which do not require companies to make disclosures about compliance with the Combined Code. However, in 2004 the Quoted Companies Alliance urged AIM companies to use the principles of the corporate governance codes as goals to be achieved (Deloitte, 2005). The disclosure practices of AIM-listed companies can therefore be said to be voluntary (Nordberg, 2004). Traditionally, managers of smaller companies are expected to concentrate on their companies' day-to-day issues and challenges, ensuring their corporate survival rather than dealing with such strategic issues as compliance with governance codes. Nevertheless, managers of AIM-listed companies intend to have access to external funds through the capital market, hence they need to assure their existing and potential investors (private or institutional investors or venture capitalists) of the soundness of their governance mechanisms. Many SMEs are said to be striving to comply with Combined Code requirements (Nordberg, 2004; Collier, 2004) and that such compliance will engender investor confidence (Deloitte, 2005) and institutional investor interest (Nordberg, 2004). At the same time, it is important to note that the extent of compliance with governance codes is likely to be at a higher level for SMEs that intend to grow and eventually move on to the main capital market than for SMEs with no such intention.In the next section, we discuss two theoretical perspectives that can be used to explain why AIM-listed companies comply with governance regulatory codes.\n3. Theoretical perspectives: The crux of principal-agent theory is that managers as agents of shareholders can exhibit behaviour and take decisions that may not necessarily maximise shareholder wealth. Opportunities for managers to use shareholders' funds in ways not in the interests of principals or squander them on non-wealth maximising activities are well documented (Shleifer and Vishny, 1997). According to this theory, adequate monitoring or control mechanisms need to be established to protect shareholders from conflicts of interest (Kiel and Nicholson, 2003). An effectively structured board, performance-related compensation contracts, and concentrated ownership holdings exemplify internal mechanisms that encourage active monitoring of managerial decision making processes (Kiel and Nicholson, 2003). The benefits of a sound and effective governance mechanism are expected to outweigh and not necessarily to eliminate but to mitigate its costs. It is argued that a large group of corporate outsiders need to control the incentives of a large group of insiders (Weinberg, 2003).Traditionally, most studies of principal-agent theory have focused on investors as the principal. Shankman (1999, p. 332) challenges this view and calls for a new paradigm in which corporate interests are seen as an extension of the interests of a larger community and contends that stakeholder theory is a logical conclusion of agency theory where different stakeholder groups are recognised and considered. At the same time, Heath and Norman (2004) warn against a multi-principal problem if managers are held accountable to the interests of all different interest groups and take the view that, with the growing interests in corporate social responsibility and the increasing recognition of a range of stakeholder groups, there is a need for reforms in governance mechanisms whereby stakeholder groups are represented in governance mechanisms and the regulatory codes and requirements (Corporate law) formally consider the interests of non-shareholding stakeholders.Even though the existing governance regulatory codes do not explicitly relate to non-shareholding stakeholders, there are some aspects of the codes that do so indirectly (Smith Report, 2003, para.s 5.1-5.36). Identification of the type of SMEs that tend to comply with the codes and report more on their governance mechanisms should help us with a better understanding of which one of the two theoretical perspectives discussed above provide the most powerful explanation for governance practices among AIM-listed companies. We therefore focus on those types of companies with relatively transparent governance arrangements.\n4. Research questions and hypotheses: We concentrate on finding: relationships between various corporate governance characteristics and the level of governance information disclosure, and relationships between firm structure-specific characteristics and the level of governance disclosures.To answer the first question, we selected five corporate governance characteristics considered by previous studies to impact on governance disclosure practices adopted by large companies. For the second research question, two firm characteristics - financial leverage and ownership structure - were selected. All the studies reviewed in the following sub-sections to form the basis of our hypotheses are of large companies. This approach creates knowledge on the extent of corporate disclosure by SMEs, and allows comparison of the results with larger companies.4.1. Corporate governance characteristics\n5. Sample selection and data collection: The sample comprises companies listed on AIM on or before 30 June 2001. This study concentrated on three years, 2002, 2003 and 2004[2]. We included all the companies that were active in each of the three years. Overall, 89 companies, excluding all the financial companies, were included in the final sample. Company details were obtained from two separate sources. First, Datastream for data pertaining to financial leverage, profitability and firm size; and secondly, the respective companies' web sites and annual reports for information including corporate governance characteristics and firm structure-related characteristics.5.1. Corporate governance disclosure index\n6. Data analysis: The Mann-Whitney test was applied to examine whether there were significant differences in the disclosure levels of specified groups of companies. This test was suitable as all the assumptions, including the normality assumption and the equal variance assumption discussed by Huck (2004), were met for the sample. We also used the Spearman Rank Correlation to measure the significant level of association between governance disclosure scores and the above variables. We applied multiple regressions to assess variability of disclosure by the individual corporate governance and firm structure-specific characteristics. The regressions helped establish the strength and direction of existing variables and instances of co-linearity among the independent variables.6.1. Corporate governance disclosure\n7. Summary and conclusion: Our evidence illustrates that, in all the three years under investigation, companies had reported on about 50 percent of governance items specified by the regulatory bodies. Of the seven hypotheses developed, five were rejected. Only two, board independence (H2) and composition of the audit committee (H4), were supported. Positive associations exist between the presence of non-executive directors on boards of directors and audit committees and governance information disclosure. This illustrates that the higher the number of non-executive directors, the more fully SMEs report on their governance mechanisms. The regression analysis reveals the causality between the number of non-executive directors on the audit committee and the level of governance disclosure, indicating that the presence of non-executive directors is a determining factor in the transparency of a given governance mechanism.Our finding on board independence is consistent with those of Haniffa and Cooke (2002) and Cheng and Courtenay (2006) who found that the proportion of non-executive directors on boards was positively related to the quality and extent of financial disclosure. In the case of the audit committee, our findings provide supporting evidence to those of Forker (1992) and Ho and Wong (2001), indicating that the number of non-executive directors on the audit committee determines the transparency of governance disclosure. The audit committee has a key role in governance structures with a wide range of responsibilities including identification, assessment, management and monitoring of financial and non-financial risks that a company may face (Smith Report, 2003: para. 5.6), dealing with whistleblowers (Smith Report, 2003: para. 5.9), and the quality and integrity of the internal control processes (Smith Report, 2003: para.s 5.10-5.13). These responsibilities may concern issues related to shareholders or non-shareholding stakeholders, such as employees and society at large. In the absence of any mandatory requirements, the fact that more than 50 percent of AIM-listed companies complied with the recommendations on the audit committee reflects their recognition of the significant roles and responsibilities the committee has on issues and matters that may relate to non-shareholding as well as shareholding stakeholders. As fast growing companies facing limited resources in an increasingly competitive capital market, AIM-listed companies seem to consider audit committees as serving a wide range of purposes for them. Hence, our evidence points towards AIM-listed companies being more likely to have adopted a stakeholder approach than merely concentrating on their shareholders and shareholder interests. In the case of growing companies therefore, we conclude that stakeholder theory has more explanatory power than principal-agent theory.Our policy conclusion is that in view of the increasing attention that institutional investors pay to corporate transparency and governance mechanisms, the presence of more non-executive directors within such mechanisms seems to be the key to better adherence to such codes and regulations. Increasing disclosure volume may remedy declining investor interest in SMEs.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to examine the extent of compliance with the governance regulatory requirements by small and medium-sized companies (SMEs) listed on the alternative investment market (AIM).\nDesign/methodology/approach: - The paper focuses on AIM-listed companies over a period of three years (2002, 2003 and 2004) and concentrates on their extent of compliance with the corporate governance disclosure guidelines set out by the regulatory bodies. To measure the extent of disclosure, a checklist was compiled based on the Combined Code and FSA listing rules. Having reviewed the literature on large companies (as there were no studies on SMEs), a number of governance and firm structure-specific characteristics were selected. The relationships between the level of governance disclosure and the chosen characteristics were examined in order to highlight those factors that are associated with and affect the level of governance disclosure.\nFindings: - On average about 50 percent of governance items required to be disclosed had been reported by AIM-listed companies. As for large companies, there is a positive relationship between the number of non-executive directors in governance mechanisms and the extent of disclosure. Considering that there has been a declining interest in AIM-listed companies, the presence of more non-executive directors is recommended, as this would ultimately mean more compliance with the governance disclosure requirements and could result in restored investor interest and confidence. The findings also indicate that SMEs are more likely to have adopted a stakeholder approach when concentrating on their governance arrangements.\nOriginality/value: - The study presents evidence on governance disclosure practices of SMEs and indicates that, in spite of a less stringent regulatory regime set up within AIM, non-executive directors, who bring in an element of independence to boards, play a significant role in elevating corporate transparency.\n",
"role": "assistant"
}
] |
Do my staff trust me? : The influence of organizational structure on subordinate perceptions of supervisor trustworthiness
|
[
"Trust",
"Organizational structure",
"Centralization",
"Personality",
"Formalization",
"Big five",
"Trustworthiness",
"Routinization",
"Span of control"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: For supervisors, having the trust of their subordinates is of crucial importance. Research has shown, for example, that subordinates who trust their supervisors have higher job satisfaction, higher job performance, and lower turnover intentions. In addition, they demonstrate greater citizenship behavior and less deviant behavior (Dirks and Ferrin, 2001, 2002). Davis et al. (2000) showed that when subordinates' trust in managers is widespread throughout an organization, overall sales and profits increase. Not surprisingly, therefore, most of the research on trust has focussed on subordinates' trust in their supervisors (Schoorman et al., 2007). Likewise, organizations are keen about promoting it in the workplace.\nTheoretical overview: trust and perceived trustworthiness: Mayer et al. (1995) developed a widely accepted model of trust in organizations. According to the model, trust refers to the willingness of a trustor (e.g. subordinate) to be vulnerable to the actions of a trustee (e.g. supervisor) based on the expectation that the trustee (i.e. supervisor) will perform a particular action important to the trustor (i.e. subordinate). The model stipulates that trust is directly influenced by two factors, namely, the propensity to trust of the trustor (i.e. subordinate) and the perceived trustworthiness of the trustee (i.e. supervisor).\nInfluence of organizational structure on subordinate perceptions of supervisor trustworthiness: Formalization\nMethodology: Design and respondents\nResults: Table I shows the descriptive statistics, intercorrelations, and reliabilities for the variables. H1, H5, H6, and H9 related to the main effects of organizational structure on subordinates' perceptions of their supervisors' trustworthiness. These hypotheses were tested with hierarchical multiple linear regression. The control variables gender, age, and propensity to trust were entered as the predictors in step one, the independent variables formalization, routinization, hierarchy of authority, participation in decision-making, and span of control were entered as the predictors in step two, and the dependent variable subordinate perceptions of supervisor trustworthiness was entered as the criterion. H2, H3, H4, H7, and H8 related the moderation and interaction of subordinates' personality traits. These hypotheses were tested with Hayes and Matthes' (2009) modeprobe SPSS macro which enables probing of single-degree-of-freedom interactions in ordinary least squares and logistic regression.\nDiscussion: Summary of findings\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the influence of organizational structure on the degree to which supervisors are perceived to be trustworthy by their subordinates. The manner in which subordinates' personality traits interact with and moderate these relationships is also examined.\nDesign/methodology/approach: A three-part survey was completed by 137 full-time workers from various industries.\nFindings: Formalization was positively related to subordinates' perceptions of their supervisors' trustworthiness and more strongly for subordinates who ranked higher on straightforwardness. Routinization was also positively related but only for subordinates who ranked low on excitement-seeking. Centralization in the form of hierarchy of authority was negatively related and more strongly for subordinates who ranked higher on competence. Routinization for subordinates who ranked high on excitement-seeking, centralization in the form of participation in decision-making, and span of control were not related.\nPractical implications: Before pursuing trust-building efforts, supervisors should consider whether and in what way they may already be perceived to be trustworthy by their subordinates based on the structure of their organization and the personality traits of their subordinates.\nOriginality/value: This study is the first to examine the main effects of organizational structure on subordinates' perceptions of their supervisors' trustworthiness.\n",
"role": "assistant"
}
] |
Application of agent technology for recycling-oriented product assessment
|
[
"Agent technology",
"Design for recycling",
"Recycling",
"Computers",
"Design"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Presently, the handling of products withdrawn from usage is one of the most important problems emerging in modern society. It concerns not only Europe and the USA, but the whole world (Zia et al., 2008; Chen et al., 2009; Bezzina and Dimech, 2011). It is a common conviction that one of the possibilities to solve this problem is taking into account the environmental issues in designing the product. The earlier the environmental aspects are identified, the greater the possibility to design an environmental-friendly product (ISO, 1995; Santos-Reyes and Lawlor-Wright, 2001; Lin et al., 2001; Sundin et al., 2012).Widely understood environmental issues manifest themselves by appearing and evolving the term \"design for environment\" (Lin et al., 2001). Within this term one can distinguish: design for remanufacturing, design for reusability, design for disassembly and design for disposal. The holistic approach to the product is also reflected in the term product lifecycle management (PLM), which is also a broad one. It can be perceived both as a management paradigm (Saaksvuori and Immonen, 2002) and as an IT solution, whose components are dedicated to deal with particular phases of product lifecycle (Hadaya and Marchildon, 2012). Although, in the authors' assessment, within the scope of functionality of PLM solutions available on the market, the particular phases of product lifecycle are not represented in an equal way. It mainly concerns final phases, especially the recycling phase. The solution presented in this paper is a step towards including environmental issues into design in a practical sense. Its essence consists in using agent technology in design process. The use of this type of system architecture is grounded in the complexity of modern products, both in design and material aspects. The research question of this paper is to determine the usefulness of agent technology for supporting the designer's work in obtaining products that are recycling friendly. An additional research goal is to define a new set of measures to assess the susceptibility of a product to be recycled.The remainder of this article is structured in six sections. Section 2 presents the measures used in quantitative assessment of products toward their susceptibility to recycling. Section 3 characterizes agent technology, taking recycling issues into account, and justifies usage of such an approach for the purpose of design. The structure and operation of the system is presented in Section 4. Section 5 presents the set of recycling-susceptibility measures developed by the authors. An example of assessment of a household appliance through use of the system is also shown in Section 6. The paper concludes with implications arisen from this study, including the limitations of the presented solution and further research directions.\n2 Methods and indices of recycling-oriented product assessment: Existing computer tools aiding a designer's actions in the area of eco-design allow selection of a method of recycling, eco balancing of products and materials and help defining paths of disassembly. They can be individual tools or integrated systems, joining together many modules. Most of these tools, aiding in the environmental assessment of products, are based on various methods, e.g. LCA - life cycle assessment, LCC - life cycle cost, LCS - life cycle sustainable, LCIA - life cycle approach (Guinee et al., 2001; Bouman et al., 2000; Weidema et al., 1999; Sadiq and Khan, 2006; Goralczyk and Kulczycka, 2005; Santos-Reyes and Lawlor-Wright, 2001; DeMendonca and Baxter, 2001; Dove, 2005) and eco-indices, which are created based on obligatory standards.The eco-indices have a form of single values, representing an influence of the product or material on the environment. For example, material input per unit of service (MIPS) determines the quantity of materials used for the product to be manufactured (Ritthoff and Liedtke, 2002). A MIT index is a complementary value, which determines a use of resources per the unit of material. It takes into consideration use of materials or resources and energy (MIPS, 2009). A level of recycling is a ratio of material mass which will be subjected to recycling and the total mass of the product. Producers of electrical and electronic hardware are obligated by law to achieve a defined value of the recycling level (WEEE, 2002; Act, 2005).Eco-indices are an attempt to model a wide range of influences on the environment. Each one represents different characteristics and requires individual interpretation. A selection of a specific eco-index or a set of indicators is dependent on the individual characteristics of the product, place and method of manufacturing, recipients, etc.One of the first attempts at creating such a multi-point model is the eco-point index, developed in 1990 by the Swiss Ministry of the Environment. It is based upon 14 various coefficients, which are weighed and aggregated. Although this method was in practical use only in Switzerland (due to the specific conditions for which it was determined), creators of other methods were using it as an example (Lewis et al., 2001). Nowadays, one of the most frequently used eco-indices is the Ecoindex 99, which is a modification of the previously used Ecoindex 95. It was calculated by PRe Consultants on the order of The Netherlands' Government, in partnership with different types of manufacturing companies and research and development units (Lewis et al., 2001).For the ecological aided design system with special regard to recycling (described in this paper), an index that defines the level of product adaptation for disassembly was created. It is a sum of three separate coefficients: a level of recycling (calculates product compliance to recycling and compares compatibility with legal requirements), a variety of connections (calculates a level of complexity of the disassembly process and a quantity of one of the disassembly cost components) and a material variety coefficient (calculates a level of complexity of a material sorting operation and compatible materials selection).Ecological indexes and other methods of assessment of a product's influence on the environment are focused on the selected aspects of interaction with the environment. Determination of the index value provides information of a current state and allows evaluation and identification of desired directions of further actions. A process of design is an iterative process and values of particular indices change after each stage. Current tracking of their values provides valuable information of the results of actions taken but is considerably work consuming. Therefore, it is necessary to use computer-based techniques for this task.\n3 System for aiding the pro-ecological design: 3.1 Agent technology\n4 Description of the system: 4.1 Functional assumptions of the system\n5 Recycling assessment of the product: 5.1 Calculation of recycling level and index\n6 An example of application: An exemplary recycling analysis of a product using the agent system was performed for a small appliance - an electric kettle of typical design (ZEL, 2010). In Figure 6, a scheme of the kettle construction is presented.Before implementing the data in the system to aid the pro-ecological design, it was necessary to transform the standard structure of the product into the recycling structure. Lists were expanded with masses of particular elements and materials and types of joints were defined. Table IV contains the recycling parts list of the product; Table V presents a representation of connections in the product structure.After entering data into the agent system, recycling analysis of the device was conducted. Recycling level (PR) was 68.61 percent and WPR=4.0 (Table III); WP=1.5, WM=1.5. Thus, overall recycling index (WR) amounted to 7.0 (Figure 7).The conducted analysis indicated the elements in the design, that influences on decreasing recycling level to the greatest extent (these elements are underlined within the structure of the product - Figure 7).According to the advices from the agents, the changes in the analyzed device were made. It was changed the way of joint between parts in upper controller. The renewed recycling analysis showed the increasing of PR to 100 percent, and WR decreased to 3.5. It expresses the increasing recycling susceptibility of the product (Figure 8).\n7 Conclusions: The described agent system can be used by designers as a complementary system to a CAD system, expanding its possibilities with aspects related with later disassembly and recycling of the product. The system informs a designer of current values of the recycling level and index and about places in the product that have the highest negative impact on the overall product assessment. It gives a solid foundation to determine further direction of work over the product design.In the present version, the system is incompatible with existing CAD systems. Compatibility should manifest itself as a mutual exchange of data about the product structure.The main problem here is an interpretation of connections between structure elements. In CAD systems, a structure and geometrical features of elements are defined. Joints between parts are not directly defined, as they result from the spatial arrangement of the elements. However, from the point of view of susceptibility to recycling, an open definition of joints and their attributes is necessary. The system creators had in mind a potential usefulness of such a function of the system and a solution has been applied to allow further expansion of the system in this direction.Simultaneous cooperation of the system with many databases on different servers is possible. Adding servicing of another database server requires programming and compiling new versions of the designer interface and the information agent. Access to the database in the designer interface is realized through the ADO interface and in the information agent through the ODBC interface; both interfaces are independent of the database. As such, their new versions will be different only regarding the configuration parameters of these interfaces - assuming that the database structure will not change.\n",
"role": "user"
},
{
"content": "Purpose: - The paper aims to describe ideas and implementation of the computer tool for computer-aided and recycling-oriented design. Currently, there is a strong tendency to take into account the impact of a product on the natural environment. The authors concentrated on the issue of the recycling process of the product, taking into account the phase of its design. The purpose, structure, technology and example results are presented in this paper.\nDesign/methodology/approach: - It was assumed that analysis will be performed based on a set of measures of a product. These measures (developed by the authors), describing the product from the recycling point of view, are calculated automatically, according to the changes (variants) in the product's model. The presented tool is based on agent technology. The structure of the system - the agents, its roles and communication between them - has been described.\nFindings: - The main achievement of the work presented in this paper is the method supporting eco-design based on agent technology. Based on the analysis of the process of recycling-oriented eco-design, the authors designed and then implemented the tool that aids designers' activities in the area of eco-design. The drawn-up method supports decision making concerning designing environmental-friendly products. Thanks to agent technology, the design process can be conducted in the distributed design environment.\nResearch limitations/implications: - The usage of the presented computer-aided and recycling-oriented system during the design process requires the duplication of some of the engineer's work. The analysis is performed within the presented tool, outside the typical contemporary design environment, such as CAD 3D systems. As a consequence, there is a need to enter the product structure into the described system, which can be inconvenient. The further work of the authors assumes implementation of this concept into a CAD 3D system.\nOriginality/value: - There has been a limited amount of research work regarding application agent technology in the field of end-of-life oriented design. The structure of the system, the measurements and idea of extension of product structure are the original results of the work. The results could be implemented into commercial computer-aided design systems, especially into its PLM (product lifecycle management) group, due to the weak representation of end-of-life phases in these tools.\n",
"role": "assistant"
}
] |
Emotional capacity in the public sector - an Australian review
|
[
"Competencies",
"Emotional intelligence",
"Organizational change",
"Public sector reform",
"Emotion",
"Emotional capacity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The essence of high performing leaders lies in the core emotional preparedness to search for the common emotional ground with courage and commitment toward connectedness(Leithwood and Beatty in Blackmore, 2011).\nIntroduction: Capability frameworks are vehicles used to enhance leadership competencies, promote public sector competencies and effect cultural change (Bhatta, 2001; Hondeghem and Vandermeulen, 2000; Horton, 2000; Podger, 2004). Public sectors across many nations have adopted competency development as a key strategy of public sector reform. The NSW Public Service Commission (PSC), in Australia, developed a capability framework implemented across all agencies, to drive the cultural change necessary to ensure its workforce effectively adapts to the changing future context.\nMethodology: A literature review provides the case highlighting emotional capacity as a fundamental capability in achieving organisational change. The scene of the reform context is established informed by a synthesis of literature retrieved through search terms including public sector, public administration, organisational culture, organisational change, change and emotion. A specific search was conducted to explore EQ as the concept defining emotional capacity with particular focus in the public sector. A ProQuest search using keywords:\nLiterature review: The case for emotion at the heart of public sector change\nResults and discussion: A review of the 490 behavioural indicators described in the framework revealed that only six explicitly indicate emotional components in three of the 20 competencies. Emotional capacity is not explicitly evident in the framework as a backbone capability that drives the effectiveness of these competencies. The review highlights a focus on the technical and managerial areas, reinforcing the \"what\" rather than the \"how\". This is possibly a legacy of public sector management still encased in traditional paradigms. Rosete and Ciarrochi (2005) examining the Australian public service noted that leaders can have the \"what\" but lack the \"how\" skills. The how skills are significantly important in interpersonal interactions.\nConclusion: The focus of the PSC to build capability, improve performance management and enhance recruitment and selection processes is commended. At the heart of managing this change is influence and leadership, present and active across all levels of the organisation, formally and informally. Managerial and technical ability is not enough (Dreyfus, 2008), as leaders need to successfully re-negotiate the psychological contract with their colleagues to engage and obtain commitment from the workforce to collectively arrive at the redefined state.\n",
"role": "user"
},
{
"content": "Purpose: This paper presents a review examining an Australian public sector competency framework through the lens of emotional intelligence (EQ) to answer the question \"To what extent is the concept of EQ used to facilitate NSW public sector reform?\" The purpose of this paper is to accentuate the importance of emotional capacity as an important capability to achieve reform goals, recognising the public sector's deep organisational history and accepting that change is an emotional event, and that people achieve change.\nDesign/methodology/approach: A literature review drawing relationships between culture, change and emotion is applied to a capability framework for the public sector in the State of NSW. This review serves two purposes. First, it examines interacting factors that define the public sector context - a culture developed over generations, identity, the impact of culture on change and the relationship between change and emotions. The second examines a concept for its ability to transform this culture in a comparatively short time compared to its evolutionary history. Emotional capacity is framed by the EQ literature and is explored as a competency with particular focus in the NSW public sector. A ProQuest search using keywords Emotional intelligence and Public Sector or Civil Service; and Emotional intelligence and Public Administration located 22 studies across 14 countries looking at EQ in the public sector. These are supplemented by additional studies on EQ. The capability framework is examined against the elements of the only recognised standardised test for EQ (Fiori and Antonakis, 2011), the Mayer-Salovey-Caruso Emotional Intelligence Test.\nFindings: The examination concludes that emotional capacity is implicit, if not overlooked within the framework, with continued emphasis on technical and managerial competencies, evident of public sector management still encased in traditional paradigms. The discussion positions the development of emotional capacity as a high-order competency in a challenging reform environment.\nResearch limitations/implications: The literature review may suffer from publication bias in both the literature cited in this review as well as those studies that have been published, particularly given the small amount of studies available within the public sector environment. The theoretical nature of the matching assessment is subjective and allows potential for variation in interpretation in both the meaning of the competencies and the matching to the four branches of EQ.\nPractical implications: Empirical research in EQ is limited in the public sector domain. The public sector has an embedded culture weighed with assumptions steeped in history. A public sector organisation is valuable for longitudinal studies as many employees stay for considerable years if not their whole career. Further empirical research within this sector in examining the impact of emotional capacity on cultural reform would enhance the knowledge in this field.\nOriginality/value: The paper contributes to the limited literature examining the optimal competencies in particular emotional capacity for reform in the public sector.\n",
"role": "assistant"
}
] |
Office building efficiency and capacity benchmarks
|
[
"Buildings",
"Resource efficiency",
"Real estate",
"Benchmarking",
"Workplace"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Many organisations increasingly view corporate real estate management (CREM) as a vital part of their business. This perception may be based on the handling of real estate by CREM, which according to Avis and Gibson (1995) is the second or third largest cost factor for most companies (after staff costs). On the other hand, the support of the core business is considered to be the primary task of corporate real estate (CRE). (Markides, 1999) For further information about CREM see Bon (1992, 1994a, b, c).In the service sector, which is examined more closely as part of this paper, office buildings are, as a rule, a key production factor. The CREM of service companies and its sub-segment, premises management, therefore focus on office workspace. As part of their premises management, companies in the German-speaking region, for example, are increasingly operating with benchmarks (see International Facility Management Association, 2001) which are used to determine whether CREM has achieved its performance targets within the company and also in comparison with competitors.These benchmarks are used as core indicators, e.g. m2 of usable floor area per existing workplace and m2 of office space per existing workplace (for definitions see Glossary), which must be achieved through CREM activities (by way of \"management by objectives\", for example). Such core indicators or benchmarks are agreed or set as objectives. However, when deciding in favour of one or more of these benchmarks, the question arises as to the ranges of realistic benchmarks for a particular property or portfolio.As a rule, the literature on benchmarking can be broken down into two categories (Garnett and Pickrell, 2000):1. general benchmarking guidance; and2. construction-specific commentaries.The first group largely comprises publications from the 1990s, when the use of benchmarking increased (Camp, 1989, 1995; Codling, 1992; Bendell et al., 1998). These publications concentrate on presenting the fundamentals of benchmarking and, above all, the benchmarking process. The second group focuses on the specific adjustments to construction and real estate management, mostly in connection with case studies. By way of a representative example for the German-speaking region, we highlight the studies by Jones Lang LaSalle (2003) and the Baukosteninformationszentrum (2003) (BKI) as part of the present paper.While the former organisation focuses primarily on cost aspects, the latter also addresses the space benchmarks examined more closely as part of the present paper. However, our study goes beyond the benchmarks to also include the main drivers of benchmarks. This enables us to determine benchmarks that are characterised by lower spreads and therefore more closely approximate the property to be assessed. In addition, the present study examines benchmarks outside the scope of standardised definitions and also includes workplace benchmarks and their drivers. The above-mentioned studies form the basis for these examinations, as they provide both survey methods and standardised space benchmarks.This project is based on an empirical study at the Swiss Federal Institute of Technology Zurich. As a first step, the floor data of 116 owner-operated office buildings was collected. As a second step, the drivers of the two relevant building efficiency benchmarks (namely m2 usable floor area per m2 gross external floor area and m2 office space per m2 gross external floor area) were ascertained using regression analyses, and their benchmarks were determined. In addition, as a last step, capacity benchmarks and their spreads were ascertained using the same data. These are the above-mentioned benchmarks for m2 of usable floor area per existing workplace and m2 of office space per existing workplace, which are the principal core indicators of CREM.All spaces surveyed are based on DIN 277. This standard was selected because it is widely used both in Switzerland and in Germany. By comparison with other European standards (e.g. RICS), DIN 277 is characterised by a higher degree of detail. This standard serves exclusively for the definition of space. For this purpose it is divided into three parts, with each part covering an increasing degree of detail:* DIN 277-1: Grundflachen und Rauminhalte von Bauwerken im Hochbau: Begriffe, Berechnungsgrundlagen (Plan areas and volumes in building: terminology, bases of calculation) (Deutsches Institut fur Normung, 1987a);* DIN 277-2: Grundflachen und Rauminhalte von Bauwerken im Hochbau: Gliederung der Nutzflachen, Funktionsflachen und Verkehrsflachen (Netto-Grundflache) (Plan areas and volumes in building: classification of utilisation areas, operating areas and circulation areas (net ground area)) (Deutsches Institut fur Normung, 1987b); and* DIN 277-3: Grundflachen und Rauminhalte von Bauwerken im Hochbau: Mengen und Bezugseinheiten (Plan areas and volumes in building: quantities and reference units) (Deutsches Institut fur Normung, 1998).The structure of the present paper is based on this approach. In the first part, the authors discuss the usage of core indicators in CREM, while presenting the empirical study and its results in greater detail in the second part. Both the building efficiency and the capacity benchmarks must be considered. In the final part of this paper, the results achieved are re-examined and further aspects are identified which must be taken into consideration when working with core indicators.\nTheoretical framework: Goal and result-oriented CREM\nEmpirical study: Design of the empirical study\nDiscussion and outlook: As the present study shows, it is feasible to determine building efficiency benchmarks. Indeed, we were able to provide such benchmarks for office buildings. However, no conclusions may be drawn from building efficiency to capacity benchmarks. Among other things, we established that the two different benchmark types are driven by differing factors. The capacity benchmarks outlined in this paper are therefore not specifically associated with any specific drivers, as we have not been able to ascertain any such drivers during this study. This may be the starting point for a further exploration, including another survey. The basis for a further examination would be the building efficiency benchmarks and their drivers, as an effective building efficiency is the prerequisite for a high level of building capacity.When working with building efficiency and capacity benchmarks, it should always be taken into consideration that there are also other metrics that may be used as core indicators. One of the most common indicators mentioned in this context is staff satisfaction. In addition, the shareholder value aspect, for example, must not be ignored in a CREM context.Staff satisfaction can be translated into core indicators and thus used when measuring the performance of CREM. Using practical examples, Granath (1999) examined the different implications: \"workplace making potentially has a greater importance to companies than just housing their business, communicating an image of the company, being a sound investment or even being an appropriate physical work environment. All these are traditional values associated with a good workplace, but suggesting workplace making as a tool beyond this is a way to enhanced organisational behaviour and learning\" (Granath, 1999). Pinder (2003) examined the criteria that influence staff satisfaction on the basis of interviews with 1,800 users. He identified four factor categories, containing 22 drivers, which relate solely to the building:* space configuration of the building;* environment of the building;* appearance of the building; and* functionality of the building.Against this background, the function of CREM is to provide quality workplaces. This function cannot be fulfilled independent of building efficiency and capacity. However, the availability of a quality workplace is also determined by other factors (Pinder, 2003).The growing trend to include shareholder value aspects when assessing the capabilities of CREM is based on the value and performance of CRE. According to Krumm and Linneman (2001), CRE accounts for 39 percent of the total company value. Hence, CRE ties up an enormous amount of capital and yet its performance is comparatively low. According to Krumm and Linneman (2001), this is reflected by the following figures: \"the FTSE Euro top 300 had an average return on equity of 38.3 percent over the last 2 years, the GPR General European Property index only gained a 3.8 percent return over the same period\". Against this background and on the basis of an empirical study, Deng and Gyourko (n.d.) conclude that a low level of CRE has a positive effect on shareholder value. It is therefore the task of CREM to reduce available CRE to the lowest possible level. To achieve this, however, both the building efficiency and above all the building capacity must be positively influenced by CREM.Accordingly, both building efficiency and capacity are key elements of the CREM capabilities and determinants of staff satisfaction and shareholder value of the entire company.\n",
"role": "user"
},
{
"content": "Purpose: - As corporate real estate is increasingly being recognised as an important cost and production factor, senior management now pays considerable attention to this asset group. These assets are managed, inter alia, by using benchmarks. In addition to monetary benchmarks, building efficiency (e.g. m2 of usable floor area per m2 of gross external floor area) and capacity benchmarks in particular (e.g. m2 of usable floor area per existing workplace) must be highlighted. Previously, specific values and their drivers for the assessment of buildings or portfolios were not available.\nDesign/methodology/approach: - This study is based on a survey carried out in Switzerland, involving the collection of floor data (in accordance with DIN 277) of 116 owner-operated office buildings. In addition, their drivers were ascertained using regression analyses.\nFindings: - The building efficiency may be assessed on the basis of the share of usable floor area or the share of office space in the gross external floor area. The main drivers are the shares of vehicle parking space, areas for residential and recreational purposes and areas for storage, distribution and retail. These drivers must be taken into consideration when assessing the building efficiency. It became apparent that capacity benchmarks are determined primarily by factors such as the space use management strategies, and only to a lesser degree by the building itself.\nOriginality/value: - The study provides space benchmarks and their drivers. The results therefore permit an objective evaluation of office buildings. However, further work transcending the influence of the building itself will be required with regard to the capacity benchmarks.\n",
"role": "assistant"
}
] |
Employee ownership + accountability=competitive advantage: Precor drives organizational change
|
[
"Management accountability",
"Employee accountability",
"Organizational culture",
"Employee involvement"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Precor has a history of entrepreneurial success, but we weren't satisfied with staying in place. Our ability to step up to the next level, to achieve our objectives and potential, required that we transform our culture to focus on a results-oriented, high performance culture. We needed to put in place a structure that articulates what we were trying to achieve - to put the entire team on the same page, define our culture and enroll our employees to be part of that (Paul Byrne, Precor president).\nMoving from \"small to big\": A $200 million US-based manufacturer and marketer of premium fitness equipment, Precor celebrates its 25th anniversary in 2005. Operating under several ownership groups since its inception, Precor in late 2002 became part of Helsinki-based holding company Amer Sports. Today, Precor employs 750 people in subsidiaries or reporting companies in the USA, Germany, Spain, Japan, Benelux, China and the UK.The Precor story is one of growth, change and transformation of a company that originated in a garage, and is entering its early adult years.From the outside, the Precor story appears to be one of progressive evolution and development. What's less apparent is that as Precor entered its third decade with big business goals, the company in many ways was operating as a small business.Externally there was evidence of great change, including a stair-stepping series of line expansions and market growth in the 1980s and 1990s. Further, a major initiative in the late 1990s by Precor's former parent, Illinois Tool Works (ITW), transformed Precor manufacturing into a world-class lean manufacturing operation.A global leader in lean manufacturing, ITW pared Precor manufacturing to core activities through outsourcing, workforce reduction, and investment in production technology and processes.Entering 2000, Precor had finely tuned its production operations, and generated an unprecedented string of double-digit increases in annual revenues and profitability. New leadership also began outlining a vision for Precor that extended beyond traditional fitness equipment, to a \"total product\" offering of products and services leveraging technology.\nRecognizing the need for change: As Precor senior management assessed resources to achieve the aggressive \"total product\" vision, the divide between Precor's world-class assembly-test-ship operations, and a surrounding \"culture of complacency\" clinging to small-company ways became more defined. Despite world-class operations practices, the Precor culture was characterized by a focus on effort, not results, and lack of rigor and process.Management recognized that to achieve its \"total product\" vision, it was necessary that Precor displace the reigning culture through training to establish new structures and processes.Fortunately, as a mid-sized company with few enterprise-wide processes or initiatives that impede change, corporate change at Precor could be achieved relatively rapidly. Also, there were no major functional or organizational divisions to be bridged between Precor human resources, and training and development groups. With few institutional impediments, programs that fostered employee performance could be relatively easy to integrate, and systems that rewarded and recognized recruiting, hiring and training could be tied together seamlessly.\nFocusing on people, processes and products: Precor operations had established and mastered core competencies \"assemble, test and ship.\" Now Precor management applied an organization-wide focus to people, processes and products, beginning with an initiative to build accountability and ownership within every Precor employee.While the concept of individual and organizational accountability seems basic, Precor management was surprised to discover that few employees were aware of specific results or goals the company had committed to deliver.Further, even those who did recognize general objectives of revenue, profit and service were unable to articulate how their individual role and performance related to corporate results in any way.It's a common issue, similar to the way the public approaches weight loss, one of the fitness industry's key objectives. Countless consumers grasp the idea that they want to lose weight, but one of the fitness industry's other primary challenges is instilling in consumers the discipline and behaviors necessary to lose weight.Similarly, it was essential that Precor formalize a framework of behaviors and processes that drive personal accountability, and establish the direct connection between an individual's performance and Precor's success in accomplishing its corporate objectives.\nPeople: setting a corner-stone of accountability: it all started with senior management: During the fall of 2001, Precor's executive team made a commitment to a cultural change and alignment needed at Precor to achieve our growth plans.To crystallize the connection between employee performance and organizational success, Precor introduced a cornerstone in the system and structure of the OZ Principle(r) - Getting Results Through Individual and Organizational Accountability, a training program provided by Partners in Leadership.The executive team began by strategizing content and principles in leading and sustaining an accountability initiative. As a beginning, Precor's management team graded the company on the four main behavioral areas that define accountability - \"See It, Own It, Solve It, and Do It,\" and found that we averaged a \"C to D minus.\" Nor surprisingly, the Precor team committed to improving the grades.The first action step was to identify the big picture, results to focus on, and the language and actions to create the experiences that would generate the beliefs desired in Precor employees of the future - basically, the methodology for achieving results.Actually, Precor's team had good cause for excuses that would justify less-than-optimal performance at this time. ITW, Precor's parent company, had placed Precor on the sales block, thrusting the company into the unknown. Also, Precor's single largest customer had filed for bankruptcy, resulting in a scramble to rebuild the Precor dealer network - as well as a multi-million dollar write-off against the bottom line.However, management held firm to the beliefs of a no-excuses approach, and ultimate accountability. Senior management continued to stress continuous improvement in Precor products and services, and building the tradition of quality, innovation, and customer focus through great people.Management's attitude was exemplified by Precor president Paul Byrne's introductory remarks at Precor's January 2003 all-employee kickoff meeting: \"The world's a mess, the economy is slow and the weather is bad. We are simply not going to accept any of those old excuses anymore\".\nTraining reinforces strategy: Training was rolled out to cross-functional groups in all technical professional positions, followed by programs tailored for field sales, and international offices.Oz Principle(r) models of \"See It, Own It, Solve It, Do It\" - the Steps to Accountability models - were introduced adopted as the definition and measurement of success.In a very short time, a common language was established, and employees could identify when they were being accountable, or staying \"Above the Line,\" and when they fell \"Below the Line,\" with behaviors that can lead to a spiral of deferral, denial, blame and turf-guarding.In addition to building the case for change and creating a culture of accountability, the Precor senior management team developed strategies for content and principles to extend and sustain changes in the organization.Systems to improve employee communications, recognize ownership for results, and support and reinforce accountability throughout the company were created, or revised to align with the new framework of organizational values.Additionally, building new skill sets took on a major focus, beginning with a company-wide training program on basic business finance and strategies.Through a simulation/game-playing training program created by Paradigm Learning, teams of employees tried their hand at buying and running a company.In training sessions in which Precor employees manufactured and delivered products, they also confronted a variety of business challenges presented by suppliers, customers, banks and customers. They also experienced every aspect of the financial process, from collecting receivables and paying bills to making decisions that illustrated business principles such as cash flow, return on equity and cost of good sold.Though this role-playing process, Precor dramatically increased employee understanding from a CFO's perspective, featuring skill sets valued by management, and reinforcing how each individual's performance impacts the bottom line.\nProcess: attacking a \"casual culture\": In addition to processes that build new skill sets throughout the enterprise, Precor also was determined to establish process practices and supporting systems to maintain Precor momentum in personal accountability, financial awareness and project management.Among the most important processes was that each Precor division developed and institutionalized clear, performance-based success factors, then established structured reviews assessing individual performance against specific performance metrics. In some cases, performance review formats conflicted with Precor's intent to focus on results, not effort.For example, in areas such as teamwork and communication, the review format was altered from focusing on whether an employee possessed a skill, to how well the skill was used to achieve goals, with 80 percent of a review section's values assigned to goals and goal achievement, not capability.A similar performance-management process was developed for work-group performance, with well-defined and visible performance targets reviewed monthly.To reinforce the new structure, Precor established internal communications that link the employee to the bottom line, including monthly Precor's senior management team reports to the entire company, highlighting Precor performance in key revenue, operations, customer satisfaction and profitability metrics. Also, on a quarterly basis, Precor's senior management team sits down with the entire employee base in open meetings in which no business-related subject or question is taboo.Additionally, monthly \"Dials on the Dashboard\" e-mails and handouts to every Precor employee recap the company's performance. As a division of a publicly-held company, Precor can't present monthly financial performance in advance of quarterly financial reports. Even so, by communicating performance as percentages above or below plan, \"the Dials\" provide a performance score card that all can understand.Eagerly anticipated by everyone from production line workers to executive managers, \"the Dials\" portray Precor performance against plan in sales by division, customer satisfaction, safety, inventory management, on-time delivery, and most importantly, profitability.Accompanied by a narrative analysis written by Precor president Paul Byrne, \"the Dials\" are closely followed within Precor - with its tradition of aggressive profit sharing, Precor employees have become extremely attuned to nuances revealed in Dials reporting.A similar structure was developed for work-group performance, with well-defined and visible performance targets reviewed monthly.Additionally, Precor continues to focus on building financial skills and individual awareness of one's impact on the bottom line.\nA firm hand in \"fixing what isn't broken\": An underlying issue increasingly became evident as Precor change initiatives were implemented: most Precor employees understood that Precor was by no means \"broken\" prior to the change initiative. In fact, throughout its history, Precor had been successful - especially so over the last decade.While organizational change advanced, it was inevitable that in some cases change would encounter pockets of denial or resistance - particularly in areas that had experienced the greatest success.At Precor, senior management applied a firm hand to implement change.Through direct communications and fair yet uncompromising personnel decisions, Precor senior management demonstrated that change was mandatory not only for Precor's success, but the success of individual Precor employees.Management also came to recognize that good training is not a management perq, and professional training was essential to universal acceptance of change - although management often holds that professional training is appropriate for mid-range managers and above, while investments in training the rank-and-file often are considered a tough decision.Last, Precor was firm in its commitment to set the roots of the culture of accountability deeply through systems that support, manage and maintain change in the organization, ranging from publicly rewarding positive behaviors, to fostering results-based personal performance reviews, and reinforcing the connection between individual performance and organizational results.\nThe real story: the results: The results are impressive, including institutionalized and in-depth delivery of company-wide communications, performance management processes, and reward-and-recognition programs. Individual, departmental and company goals are well defined and aligned with corporate objectives. Precor has transformed from a casual culture to a \"no excuses\" culture of personal ownership and accountability.Prior to the initiative, individuals within Precor collectively self-graded the company at a C- to D+ level on key accountability principles of \"See it, Own it, Solve it, Do it.\" A little more than a year later, the self-rated score was B+.Perhaps most importantly, the benefits are visible in the bottom line: in the three years since Precor management committed to organizational change, Precor revenues have increased more than 50 percent, with profitability up more than 200 percent.\n",
"role": "user"
},
{
"content": "Purpose: - It is well recognized that fast-growing companies can outpace their systems and infrastructure. Similarly, rapidly growing companies can outgrow their corporate culture. This case study aims to address training that replaces an established yet outgrown corporate culture with a dynamic culture suitable for the growing organization.\nDesign/methodology/approach: - Global fitness equipment leader Precor has achieved cultural change through a strategy founded in increasing - and rewarding - personal accountability for results. Through training, Precor established direct connections between each employee and the corporation's financial performance. A system of ongoing communications and rewards supports the change over time. The process included institutionalized and depth delivery of company-wide communications, performance management processes, and reward-and-recognition programs.\nFindings: - Through training, an organization can successfully transition individuals and departments to embrace a culture of personal ownership, accountability and reward. Financial benefits also accrue: since Precor committed to organizational change, corporate revenues have increased by more than 50 percent, and profitability is up more than 200 percent.\nPractical implications: - Management can institutionalize a culture of personal ownership and accountability that generates superior financial results.\nOriginality/value: - Human resources and training executives interested in significantly increasing their organization's business results can gain insight into building employee accountability and individual links and ownership of their company's results.\n",
"role": "assistant"
}
] |
Cultivating entrepreneurial behaviour: entrepreneurship education in secondary schools
|
[
"Engagement",
"Education",
"Entrepreneurship",
"Student retention",
"Entrepreneurial behaviour",
"Teacher facilitation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Entrepreneurship Education (EE) is not a new concept. From the early 1980s in New Zealand and throughout the world, Western governments recognized that an entrepreneurial orientation may lead to economic growth, job creation, international competitiveness and technological advancement (Jack and Anderson, 1999; Audretsch et al., 2002; Ladzani and van Vuuren, 2002; Grebel et al., 2003, Vetrivel, 2010). Fayolle et al. (2006, p. 701) have suggested that EE is a way of encouraging economic development. EE is therefore primarily aimed at facilitating economic growth and increasing the pool of entrepreneurial talent within economies. Entrepreneurship is perceived to be a major source of innovation, job creation and growth and is recognized as a measure of a country's economic health (Audretsch and Thurik, 2000 and 2001; Audretsch et al., 2002). The way a country has established its educational system can also help lead people to develop qualities that are considered important for entrepreneurship (Reynolds et al., 1999).\nEntrepreneurship education in context: New Zealanders have a strong popular identification with the ideals of entrepreneurship. On at least two occasions, the Global Entrepreneurship Monitor (Reynolds et al., 2001, 2004) has placed New Zealand highly in total entrepreneurial activity. New Zealanders are told and encouraged through the popular press to buy into the paradigm. For example, the book by John et al. (1998), later made into a television series, celebrates the popular notion of the innovative nature that resides in the average New Zealander. The New Zealand School Curriculum for example has as one of its vision statements that students will be \"Enterprising and entrepreneurial\" (MOE, 2007, p. 8).\nResearch design: The broad thrust of this research was exploratory and based on qualitative techniques to uncover the underlying themes and concepts associated with EE at secondary school level. The theoretical position of this study is one of applied social research which is most commonly used to solve problems and explore specific phenomena within a social context. The findings of this study would ultimately be useful in compiling recommendations for educational institutions and may be used to further develop educational policy.\nCommon themes and concepts: One of the most striking features encountered in this study with schools who have almost fully integrated EE into their curriculum delivery, was the level of \"energy\" displayed by students and teachers alike. Students, teachers and leaders in these schools display an enthusiasm and focused commitment that one rarely encounters in a large majority of educational institutions. Students are focused, concentrating on the task and virtually oblivious to outside interference. Teachers act as coaches and mentors, facilitating the learning process with students and demonstrating a passion for their subjects. Principals push everything else aside and freely enthuse about the refreshing changes and positive results they are gaining from an EE approach.\nConclusions: The major objective in this study was to establish whether an EE approach was effective in bringing about changes to teaching practice in secondary schools that resulted in improved student outcomes, behavior and attitude. The findings from the spread of schools researched lead to the general conclusion that the move to an entrepreneurial approach to secondary education is a mid- to long-term strategy that generates significant positive changes in student behavior and attitude. It was found that schools could be plotted on a continuum of \"entrepreneurship\" with some schools further advanced due to the degree to which entrepreneurship had been embedded in the value structure and strategy of the school. There was no evidence that any one school had achieved full integration of the entrepreneurship concept across all subject areas in the secondary curriculum. This highlights the incremental nature of the adaptation process and suggests that it may take several years to achieve a high degree of entrepreneurial integration based on current definitions of the concept and the cultural differences encountered in each school.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper was to investigate the degree to which Entrepreneurship Education (EE) was being provided to secondary school students following changes to the Secondary School Curriculum in 2010 by the New Zealand Ministry of Education. Under these changes, secondary schools were charged with following an \"entrepreneurial\" approach to school instruction that would develop entrepreneurial behaviors in students.\nDesign/methodology/approach: The study used a qualitative methodology focused on gauging the reaction by teachers, students and their parents to this new teaching approach. The sample comprised ten secondary schools situated in Northland, New Zealand. A series of focus groups were used to solicit data among three levels under study in each school, i.e. teachers, students and parents. Individual semi-structured interviews were used to collect data from school principals to determine overall reactions to EE by the rest of the school.\nFindings: Teachers reported benefits in terms of reduced direct teaching workload, increased participation from students and significantly improved scholastic results compared to targets set in the curriculum. Students reported positively on the greater degree of flexibility allowed under this teaching approach, while parents reported changes in attitude and more engagement in school activities and projects.\nResearch implications: The continuing evolution of classroom education at secondary school level has long-term repercussions for student learning, engagement and retention as we move to the digital age. Similarly, there are also consequences for the evolving role of teaching, curriculum design and delivery.\nOriginality/value: The value of this research lies in a closer examination of the effects traditional teaching practices have had on secondary students entering the digital age. Furthermore, it investigates an alternative teaching approach through EE and the impact it has on student learning, retention and engagement.\n",
"role": "assistant"
}
] |
On the measurement and benchmarking of research impact among active logistics scholars
|
[
"Citation",
"Bibliometrics",
"Logistics scholarship",
"Research impact"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Knowledge creation, an activity that has traditionally been espoused within academia, is becoming ever more important at schools of business. Several programs that earlier may have considered research as a secondary activity now show an increased focus on research (Bennis and O'Toole, 2005). In addition, programs that already had a research inclination have increased that focus even more (Holsapple, 2008). This is true for most business disciplines, including logistics. For example, Carter et al. (2005) demonstrated an increase in the overall number of institutions that are showing an interest in publishing logistics research in the top journals. Given this enhanced focus on research and the generally high importance that research plays in a scholar's career (Ford et al., 2001), there is an interest in measuring research output and impact (Coleman et al., 2012; Chapman and Ellinger, 2009).\nLiterature review: The mathematical and statistical analysis of scholarly records usually falls in the field of study called bibliometrics, which consists of research methods usually grouped as \"citation\" or \"co-citation\" analyses (Pritchard, 1969). Citation analysis is based on the direct counts of references made to or received from other documents, while co-citation analysis exploits paired citations as a measure of association between documents or sets of documents. Citation analysis is frequently used to evaluate or compare papers, journals, academic programs, or institutions (Charvet et al., 2008; Gupta, 1997; Cronin, 1981; Reinsch and Lewis, 1993). What follows is a discussion of some of the relevant studies on citation analysis in the logistics literature.\nResearch method: Data collection\nData analysis and results: Mingers (2008) demonstrates that in the management field, cumulative citations of a paper usually follow an S-shaped curve, starting off slow and staying that way till about the fifth year after publication. A stable citation pattern is achieved in the beginning of the sixth publication year, after which the citation pattern remains similarly high for another 20 years. Finally, citations begin to taper off around year 25, when the work becomes obsolete. It can thus be seen that the citation pattern appears to progresses in five-year increments, with the first increment lower than the remaining ones (until the onset of the fifth increment around the 25th year).\nDiscussion and implications: A comparison of the metrics across the various cohorts revealed some interesting insights. First, as would be expected, the total citations kept increasing as academic age increased (Table III). This is not a surprising finding since citations can in no circumstance decrease over time. And as scholars age, they will usually see an increase in the citations of their body of work. Thus, such a measure of research impact puts ECRs at a considerable disadvantage as compared to more seasoned researchers. Thus, it appears that the raw measure of total citations is not an appropriate measure to compare the impact of scholars of various academic ages. The same can be said of the h-index as well (Hirsch, 2005).\nLimitations and future research: The conclusions of this study, while valuable, are subject to certain caveats. First, as we have observed, the h-rate is only useful in evaluating logistics scholars who have an academic age of more than five years. Thus, it may have limited applicability for evaluating the research impact of extremely young scholars.\n",
"role": "user"
},
{
"content": "Purpose: Scholarly interest in carrying out impactful research continues to remain high. Yet, given that citations of scholarly work can never decrease with time, traditional measures of research impact (such as raw counts of citations) unwittingly discriminate against early career researchers and also make it hard to identify future high impact scholars. In the current study, the paper compares several commonly used measures of research impact to identify one that best normalizes for the effect of career stage. The measure thus applies equally across most career stages, providing a usable impact benchmark for logistics scholars irrespective of seniority level. The paper also aims to present benchmarks on that metric to help logistics scholars identify their research impact vis-a-vis their peers.\nDesign/methodology/approach: Bibliometric data on the research of 702 logistics scholars were collected and analyzed by dividing the scholars into different cohorts based on seniority. Comparisons of different citation metrics were then made.\nFindings: The h-rate provides the most appropriate basis for comparing research impact across logistics scholars of various career stages. Benchmark h-rates are provided for scholars to identify their research impact.\nOriginality/value: The authors are unaware of any other work in the logistics field that measures the research impact of logistics scholars in this manner.\n",
"role": "assistant"
}
] |
A cross-cultural validation of a gender role identity scale in marketing
|
[
"Gender",
"Cross‐cultural studies",
"International marketing",
"Consumer behaviour"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Gender is a common segmentation variable and targeting criterion used by marketers (Wolin, 2003; Palan, 2001), and is regularly studied in the marketing literature (An and Kim, 2007; Milner and Fodness, 1996; Odekerken-Schroder et al., 2002). Researchers in psychology, sociology, communication, and marketing report that gender differences exist in cognitive abilities and social behaviors. However, in this research, what is meant by gender is not always clear. \"Gender\" can connote differences based on biology (male and female) or differences based on cultural roles (masculine and feminine). The majority of marketing studies evaluate the effect of gender as the biological distinction between male and female (Garnt and Bodenhausen, 1997). This biological view assumes that males and females are significantly different from birth and that the differences in their preferences and communication styles can be largely, or even entirely, explained by biological factors. As a segmentation variable, biological sex meets several requirements for successful implementation including:* it is easily identifiable;* segments are accessible;* segments are measurable and responsive to marketing mix elements; and* segments are large and profitable (Wolin, 2003).While relatively easy to implement, segmentation based on biological sex might be problematic in today's global market. Men and women exist within a culture and what it means to be a man or woman is heavily influenced by that culture (Palan, 2001). Culture is the sum total of capabilities and habits acquired by humans as members of society, encompassing their knowledge, beliefs, art, morals, laws and customs (Cateora, 1993). It represents the organized set of shared and learned values, as well as beliefs, expectations, and norms. At the cultural level, the set of roles dictated by social values and norms are referred to as masculine or feminine. Thus, beyond just the biological distinction, gender is used to connote the cultural norms, values and beliefs associated with roles of men and women.\nMethod: Participants\nResults: Internal validity: CFA of common measurement model\nGeneral discussion: One of the major criticisms of measures used in cross-cultural research is that they are not validated before being used in different countries. The primary purpose of this study was to investigate the efficacy of a shortened BSRI, the GTI measure of gender identity, across different cultures. This study has developed the first comprehensive validation of an important gender identity scale used in the field of marketing. Our findings suggest that a revised 16-item scale based on Barak and Stern's (1986) GTI scale can be used cross-culturally to measure gender identity. Theory building and testing in cross-cultural and cross-national environments is growing in popularity (Caruana and Ewing, 2006; Chan et al., 2007; Dolnicar and Grun, 2007), and this revised scale will enhance measurement applications and research conducted in foreign markets.Some limitations of the present research should be noted. First, this research was conducted using student samples. We believe this is a reasonable approach given that:* the BSRI from which the GTI scale was derived was originally validated using college students in the USA (Bem, 1974); and* the majority of studies in advertising involve student samples (Brown and Stayman, 1992).Nevertheless, a number of researchers have suggested that the lack of external validity inherent in using a student sample limits the generalizability of results (Lynch, 1982). With this consideration in mind, it would be desirable to replicate the present results using representative adult samples. Second, the findings from this study have only been validated based on three countries. We believe that these three countries provide a strong foundation for testing the cross-cultural validity of the GTI: the USA is a useful baseline in addition to being the country where the GTI items were developed, and Mexico and Norway represent extreme values on the masculinity/femininity dimension. We would speculate that the results of the present research should at least hold in countries that are culturally similar to those included in our sample (e.g. northern European and central American countries share many cultural characteristics with Norway and Mexico, respectively), but the truth of this speculation is a question for future research. It would be desirable to replicate the present results using a larger and broader sample of countries; in particular, the present results should be replicated in Japan and/or China both because these countries have substantially different linguistic and cultural foundations from the countries in our sample and because these countries are frequently examined in cross-cultural research. Third, the scalar equivalence results for the GTI were rather poor. Full scalar equivalence was not supported, and even the partial scalar equivalence model showed a noticeable decline in fit compared to both the configural E/I and metric E/I models. These results should not adversely affect research examining associations between the GTI and other variables of interest, but researchers who simply wish to compare mean GTI scores across cultures should do so with caution.Research conducted in the USA suggests that consideration of cultural values related to masculinity/femininity can lead to significant improvements in marketing effectiveness (Jaffe and Berger, 1988). Future research should examine whether this also occurs in other countries. The validated 16-item GTI gender identity scale is a useful instrument to examine this issue, and can help us further understand the role of masculinity/femininity in consumer decision making.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to explore the cross-cultural efficacy of a gender identity scale commonly used in marketing: the shortened version of the Bem Sex Role Inventory (BSRI) measure developed by Barak and Stern, the Gender Trait Index (GTI).\nDesign/methodology/approach: - Data were collected in the USA, Mexico, and Norway, and confirmatory factor analysis was used to assess the cross-cultural equivalence of the GTI.\nFindings: - Configural, metric and partial scalar invariance of a revised 16-item measure were supported.\nOriginality/value: - The validated 16-item GTI scale will enhance measurement applications and theory building in cross-cultural research, and further the authors' understanding of the role that gender identity plays in consumer decision making.\n",
"role": "assistant"
}
] |
Researchers show women at Spanish SMEs respond positively to high-involvement HRMs
|
[
"Human resource practices",
"Gender role theory",
"Employee affective commitment",
"Human resource management system"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: A study of Spanish SMEs has revealed that high-involvement HRM systems inspire emotional commitment in women workers. The results from 6,320 employees from 104 SMEs suggest HRM systems that combine different initiatives - training, participation, information and autonomy practices - are very effective at encouraging the involvement and participation of women.\n",
"role": "user"
},
{
"content": "Purpose: The authors wanted to look at the impact of both the overall HRM system and HRM sub-systems, in this case training, information, participation and autonomy.\nDesign/methodology/approach: The authors tested four hypotheses on thousands of employees from 104 Spanish SMEs. Previous studies of high-involvement HRM systems were done in large US firms\nFindings: The results showed women are more likely than men to reciprocate employer offerings of supportive HRM. They are also more likely to withdraw their commitment when the work environment is unsupportive. However, the results showed that the moderating effect of gender on the relationship between autonomy and affective commitment was not significant\nOriginality/value: The authors said that few previous studies had tried to explain the different impacts on men and women of perceived HRM practices. Their study was also unusual in investigating the impact of both the overall HRM system and HRM sub-systems, in this case training, information, participation and autonomy.\n",
"role": "assistant"
}
] |
Development of a customer satisfaction index model: An application to the Turkish mobile phone sector
|
[
"Customer satisfaction",
"Modelling",
"Mobile communication systems",
"Turkey"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The concept of customer satisfaction has attracted much attention in recent years. Satisfaction could be defined as a customer's post-purchase evaluation of a product or service. A key motivation for the growing emphasis on customer satisfaction is that higher customer satisfaction can lead to a stronger competitive position resulting in higher market share and profit (Fornell, 1992), reduce price elasticity, lower business cost, reduce failure cost, and reduce the cost of attracting new customers. Customer satisfaction is also generally assumed to be a significant determinant of repeat sales, positive word-of-mouth, and customer loyalty (CL). Satisfied customers return and buy more, and they tell other people about their experiences (Fornell et al., 1996).It is self evident that companies should listen to and satisfy their customers. Numerous studies have shown that the long-term success of a firm is closely related to its ability to adapt to customer needs and changing preferences (Li et al., 2006; Takala et al., 2006).The purpose of this paper is to provide information about the customer satisfaction indices (CSIs) and show the results of a CSI study carried out in Turkish mobile phone sector (TMPS). The rest of this paper is structured as follows: the second section briefly presents an overview of the CSIs and their developments in different countries. In the third section, a theoretical CSI model implemented in TMPS (CSI-TMPS) is described. The methodology and data analysis are discussed in the fourth section, followed by the analysis of the main results. Discussion and concluding remarks are provided at the end.\nCustomer satisfaction indices (CSI): The original interest in customer satisfaction research was on customer's experience with a product episode or service encounter (Yi, 1990; Anderson et al., 1994). More recent studies have focused on cumulative satisfaction. Cumulative satisfaction defines satisfaction as customer's overall experience to date with a product or service provider. This approach to satisfaction provides a more direct and comprehensive measure of a customer's consumption utility, subsequent behaviors and economic performance (Fornell et al., 1996). CSIs were built upon a cumulative view of satisfaction.The CSI model is a structural model based on the assumptions that customer satisfaction is caused by some factors such as perceived quality (PQ), perceived value (PV), expectations of customers, and image of a firm. These factors are the antecedents of overall customer satisfaction. The model also estimates the results when a customer is satisfied or not. These results of customer satisfaction are consequences factors such as complaints or loyalty of customer (Johnson et al., 2001). Each factor in the CSI model is a latent construct which is operationalized by multiple indicators (Fornell, 1992; Chien et al., 2002).Swedish customer satisfaction barometer (SCSB), reported in 1989, was the first national CSI (Fornell, 1992). It was applied to 130 companies from 32 Swedish industries. In 1992, the German customer barometer was introduced. The study is conducted for 52 industry sectors in Germany (Meyer and Dornach, 1996). The American customer satisfaction index (ACSI) was developed in 1993 by Claes Fornell, Founder of SCSB. The ACSI survey is conducted for seven main economic sectors, 35 industries, and more than 200 companies with revenues totaling nearly 40 percent of the US GNP (Fornell et al., 1996). The European customer satisfaction index (ECSI), developed by European organization for quality and European foundation for quality management, was first introduced in 1999 across 11 European countries (Eklof and Westlund, 2002). Other efforts for establishing national CSIs can be seen in Denmark, Austria, France, Netherlands, Switzerland, Taiwan, New Zealand, South Korea, Malaysia, Hong Kong and Russia (Grigoroudis and Siskos, 2003).The original SCSB model contains two primary antecedents of satisfaction: perceived performance and customer expectations. These both antecedents are expected to have a positive effect on satisfaction. The consequences of satisfaction in the SCSB model are derived from Hirschman's (1970) exit-voice theory which describes the results of dissatisfaction. The customer either exits (stops buying from the firm), or voices its complaint to the firm in an effort to receive restitution. It is expected that an increase in satisfaction should decrease complaints and increase CL (Fornell, 1992; Anderson et al., 1994).The ACSI model builds upon the original SCSB model specifications adapted in the distinct characteristics of the US economy. The main differences between the original SCSB model and the ACSI model are the addition of a PQ component, as distinct from PV, and the addition of measures for customer expectations. The ACSI model predicts that as both PV and PQ increase, customer satisfaction should also increase (Anderson et al., 1994). For the consequences, as in the SCSB, it is expected that when customer satisfaction increase then loyalty increase and complaints decrease (Fornell et al., 1996) (Figure 1).The ECSI, a modified adaptation of the ACSI model, considers the European economy as a whole, and thus, CSI scores of the countries can be compared with each other and with the European average (Eklof and Westlund, 2002). In the ECSI model, customer expectations, PQ, PV, customer satisfaction, and CL constructs are modeled the same as in the ACSI. There are two fundamental differences between the ACSI and ECSI models. First, the ECSI model does not include the complaint behavior construct as a consequence of satisfaction. Second, the ECSI model incorporates company image as a latent variable in the model. In the ECSI model, company image is expected to have a direct effect on customer expectations, satisfaction and loyalty (Grigoroudis and Siskos, 2003).The first studies on CSI model in Turkey were the research studies done by Turkyilmaz and Ozkan (2004) and Aydin and Ozer (2005). Aydin and Ozer (2005) developed and tested a new model for Turkish global system for mobile (GSM) users. The structural model they used includes some new constructs, such as switching cost, trust, and complaint handling. They collected the data from 1,662 GSM users in four Turkish cities using a face-to-face survey. In their study, the model is estimated using maximum likelihood (ML)-based covariance structure analysis method namely LISREL.Then the first national model, Turkish customer satisfaction index (TCSI), was reported as a pilot study in the fourth quarter of 2005 by Turkish Quality Association (Kal-Der) and KA Research Limited. Since, the measurement model of the TCSI is same as the ACSI model, it includes customer expectations, PQ, PV, customers satisfaction, CL and customer complaints constructs. In 2006, the TCSI was implemented for 100 companies from 25 sectors (TMME, 2006).\nTheoretical model of CSI-TMPS: The ECSI constitutes the framework for CSI-TMPS. CSI-TMPS is based on a structural equation model (SEM) which comprises the antecedents and consequences of customer satisfaction. Shown in Figure 2, the left sided factors (i.e. image, customer expectations, PQ, and PV) are the ancients of the customer satisfaction while the right sided factor (i.e. CL) is the consequence. The hypothesized relationships between the latent constructs are depicted with lines. Although in reality there may exist some more relations between the factors, the most important ones are considered.Image\nMethodology: Research instrument and data collection\nData analysis: PLS estimation of CSI-TMPS\nIndex scores for mobile phone sector: The index scores of the latent variables are calculated as the weighted average of manifest variables pertaining to their own block. Then the CSI score is calculated as following: Equation 1 Where y is the manifest variable related to latent CSI, and w is the unstandardised outer weight between latent variable and related centered manifest variable.Considering the survey data from 652 mobile phone customers, the CSI score for Turkish mobile phone industry is found 64.09 (for 0-100 scale). The other latent variable scores are shown in Figure 5.The results show that mobile phone users' expectations are very high (71.71). On the other hand, the quality level of the products they perceived is also very high (70.05). The lowest score of the CSI-TMPS model is PV (59.76), which shows the price/quality and quality/price evaluation. Since, mobile phone firms are known as the leading firms for technological products their image is expected to be high. Image score for TMPS is 67.47. CL score is found to be 66.83. Loyal customers are those who keep buying from the same company, recommend to others, and have price tolerance.\nDiscussion: The CSI-TMPS model is a SEM which consists of well-established theories and approaches in customer behavior. The constructs of the CSI-TMPS model are latent variables indirectly described by a block of measurement variables. A major advantage of the measurement model is the use of generic questions, which are sufficiently flexible to be used across a wide variety of products, services, and public sector services.The structural model of the CSI-TMPS was analyzed using the PLS method. The general applicability of a CSI model depends on the reliability and validity of the modeling results (Chan et al., 2003). Reliability and validity of the CSI-TMPS model was assessed by checking unidimensionalty of the blocks, individual item reliability, convergent validity and discriminant validity. All test results are found to be satisfactory. The primary objective of PLS is the minimization of error (or maximization of variance explained) in all dependent constructs, which can be measured by R2 values of structural models. The R2 value for customer satisfaction is 0.64. This result satisfies the crucial requirement for validity of structural model (i.e. predictive power). Finally, all the relationships between the latent constructs in the structural model are significant (p<0.001).From the results for TMPS, we can conclude that satisfaction is mostly effected by PV. That is, when customers perceive that the money they pay for the product is worth of its quality, their satisfaction increases. On the other hand, PQ and image are other important factors effecting satisfaction. These results suggest that company managers should pay more attention to their pricing policies, and to the activities which improve product quality and company image.CL, the ultimate factor in the model, is another important construct that should be taken into account. The findings show that customer satisfaction and company image have positive and significant effects on CL. Customer satisfaction is found to be the most important factor for enhancing CL.\nConclusion: A number of CSIs have been introduced in the last decade. While some index models (i.e. SCSI, ACSI) are applied in national level, the ECSI being implemented by most of the EU member countries is an international satisfaction index model (Eklof and Westlund, 2002). We developed and applied the first CSI model for the TMPS as an academic research project in May 2003 (Turkyilmaz and Ozkan, 2004). This present study is identical to the previous one with the exception that the relationship between customer expectation and PV is removed from the previous model. In order to help the respondents understand and answer survey questions easily, wording of some questions have been updated.The results of the CSI-TMPS model as a whole can be a valuable guide for the managers in formulating competitive marketing strategies. For the companies, independent and uniform measurement characteristics of the CSI-TMPS model provide a useful tool for tracking performance and systematic benchmarking over time. Considering the results of the CSI-TMPS model, the limited resources of the firms can be allocated for critical factors which have important impacts on satisfaction. In conclusion, the CSI-TMPS model provides important information for the purchase decisions of the customers and lead to improvements in the quality of goods and services they consume.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this study is to develop and test a new customer satisfaction index model for Turkish mobile phone sector (TMPS).\nDesign/methodology/approach: - Based on relevant literature and theoretical considerations on customer satisfaction, a new measurement model was developed. A survey based on the proposed model was conducted to 700 mobile phone users in Istanbul. Partial least squares method was employed to estimate the proposed structural equation model (SEM).\nFindings: - The general applicability of a SEM depends on the reliability and validity of the modeling results. The proposed satisfaction model exhibits strong explanatory power with its satisfactory reliability and validity results.\nResearch limitations/implications: - The proposed model was tested for the TMPS. It should ideally be tested periodically for different sectors with more data, and the index results should be compared.\nPractical implications: - The results of this study provide very important information for the managers in formulating competitive marketing strategies. The results show the critical points where the limited resources of the firms should be allocated to improve satisfaction and loyalty. Independent and uniform measurement characteristics of the studied model provide a useful tool for tracking performance and systematic benchmarking over time. It also provides information about the weaknesses and strengths of the company from the eyes of its customers.\nOriginality/value: - This study is an attempt to develop and implement a satisfaction model for the TMPS. The proposed model consists of well-established theories and approaches in customer behavior.\n",
"role": "assistant"
}
] |
Prolonged time taken to complete the degree programmes at the Faculty of Engineering Technology of the Open University of Sri Lanka: Barriers and remedies
|
[
"Time",
"Engineering",
"Barriers",
"Degrees",
"Remedies"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Faculty of Engineering Technology (FET) at the Open University of Sri Lanka (OUSL), the only national university, offers engineering programmes via open and distance learning (ODL) in Sri Lanka. The faculty has completed 30 years of existence and has produced graduates in the disciplines of agriculture, civil, mechanical, mechatronics, computer, electrical, electronics and communication engineering and in industrial studies in agriculture, textile manufacture and apparel technology.\nDegree programmes: Bachelor of Technology (Engineering) (BTech) and Bachelor of Industrial Studies (BIS) are the two degree programmes offered by the FET. These degrees are designed in accordance with the requirements of the Sri Lanka Qualification Framework and are recognized both academically as well as professionally. The BTech (Engineering) degree of the OUSL is recognized by the Institution of Engineers Sri Lanka as a fulfilling academic qualification for associate membership. Furthermore, the BTech (Engineering) degree is equivalent to the engineering degrees offered by other national universities (Student Guide Book, 2014).\nMethodology: The research design involved extracting a sample of graduates from FET and distribution of a structured questionnaire among them. Exploratory interviews with the easily contactable graduates from FET were conducted before developing the questionnaire.\nResults and discussion: The duration for the completion of the degree programmes by the respondents was analysed, and after removing the outliers, it was found that more than 90 per cent of the students have taken 6-15 years to complete the four- and five-year degree programmes (Figure 1).\nConclusions and recommendations: Findings of this study provide some insights for shortening the duration of the BTech and BIS degree programmes offered by the OUSL. Institutional barriers such as poor academic counselling and guidance, laboratory sessions being inappropriate/not sufficient to understand the course material, inadequate resources and facilities such as library resources and computer and internet facilities, poor teaching and learning environments and lack of laboratory facilities at the regional/study centres have contributed immensely to the lengthy duration taken to complete the degrees. But the institutional barriers can be easily overcome by strengthening the centre network that will enhance learner support and updating and upgrading of laboratory facilities. Counselling at the time of registration for each academic year should be carried out by the teachers of FET, and proper guidance should be given in selecting appropriate course loads. Personal factors such as work-related challenges, travelling time and cost of commuting to the main centre located in Colombo and inability to spend the required time expected of the programme have played major roles on the prolonged completion of the degree. Results indicate that the cost of tuition is not a major barrier for on-time completion. Psychological barriers such as possessing a limited repertoire of study strategies, lack of understanding of ODL methods and deficiency of constant motivation throughout the study period had a tremendous impact on the delayed graduation. Personal barriers can be alleviated by molding the learners to be good time managers and independent learners with strong focus. Providing the learner with a greater repertoire of study strategies may remove physiological barriers. All these can be achieved via \"Student Orientations\" which should be conducted with a strong focus on time management, developing academic skills and ODL methods. The remedies suggested for external means of shortening the duration of the degree programmes were well accepted by the respondents.\n",
"role": "user"
},
{
"content": "Purpose: It has been observed that the students attending the bachelor's degree programmes offered by the Faculty of Engineering Technology (FET) at the Open University of Sri Lanka (OUSL) take extraordinarily longer period to complete their degrees. The purpose of this paper is to identify the institutional barrier, student-related personal barrier and psychological barrier behind the prolonged duration of completion of the said degrees.\nDesign/methodology/approach: A well-developed questionnaire was administered among a selected sample of graduates who were awarded the Bachelor of Technology (Engineering) and Bachelor of Industrial Studies degrees within the past ten years.\nFindings: Results indicate that institutional barriers such as poor academic counselling and guidance; laboratory sessions being inappropriate/not sufficient to understand the course material; inadequate facilities, such as library resources, and lack of laboratory facilities at the regional centres have contributed immensely to the lengthy duration taken to complete the degrees. Personal factors such as work-related challenges, travelling time and cost of commuting to the main centre located in Colombo and inability to spend the required time expected of the programme have played major roles in the prolonged completion. Results indicate that the cost of tuition is not a major barrier for on-time completion. Psychological barriers such as possessing a limited repertoire of study strategies, lack of understanding of open and distance learning (ODL) methods and deficiency of continuous motivation had a tremendous impact on the delayed graduation. A high satisfaction rate was observed regarding the ODL tools that were used within the programmes and the formative and summative evaluation criteria. Approximately 75 per cent of the sample approved of the five remedies suggested.\nOriginality/value: Findings of this study provide insight for shortening the duration of the bachelor's degree programmes offered by the FET at the OUSL.\n",
"role": "assistant"
}
] |
Learn how to grab others' knowledge to do your job properly
|
[
"Organizational culture",
"Innovation",
"Knowledge management",
"Knowledge organizations",
"Knowledge sharing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Management scholars (e.g. Drucker, 1992; Quinn, 1992; Grant and Baden-Fuller, 2004; Jasimuddin et al., 2005) claim that knowledge in organizations is the most important strategic resource for their sustainable competitive advantage. Resonating with this, companies, including manufacturing, financial services, and consulting firms, recently started viewing themselves as knowledge-intensive. Every organization pays special attention to the notion of knowledge as one of its top corporate agenda and spends a huge amount of time and money to find ways to better manage its knowledge resources. As a result, knowledge management has received increased attention in practitioner circles.\nStatus: Part of the status issue depends upon who the knowledge contributor is. A knowledge seeker may interact with colleagues of the same status or higher status (i.e. immediate boss or manager of managers) to get technical advice. Working with the same level of colleagues may be just around the corridor, while someone within the team or located nearby visiting and sitting down together in an office to get ideas is quite effective.If the contributor is a senior member of the organization, then it would be wise to adopt the formal technology-mediated approach. If the individual is known at a personal level, then the employment of a more informal people-focused approach is workable. In contrast, when the manager's time has to be taken seriously, sending an e-mail seems to be perfect.\nPersonal preference: The personal preference of the knowledge owner and how he or she likes to be approached has considerable influence in deciding the way that knowledge seekers should approach them. It's most effective to follow a mechanism that the contributor prefers to use. The fact is that if a knowledge contributor is an immediate boss or a boss of the boss, the knowledge seeker's preference should be guided by the manager's personal choice of medium. For example, if the potential knowledge contributor prefers e-mail to communicate, it seems to be the most effective medium for the knowledge users to interact with him or her. Hence the knowledge user needs to gather information about the personal preference of the manager - that is, what the contributor's preferred style is - in the first place before initiating any interaction.\nIntimacy: Generally speaking, approaching somebody also depends upon how well the parties engaged in the knowledge transfer process know each other. The fact is that intimacy is crucial for successful knowledge transfer. There are two situations in relation to intimacy:1. strong social ties; and2. weak social ties.A knowledge seeker may interact face to face with a potential knowledge contributor with whom they have a strong social tie. On the other hand, weak ties restrict the use of an informal approach such as face-to-face interaction. When a knowledge seeker approaches a knowledge contributor he or she does not know, then it might irritate the contributor, because he or she has to respond instantly, which may not be his or her priority. So the most expected approach would be to use e-mail due to the lack of intimacy between the two parties.However, technology-mediated communication is not very effective in building social ties. Therefore, knowledge users must attempt to have face-to-face interaction in order to build personal networks, even if they do not have intimacy. That's why getting an appointment for a meeting and discussing things face-to-face help to build up a personal relationship with people with whom the knowledge seeker has less acquaintance. There are very good reasons to talk to the individual who knows the answer, and this helps to develop a social network.\nTrust: An environment of trust is conducive to knowledge exchange (Keong and Al-Hawamdeh, 2002). There are two situations with respect to trust:1. high trust; and2. low trust.High trust is a situation in which there is a set of high expectations shared by the actors in a knowledge exchange process and low trust is a situation in which there is a lack of such expectations. Since trust encourages organizational members to be collaborative and open with each other in the sense that the knowledge seeker can approach directly without hesitation to collect a provider's knowledge. However, due to lack of trust, both the prospective user and knowledge contributor may prefer to have explicit knowledge transfer using a technology-focused tool to start with. Under this circumstance, sending something by e-mail or another electronic medium is also good because text can be stored straight away and the big point is that there is no dispute between the parties. They can look back and read something that is written down; later, they can prove it. With talking, there is no proof.It is crucial to create trust between a knowledge seeker and a knowledge contributor, which, in turn, helps the contributor to think that the user will not misuse the knowledge acquired for purposes other than those agreed upon. Hence, building trust among the members of an organization is thought to be a necessary condition for an organization's success. However, it is hard to develop trust with someone if the actors cannot see each other face to face. In addition, trust does not build overnight. Frequent interactions between actors in the knowledge transfer process are reported to be vital to develop trust.\nCriticalness of a technical query: The way in which a prospective knowledge seeker grabs others' knowledge also depends on whether the technical query is time-critical - that is, it depends on the urgency of the knowledge. If the knowledge user needs information which is very much time critical, then he or she has to get an answer quickly so he or she should be allowed to go to an expert, i.e. get the answer quickly and interactively with an expert.For example, organizations keep receiving queries from their customers. Some of them are related to fixing customers' problems, and need an immediate response. Such problems have a high priority. If a customer-related problem is so crucial that the individual needs a quick answer then he or she should be allowed to go to any person at all, whether they have an intimacy or not. The knowledge provider's position should not matter.Management has to undertake initiatives to encourage organizational members to exchange knowledge with each other and build trust. Consequently, such encouragement will help to create a supportive knowledge-transfer culture. Even if the organization doesn't have a strong knowledge sharing culture, there should be an unwritten policy to allow a prospective knowledge seeker to discuss something urgent with somebody else, irrespective of status, personal relationship, trust, preference, or any other factor.Experience from large multinational corporations reveals that face-to-face interaction is the most effective way to grab a knowledge provider's knowledge. This is reflected by an employee: \"Using face-to-face conversation is the best way\". However, when positioned apart, it is difficult to work together face-to-face. The fact is that the distance can make getting technical help through face-to-face interaction impossible. When employees are split geographically, the sensible means for them is a technology-centric mechanism such as e-mail, instant messaging, etc.. A manager points out: \"He [contributor] is in America now and I am here [UK] [...] that communication is more convenient to be in e-mail. If he were here I would probably talk to him because there is something about context\". This adds another dimension, such as proximity, in addressing the way in which organizational members who are working apart will carry out knowledge transfer.Academics need to make efforts to explore the way in which managers select an appropriate mechanism of knowledge transfer, and to identify the determinants of the selection of a particular mechanism. The potential findings of their efforts might also help to guide practitioner circles in concrete terms.\n",
"role": "user"
},
{
"content": "Purpose: - This paper addresses the ways in which a prospective knowledge user can approach seeking a potential contributor's knowledge.\nDesign/methodology/approach: - Qualitative data were collected in 2004 using semi-structured interviews from a large multinational corporation situated in the UK.\nFindings: - The main findings of the paper suggest that context and content - such as status of the knowledge contributor, personal preference of superior, intimacy with the knowledge provider, trust between the actors, and criticalness of the technical query - have a great role to play when asking for technical help.\nResearch limitations/implications: - The paper ignores to identify the determinants of the selection of knowledge transfer mechanisms, which is an interesting area of future research for academicians.\nPractical implications: - The paper will help organizational members to consider the contexts before approaching their colleagues for a technical advice.\nOriginality/value: - The paper is of value to practitioners because it develops a view of grabbing others' knowledge without hesitation.\n",
"role": "assistant"
}
] |
Dealing with the "Grumpy Boomers": re-engaging the disengaged and retaining talent
|
[
"Job satisfaction",
"Organizational commitment",
"Baby Boomers",
"Supervisor support",
"Intent to turnover",
"Work-role overload"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Shediac et al. (2011, p. 2) opened a recent report on global population trends by stating that \"demographics shape our world\". More specifically, their analysis indicated that the social, economic and environmental impacts of one demographic megatrend in particular, increased aging, would be significant and global in nature.\nTheoretical framework: The theoretical framework shown in Figure 1 was developed to guide our analysis. Justification for using organizational commitment and intent to turnover to categorize our group of over 50 knowledge workers is provided below.\nMethodology: Sample\nResults: Tables I (demographics) and II (outcomes) provide key information on the different groups of Boomers in the sample. With only a few exceptions, group membership was not associated with any of the demographic variables considered in this analysis. The following significant between-group differences were, however, noted:\nDiscussion: The purpose of the current study was threefold: first, identify those work factors and practices that predispose older knowledge workers to fall into one of the four groups identified in Figure 1; second, offer suggestions on how organizations can reduce the turnover intentions of highly committed knowledge workers over the age of 50; and third, offer suggestions on how organizations can increase the commitment of knowledge workers over the age of 50 with low intent to turnover.\nConclusions: The findings from this study are supportive of our theoretical framework and reinforce the idea that how an employee is treated at work (supportive manager, work-role overload), the values and norms enacted within the organization (supportive culture) and the work itself (job satisfaction) influence organizational commitment and intent to turnover. They implicate workloads as a key determinant of an employee's intent to turnover and they reinforce the importance of supportive management to both commitment and intent to turnover.\n",
"role": "user"
},
{
"content": "Purpose: The aging of the workforce and the impending labour force shortage at the skilled end of the labour market increases the need for organizations to understand how to \"re-engage\" older workers with low commitment and reduce the turnover intentions of committed older knowledge workers. The current study addresses this issue by using employee commitment and intent to turnover scores to classify older knowledge workers into four groups: Disengaged-Exiters, Engaged-High-Performers, Retired-on-the-Job and Exiting-Performers. The purpose of this paper is to identify a set of work factors and practices that predispose older knowledge workers to fall into one or another of the four groups and offer suggestions on how organizations can increase commitment and decrease intent to turnover of their older workers.\nDesign/methodology/approach: The paper used survey data (n=5,588) from a Canadian national study on work, family and caregiving to test the framework. Data analysis was performed using a MANCOVA with one independent variable (Boomer group), four dependent variables (job satisfaction, non-supportive culture, supportive manager, work-role overload) and one covariate (gender).\nFindings: The results support the framework. The findings suggest organizations that wish to retain committed Baby Boomers need to address issues with respect to workload. Alternatively, organizations who wish to increase the commitment levels of Boomers who have \"Retired-on-the-Job\" need to focus on supportive management, organizational culture and career development.\nOriginality/value: This paper contributes to the literature on organizational commitment and intent to turnover by re-conceptualizing the relationship between these traditional concepts.\n",
"role": "assistant"
}
] |
Does income matter in the happiness-corruption relationship?
|
[
"Corruption",
"Happiness",
"Income thresholds"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Economists have studied the relationship between happiness and numerous variables for decades. Early studies (most notably, Easterlin, 1974) concentrate on the relationship between income per capita and happiness. The conclusion from these enquiries is clear. In the primarily western countries studied, the change in happiness is small relative to remarkable increases in incomes and living standards which have occurred over time. Indeed, in some countries people are no happier than before despite improvements. Similar findings emerge across countries: there is little or no association between increases in per capita income and mean happiness levels. These conclusions are in contrast to many other studies which indicate that within a given society, richer people are, on average, happier than poorer; see Deaton and Stone (2013) for a review[1]. More recent investigations comparing rich and poor citizens within a country, comparing rich and poor countries, or measuring happiness in countries over time support the notion that there is a critical level beyond which income no longer contributes to happiness. For instance, Kahneman and Deaton (2010) estimate the satiation point to be around $75,000 for the USA[2]. Still other studies maintain that income always has a positive effect on happiness (Stevenson and Wolfers, 2013).\n2. Literature survey: 2.1 Defining happiness and the literature on happiness\n3. Rationale for the hypothesis, data, and empirical model: Our model must capture the effect of corruption on happiness controlling for income and other variables. Specifically, we estimate the marginal effect of corruption on happiness controlling for levels of real gross domestic product (GDP) per capita. While we do not hypothesize the existence of a Kuznets-type curve for the relationship between happiness and corruption[8], we postulate that the nature of correlation between happiness and corruption depends on a country's level of per capita income. There are reasonable grounds for advancing this hypothesis. Intuitively, anti-corruption measures are income elastic. At low-income levels, corruption may be endemic, but it may not correlate with happiness for several reasons. Poor people may simply have little at risk, or more likely, corruption itself may be merely one of many depravities they face and as a consequence it alone has no discernible impact on happiness. At higher income levels, corruption will begin to sting, and people will increasingly become concerned. As countries approach this threshold, anti-corruption measures are adopted[9]. As incomes rise further, anti-corruption measures become tighter and corruption becomes less evident. As a consequence, for income levels beyond this threshold, corruption declines and therefore has an increasing effect on happiness.\n4. Empirical findings and discussion: 4.1 Basic results\n5. Conclusions: This study is motivated by the postulate that happiness in countries is related to corruption. However, to assume that this relationship is monotonic would be too simplistic. Thus, we test the hypothesis that the association between happiness and corruption is income dependent.\n",
"role": "user"
},
{
"content": "Purpose: Empirical evidence on the relation between happiness (life satisfaction) and corruption is barely perceptible in the literature. The purpose of this paper is to contribute to closing this gap by presenting some estimates using a large cross-section of countries over the period 1996-2010.\nDesign/methodology/approach: The empirical model allows both corruption and per capita income to enter as arguments of a happiness \"production function\". The correlation between happiness and corruption is presumed to be non-linear.\nFindings: While the results do not support the existence of a Kuznets-type trajectory, the study finds that the level of per capita income determines whether happiness and corruption are related and in what way. The authors estimate cutoff income levels at which corruption has a discernible effect on happiness. The results show that corruption reduces happiness, but only for high-income countries - roughly the upper half of the income range in the sample.\nPractical implications: Results nullify the oft-asserted statement that happiness is negatively linked to corruption in all countries. The nature of correlation is more complex.\nOriginality/value: The paper goes beyond simply testing whether happiness is related to corruption. It conjectures that the relationship between the two variables is non-monotonic. Thus, the analysis considers the notion that the association between happiness and probity is income dependent. A novel feature of the empirical model is that the estimated income cutoff levels are endogenously determined. That is, income thresholds are not pre-determined. The authors also test for the robustness of the results by addressing the issue of potential endogeneity of corruption.\n",
"role": "assistant"
}
] |
Franchising and initial public offering: a signaling perspective
|
[
"IPO",
"Franchising",
"Performance",
"Key success factors",
"Going public",
"Stock market listing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Franchising, defined as a contract-based business system by which one party grants to the other, for a consideration, the use of a combination of intellectual property and/or industrial rights, know-how, technical and commercial assistance, as well as the opportunity to be part of a network, has had a considerable impact on the USA and other countries and has been hailed as one of the best business system innovations of the twentieth century (Alon, 2010). According to a study conducted by the International Franchise Association (IFA) (HIS-Markit, 2018), franchising business in the USA supports more than 745,000 franchise establishments and 7.8m direct jobs, with an output of 713bn USD contributing about 425bn USD to the US gross domestic product (GDP). Business format franchising, a model for transferring know-how and intellectual capital in return for consideration, usually in the forms of royalties and fees, covers numerous business lines including: automotive, business services, commercial & residential services, lodging, personal services, quick service restaurants, real estate, retail food, retail products and services and table/full-service restaurants. Franchising is popular with service businesses, as it alleviates the problems of heterogeneity, perishability and scalability through quality standards, a common brand and distributed networks of service delivery. With such scale and scope, understanding the determinants of franchising financial success is paramount.\nModel development: Franchising is viewed as a hybrid organization form that has elements of markets and hierarchies because the franchisor retains some ownership and control, but gives up the operations of the units to franchisees (Elango and Fried, 1997). Two popular approaches to explaining the use of franchising have been resource-scarcity and agency theories. Most studies using these theories have tried to explain the factors associated with the proportion of franchising. The theory of resource scarcity suggests that firms initially franchise because they lack the resources for expansion. Such resources are not merely financial (capital scarcity), but also managerial, knowledge-based, organizational, etc. (Elango and Jawahar, 2002). The agency perspective, on the other hand, focuses on the monitoring skills of the franchisor. According to the agency perspective, firms franchise because they are unable to monitor their managers efficiently. Because franchisees have a residual claim on sales, shirking is avoided and the need for monitoring is diminished (Shane, 1996a). A number of studies have shown that these theories are complementary in explaining franchising proportion (Combs and Castrogiovanni, 1994; Combs and Ketchen, 1999; Boulay, 2010). For example, Combs and Ketchen (1999) suggested that resource scarcity can help agency theory explain franchising. The authors examine public firms and suggest that those firms are in less need of capital by virtue of their public listing. Resources obtained from public listing reduce the capital scarcity of the firm, allowing them to use less franchising in their expansion. In addition to a change in research foci, a number of franchising scholars have called for theoretical plurality beyond resource scarcity and agency explanations and have suggested signaling theory as an alternative (Lucia-Palacios et al., 2014; Michael, 2009; Lafontaine, 1993).\nMethodology: Sample selection\nResults: Table I provides the descriptive statistics of the sample along with the correlations among variables of interest. A review of Table I indicates the correlation values between the independent variables to be less than 0.7. Age had the highest correlation (-0.672; p<0.01) with time to franchise, as one would expect, as these two are linked with each other. However, we also checked and confirmed that the VIF score for each of the independent variables was below 10 to reduce the threat of multicollinearity invalidating this study's findings. For the independent variables in our model, the highest VIF score was 5.06 and the lowest was 1.02.\nDiscussion: Perhaps the experience of Domino's Pizza mirrors that of our sample of public franchisors. Domino's started in 1960 with brothers Thomas and James Monaghan who borrowed $900 to start a pizza shop in Ypsilanti, Mich. Domino's opened its first franchise store in 1967, went global in 1983, and public in 2004. Despite a rough IPO start, the stock sold for about 14 USD, and about $220m of the IPO revenue went to shareholders Monaghan, Bain Capital and J.P. Morgan, with Domino's receiving about $136.7m (Market Watch, 2004). As of the writing of the last iteration of this paper (February 21, 2018), the company was trading with a Price-Earnings Ratio of 42.2 at about 222 USD per share, a substantial multiplier over its original IPO price in 2004.\nConclusion: This study makes a singular contribution to the franchising literature by invoking signaling theory. Signaling theory offers powerful, unexplored explanations of franchising (Lucia-Palacios et al., 2014; Michael, 2009; Lafontaine, 1993). The models in this paper have Nagelkerke R2 values ranging from 0.38 to 0.537, suggesting that signals have a strong collective explanatory power for public listing. Managers desiring to go public might wish to signal through IFA membership, rapid franchising expansion, internationalization and, to a lesser extent, TMC and earnings disclosure. Costly signals, such as IFA membership and franchising, may have a stronger influence on the IPO potential. Signaling theory in franchising research may provide a powerful alternative or complement to the already well-entrenched resource-scarcity and agency theoretic explanations. As well as increased study outside the North American franchising context, more research is needed to show how signals may explain other franchising factors, such as ownership redirection, internationalization, pricing and performance.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the factors associated with franchisors going public using signaling theory. Listing on the stock market is a sign that the business concept has reached a threshold level of acceptance and success. To increase the relevance of this study to practitioners, the authors focus on franchising-specific controllable variables.\nDesign/methodology/approach: This study uses a sample of 2,134 franchisors from US drawn from a survey by Entrepreneur magazine during the years 2015-2016. Binominal logistic regression models are used for analysis of the data.\nFindings: Findings indicate that time to franchise, international operations, franchise association affiliation, disclosure and extent of top management commitment are factors positively related to the likelihood of a franchisor being publicly listed.\nResearch limitations/implications: Study findings are based on a sample of franchisors from North America, where financial markets are well developed, and due caution should be exercised before generalizations are made to other contexts. A major implication of this study is that signaling theory may provide an important supplement to the already well-entrenched resource-scarcity and agency theoretic explanations in franchising research.\nOriginality/value: While signaling theory is growing in importance in the franchising literature, this study is the first to uncover the relationship between company signals and initial public offering.\n",
"role": "assistant"
}
] |
Lessons in spiritual leadership from Kenyan women
|
[
"Social justice",
"Leadership",
"Kenya",
"Women"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nLessons in spiritual leadership from Kenyan women: Whereas there are growing numbers of leadership studies about women from North America and other Western countries, there remains a dearth of literature about women's leadership in Africa (Ngunjiri, 2009, 2010; Nkomo and Ngambi, 2009). As Nkomo and Ngambi argued, African women leaders are missing from the discussions, necessitating concerted efforts to study and publish about them. Furthermore, the two authors posit that there is a tremendous need to capture and develop all available talent in order to resolve the myriad of issues facing the African continent:Thus, scholars cannot afford to neglect understanding the status and experiences of African women who occupy leadership and management roles in African organizations, or the barriers to their success (Nkomo and Ngambi, 2009, p. 50).With this call to action in mind, the paper aims at explicating the leadership experiences of select women from Kenya, who occupy leadership positions in various institutional types, and whose leadership also involves an educational and social justice perspective. However, the focus of this paper is not on the barriers or challenges per se. Instead, taking an appreciative inquiry perspective, the paper demonstrates how select women utilize Africana spirituality in articulating leadership for social justice. In that sense then, the paper focuses on \"what works\" in the midst of, rather than the absence of challenges and barriers (Jones, 2003; Lawrence-Lightfoot and Davis, 1997; Ngunjiri, 2006, 2007).\nSocial-historical context: Kenya is a small East African country that is only about 583,000 square kilometers or twice the size of the state of Nevada, US. It is a former British colony that attained independence in 1963, and has experienced tremendous social, economic and political changes in its five decades of existence as a country. The twin factors of colonialism and Christianity helped to interrupt traditional communities and their ways of life, including their gender roles and social arrangements, with effects that are still felt today. Whereas the country has been stable for most of its history, the beginning of 2008 dawned with the worst political unrest and violence ever experienced within its borders. Prior to that, the political unrest of the late 1980s and early 1990s were much less chaotic, resulting in Kenya advancing into an emerging democracy. Many of the leaders who were interviewed for the larger study of which this paper is derived advanced into leadership within that tumultuous political environment which had an impact on all organizational types - be they government, business, education or civil society (Ngunjiri, 2010).\nMethods: This paper is derived from data collected for my dissertation in the summer of 2005 which was based primarily on qualitative interviews with supplemental archival data and observations of 16 women who I describe as being \"broadly educational leaders\" - meaning that whether their leadership was within an educational institution or a different type of organization, they considered themselves as responsible for educating their constituents and those within their spheres of influence formally or informally on issues relevant to social justice. Each interview lasted between two to four hours, and I transcribed them verbatim, coded them and analyzed them utilizing appropriate qualitative methods (e.g. open coding, axial coding). I hand coded all the data. For this particular paper, I have re-analyzed the data with a view to understanding how the type of spiritual leadership exemplified by the women leaders was similar to, and different from, western conceptualizations of leadership, particularly those described by management scholars (e.g. Fry, 2003; Kriger and Seng, 2005; Reave, 2005; Whittington et al., 2005) and education scholars (e.g. Dantley, 2003a; Keyes et al., 1999; Woods, 2007). For more on the methods employed in the larger study, please see Ngunjiri (2006), available free on the internet.Theoretical framework\nIntroducing the women leaders: This paper is derived from a much larger study of women's leadership in Africa which focused on 15 Kenyan and I Sudanese woman; their leadership had both an educational and a social justice perspective (see Ngunjiri, 2006, 2010, for deeper explication of other themes relating to women's leadership in Africa). The women's spheres of influence ranged from grassroots to global - from the principal of an elementary girls school to president of a global organization, from national politicians to pan-African leaders. Each of the women described their role in terms of leading for social justice - whether this be in the form of economic, legal, educational, or political justice for the marginalized. In addition, each woman understood her role as broadly educational, such that even those whose position was not as \"head in command\" of an educational institution, they still recognized their educational leadership role in the community, nation and international spheres of influence. In that sense, their leadership involved being both institutional leaders and opinion leaders, with the ability to influence change from within organizations, communities, and the nation and for some pan-African and Global organizations. The women ranged in age from early forties to early eighties, thus providing a rich diversity of ideas and experiences in leadership from the pre-independence to contemporary times. Their stories provided rich lessons in spiritual leadership, helping to expand our current understanding of the phenomenon by adding African data, particularly the critical perspective derived from practical wisdom, and the cultural values inherent in ubuntu philosophy.Lessons in spiritual leadership\nImplications for research and practice: These four values: beneficence, hope, courage and ubuntu together add to our understanding of leadership praxis as far as the integration of spirituality with leadership is concerned. These stories demonstrate how leaders can employ their identity - including their spiritual identity - in their leadership practice, thus becoming more authentic leaders by being able to act from deeply held spiritual values. In particular, these four values contribute to the leaders' ability to build and sustain community. As the literature on spiritual leadership and spirituality in organizations demonstrates, people are searching for community within their organizations and workplaces (Benefiel, 2003, 2005; Fairholm, 1997; Fry and Whittington, 2005). As such, educational leaders can learn from these snapshots of Kenyan women leaders how to articulate a leadership that enables them to be effective in creating community within their educational institutions, thus enabling their constituents to experience a deeper sense of belongingness (Fry, 2003; Keyes et al., 1999).The role of spirituality in leadership practice can therefore not be overemphasized - perhaps it is time for those in educational leadership to catch up to those in the management disciplines in researching about spiritual leadership in educational administration. So far, there are a few such studies and conceptual papers available (for example, Dantley, 2003a, b, 2005a, b; Dillard et al., 2000; Keyes et al., 1999; McClellan, 2006; Murtadha-Watts, 1999) but there is room for more, particularly those studying non-western forms of spiritual leadership.\nConcluding thoughts: At the present time, there is a greater need for effective international and cross cultural communication, collaboration and cooperation, not only for the effective practice of management, but for the betterment of the human condition (House et al., 2004, p. 1).With this in mind, studies on spiritual leadership and gender from many diverse perspectives, including those utilizing non-western theories and non-white populations are welcome for both expanding knowledge and improving human relationships across differences. Stories such as those used to demonstrate the four themes of courage, hope, beneficence and ubuntu/humanness help to remind us that there are leadership lessons to be learned from the margins and the frontiers; that there are important contributions to knowledge and praxis that can be derived by studying those that have historically been excluded from theorizing. The value of such studies is immense in expanding our understanding of leadership theory and practice beyond the current canon. Even cultural worldview such as Ubuntu with its African roots has applicability on a global scale, because its values of solidarity, generosity, mutuality and commitment to community are values that can find resonance well beyond Africa's borders.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to explicate spiritual leadership lessons of beneficence, courage, hope and ubuntu/humanness that are derived from the experiences of women leaders in Kenya. The paper seeks to connect African data with existing literature on spiritual leadership, to demonstrate where African spiritual leadership is similar to, or different from, western conceptualizations of spiritual leadership.\nDesign/methodology/approach: - The study from which this paper is derived employed qualitative methods, specifically interviews with supplemental archival data and observations. Four major themes are explored: beneficence, courage, hope/forbearance and ubuntu/humanness as emerging from the women's leadership stories. These four themes are compared and contrasted against existing literature on spiritual leadership.\nFindings: - It is found that beneficence, courage, and hope are comparable to existing western conceptualizations, whereas ubuntu is unique to the African context.\nResearch limitations/implications: - The paper and the larger study were derived from interviews with 16 participants; as such, generalization was not a goal. The paper provides a deeper understanding of spiritual leadership as enacted by African women, with implications for the need for increased research on non-western, non-white perspectives on the phenomenon.\nPractical implications: - Readers may relate to the women's stories and be both informed and inspired towards their own social justice leadership.\nOriginality/value: - Whereas the paper is derived from field research conducted in 2005 and published variously in other sources as cited, this paper takes an original perspective in comparing and contrasting African and western understandings of spiritual leadership, and expanding the understanding of the same in a novel way not done in previous publications.\n",
"role": "assistant"
}
] |
Physical and cognitive demands of work in building construction
|
[
"Management",
"Risk management",
"Construction",
"Occupational health",
"Interview",
"Questionnaire survey",
"Job analysis",
"Cognitive demands",
"Physical demands",
"Cognitive ergonomics",
"Construction work"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: 1.1 Context of the study\n2. Methods: 2.1 Study group\n3. Results: As the aim of this study was to assess physical and cognitive demands in construction work, we present the results in the corresponding order. In addition, we present results on stress symptoms in construction workers and their association with these demands.\n4. Discussion: The aim of this study was to assess physical and cognitive demands in construction work. Results from interviews (n=35) and a survey (n=30) in three construction companies in Germany suggest that the extent of physical demands is high but not very high and the extent of cognitive demands is also in a moderately high range, especially for verbal and information processing-related demands. In this sense, we cannot confirm our hypothesis that physical demands are higher than cognitive demands. Major aspects driving cognitive demands that construction workers face are updating and using job-related knowledge, in addition to giving advice and consultation. In each construction site, workers encounter different settings, buildings and construction plans and they are expected to adapt to each of the different new work sites accordingly. New innovations, technological developments and new regulations concerning materials, safety and building construction need to be adapted to existing work processes (Sexton and Barrett, 2003; Kaplinski, 2018; Rutesic et al., 2015) and this might lead to additional cognitive demands. Another underestimated component of a construction worker's job seems to be consultation. Consultation is a service that workers provide when they inspect a building, make decisions of the relevant work steps and identify risks and problems instantaneously - a service that is, however, not frequently taken into account. We were not able to identify any scientific publication dealing with this issue, suggesting a considerable research gap. A practical implication of this finding is that construction workers in small-sized companies seem to provide a service that is not acknowledged. If this is part of the regular project delivery system, this service needs to be included in the planning of the construction site.\n5. Conclusion: The aim of this study was to assess physical and cognitive demands in construction work and outline the amount and extent of these demands that workers are faced with on a daily basis. The results shows that construction work is demanding in physical terms as well as in perceptual, psychomotor, social and cognitive terms. Using and updating specialized knowledge, giving advice and providing consultation, friendliness, assertiveness and reliability are underestimated demands in the everyday work environment of construction workers. However, these demands are relevant to the workers' productivity and well-being. Further studies need to investigate the details of the relationship between the multi-component demands profile in construction work and the effects that it has on mental health. A practical implication is that those working in the construction industry must gain awareness of these different types of demands. Managers should recognize demands in the cognitive and psychosocial domain when planning work processes and when shaping the organizational structure in the construction industry. A lack of consideration can lead to an overload of the individual worker or to constraints in the time schedule that might collapse of the time management of the project. In the long run, a lack of consideration may also lead to a mismatch between construction workers' performance and their health in the future. As employers have to protect workers from the risks arising from construction work, it is important to include cognitive demands in the occupational risk assessment as required by law (Herbig et al., 2012). Yet, at present, there is no standard application of this in the construction sector. An implication for public policy is therefore to provide resource materials that help small- and medium-sized companies to realize that. Occupational risk assessments of this type are constrained by the fact that certain methods, such as the cognitive task analysis, are limited and do not acknowledge the full range of cognitive and psychosocial demands. We strongly encourage those working in the construction industry to get together with researchers to develop urgently needed measures that include mental, social and cognitive demands in risk assessments, especially given the increasing digitization of the industry. Increasing levels of demands for construction workers in small companies (e.g. digital skills, expertise in standards and regulations and provided consultation services) can also have economic implications, as skilled labor valuable and costs in building construction may rise.\n",
"role": "user"
},
{
"content": "Purpose: Investigating demands within an occupational context has always been an essential endeavor to guarantee worker well-being and performance efficiency. In blue-collar occupations, the physical demands of manual labor have always been the major focus, but recent technological changes may come with higher demands on the intellectual capacities of workers. For this reason, the purpose of this paper is to assess physical and cognitive demands that construction workers face.\nDesign/methodology/approach: The authors conducted a quantitative study using standardized research instruments. Construction workers of three German construction companies were asked to volunteer to participate in interviews (n=35) and a survey (n=30) that assessed the extent of physical demands and a variety of cognitive demands experienced by construction workers.\nFindings: The results suggest that construction work is demanding in physical terms as well as in perceptual, psychomotor, social and cognitive terms. Using and updating specialized knowledge, giving advice and providing consultation, friendliness, assertiveness and reliability are important demands among construction workers.\nResearch limitations/implications: Construction workers face an undervalued level of demands in their everyday work environment. As high demands can affect well-being and mental health, construction companies would greatly benefit from investing efforts into evaluating the multi-component demands profile of their workers and the impact on their health, in particular through the strongly increasing cognitive demands due to the ongoing digitization of the sector.\nOriginality/value: The study identified essential demands in construction work that are relevant for the workers' productivity and well-being.\n",
"role": "assistant"
}
] |
Developing social intelligence among employees for effectively managing organizational change
|
[
"Social intelligence",
"Attitude towards organizational change",
"Readiness for change",
"Organizational change"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In a highly aggressive and complex business environment, organizations must evolve and continuously adapt to the changing dynamics of the business world in order to be sustainable. Change in an organization affects employees' attitudes, behaviors and beliefs. Employees are the driving force during organizational change as they are responsible for implementing changes their skills and competencies are likely to play a major role in bringing about any change in an organization. According to Cherniss (2000), two-thirds of competencies leading to superior performance in the workplace are emotional or social in nature. Emotional intelligence has been a buzzword in business organizations for more than two decades. It is a component of a larger but less popular form of intelligence called social intelligence (Salovey and Mayer, 1990). Social intelligence enables individuals to understand other person's behavior, moods, mannerisms, emotions and body language. It centers \"on people's interpersonal awareness and social facility, their ability or skill to deal with social relationships effectively, co-operate and collaborate with others, and create and participate in healthy, positive and caring social interactions\" (Gkonou and Mercer, 2017, p. 10). It encompasses the ability to understand socially acceptable behaviour patterns.\nResearch Design and Methods: A questionnaire based survey was conducted to collect data from employees of information technology, banking and financial services, and hospitality industries in India. The sample consisted of 450 respondents after removing incomplete cases and outliers. Social intelligence was measured by Tromso Social Intelligence Scale (Silvera et al., 2001). Developed in Scandinavia, it is the most widely used scale for measuring social intelligence. This scale measures three important aspects of social intelligence, namely - social information processing, social skills and social awareness. Attitudes toward change were measured by readiness to change scale (Holt et al., 2007). This scale measures the following attitudes toward change - appropriateness, management support, change self-efficacy and personal valence. The responses for both the scales were taken on a five point Likert scale (1- strongly disagree to 5- strongly agree). Regression analysis was used to find the relationship between social intelligence and employees' attitudes toward organizational change.\nResults and Discussion: The finding suggests that social intelligence has a significant and positive impact on attitudes toward organizational change (r = 0.421, b = 0.421, p < 0.01). Based on both the questionnaires used for the study, the results reveal that socially intelligent employees are good at entering new situations and can handle uncertainty with ease. Such employees will accept change easily without being nervous, skeptical or anxious. Socially intelligent employees can understand the context and need for change thereby viewing change as beneficial for them as well as the organization at large. Employees with high degree of social intelligence are able to understand other people's behaviour and wishes, this ability will enable them to understand the need for change and senior management's commitment toward it. In the present day team based work culture, socially intelligent employees have the ability to develop strong interpersonal relations with their colleagues and managers. This would facilitate them to share their change related fears, anxieties and doubts with their managers. As a result, managers would be able to appraise resistance to change among employees. As socially intelligent employees can easily get along well with others they will not face issues in adjusting to work when change is brought about in an organization. During an organizational change, employees high in social intelligence can adapt their behaviour according to social setting. They exhibit confidence, positive behaviors and emotions during organizational change.\nConclusion: Social intelligence is an important skill in a workplace setting. It should be an integral part of the curriculum at management schools so as to acquaint and train the future managers in this important skill. Certain organizational interventions can aid in developing of social intelligence among employees. Interventions like social skills training will help employees improve interpersonal relations. Social sensitivity training will enable to sensitize employees by developing their empathy. Communication skills training leads to improved verbal and non-verbal communication of the employees. Besides these training courses, employees themselves can make certain positive behavioral changes to be more socially intelligent. Employees should:\n",
"role": "user"
},
{
"content": "Purpose: The objective of this paper is to study the impact of social intelligence on attitude toward organizational change.\nDesign/methodology/approach: This paper is based on an empirical study conducted across information technology industry, banking and financial services industry and hospitality industry in India.\nFindings: This paper shows the positive impact of social intelligence on attitude toward organizational change. Further, it provides various recommendations to develop social intelligence of employees in business organizations.\nOriginality/value: This paper provides insight into how social intelligence plays significant role in managing attitudes toward organizational change.\n",
"role": "assistant"
}
] |
A study on the multidimensional information management capability of knowledge workers
|
[
"PLS",
"Survey",
"Information management",
"Knowledge management",
"Questionnaire",
"Information processing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Knowledge management (KM) starts from the effectiveness of personal information management of knowledge workers in an organization (Grant, 1996a, b; Tsoukas, 1996). The synergetic process of managing personalized information related to facts, procedures, concepts, interpretations, ideas, observations, and judgment, among individual, groups, and inter-groups in an organization is known as KM (Alavi and Leidner, 2001). The knowledge worker brings in new knowledge from the outside, translates this to useful information, and ultimately advances organizational performance in the KM processes. Research on identifying and leveraging the collective knowledge in an organization to help the organization to compete (von Krogh, 1998) has made significant efforts to explain the relationships between various KM organizational interventions and performance (Sambamurthy and Subramani, 2005). However, there is no specific information management capability (IMC) constructs for knowledge worker.\n2. Literature review: Knowledge workers are the ultimate agents who put information to use (Orna, 1996). Janz and Prasarnphanich (2003) clarified the relationships among organizational climate, the level of cooperative learning that takes place between knowledge workers, and the resulting level of knowledge created by team performance and individual satisfaction levels. The definitions of data, information, and knowledge have been suggested and debated by researchers in many fields, such as IS, strategy, and communication. Data are raw facts about events; information is the data that are processed (e.g. classified, summarized, and transferred) to add meaning and value within certain context (Grover and Davenport, 2001; Martz and Shepherd, 2003). Knowledge has been argued based on different point of views (e.g. Tuomi, 1999; Langefors, 1973), but this paper adopts the definition of knowledge as \"justified belief\", and recognizes it as a primary strategic resource which can be transferred, recombined, and used to create value in organizations (Leonard-Barton, 1992; Nonaka, 1994; Kogut and Zander, 1992; Grant, 1996a). Saberwal and Becerra-Fernandez (2003) proved that internalization and externalization impact perceived effectiveness of individual-level KM. Their results also supported the expected upward impact in perceived effectiveness of KM, from individual to group level, as well as from group to organizational level. Ko et al. (2005) also posit that knowledge transfer is influenced by knowledge-related, motivational, and communication-related factors.\n3. Research hypotheses: 3.1. Sensing information\n4. Research method: 4.1. Measure development\n5. Discussion: 5.1. Summary\n",
"role": "user"
},
{
"content": "Purpose: Drawing upon the extant literature from information systems (IS), information science, psychology, marketing, management, and IT training, the purpose of this paper is to propose information management capability (IMC) construct and its sub-dimensions.\nDesign/methodology/approach: New instruments were developed and validated to measure the proposed IMC constructs. The proposed model was empirically tested using the data collected from 120 knowledge workers using SPSS and partial least square.\nFindings: This research proposes and confirms that IMC has formative sub-dimensions such as sensing, collecting, organizing, processing, and maintaining.\nOriginality/value: The study findings provide important insights on enhancing knowledge workers' information management practices and subsequent knowledge management practices. The new instruments can be used as diagnostic tools for knowledge workers' recruiting, ongoing assessment, and training.\n",
"role": "assistant"
}
] |
Hedge fund performance and managerial social capital
|
[
"Hedging",
"Fund management",
"Social capital",
"Human capital",
"Alternate investment markets"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Hedge funds around the world have an estimated $1 trillion in assets from $50 billion in 1990 and the number of hedge funds has climbed to nearly 6,000 from 300 in 1990 (Bernstein, 2002). The term \"hedge fund\" is historically rooted and has evolved over time to include a multitude of skill-bases investment strategies with a broad range of risk and return objectives. The common element among these strategies is the use of investment and management skills to seek positive returns regardless of market direction (Fuss and Herrmann, 2005; de Brouwer, 2001; Lhabitant, 2002). The hedge funds are usually in the form of a limited partnership or a limited liability company that employs speculative techniques in the hope of obtaining large capital gains due to comparative freedom from restrictions which give them risk-and-return characteristics that are very different from those of traditional investment funds.The investment strategies of hedge funds tend to be quite different from those followed by conventional funds. Every fund embarks on its own preferred strategy, which leads to a very heterogeneous asset class \"hedge funds\". Fuss and Herrmann (2005) provides a classification of three broader strategies: relative value, the event driven and the opportunistic strategy. Hedge funds pursue these strategies by taking long and short position in assets, buying or selling derivatives, and use significant amount of financial leverage. In addition, hedge funds tend to invest in illiquid instruments (under researched assets trading in comparatively inefficient markets), such as private equity, restricted stock, and bank loans (Horwitz, 2004). These funds attract wealthy individuals and their fee structure consists of a management fee varying from 1 to 8 percent and an incentive or performance fee of 15-25 percent (Fung and Hsieh, 1999). The trading profits, can lead to very large rewards for successful managers. The most apparent benefit of performance-related fees is that they align the interests of the investment manager and the client. For the investment manager, the ability to derive profits directly from investment returns is an incentive to hire the most successful managers, who in turn are attracted by the prospects of sharing potentially large profits (Cowell, 2003).The empirical evidence related to the hedge funds have centred on comparing the time series properties of hedge fund indices with returns on traditional financial assets (see, e.g. Kat and Brooks, 2002). Similarly, some studies (see, e.g. Edwards and Caglayan, 2001; Ackerman et al., 1999; Fung and Hsieh, 1997) have examined the performance and persistence of different hedge funds using single or multiple-factor model in capital asset pricing model (CAPM) framework. Getmansky (2004) argues that fund- and strategy-specific factors impact the performance of the funds. In this regard, some researchers have investigated the relationship between size of fund and fund performance, for instance, some studies (see, e.g. Clark, 2003; Herzberg and Mozes, 2003) conclude that no significant difference can be found between small and large funds, but Hedges (2003) found that smaller funds outperform large funds. Ammann and Moerth (2005) analyzed the impact of fund sizes with respect not only to returns, but also to volatility measures, Sharpe ratio and alphas derived from a multi-asset class model. They found that larger funds have lower volatilities compared to smaller funds. There is a negative relationship between fund size and returns. Using the classification of hedge funds, Harri and Brorsen (2004) found that market neutral and fund of fund (FOF) styles show the greatest persistence of performance compared to event driven, global and global macro styles.Conversely, numerous studies have explored the survival and lifetime of the hedge funds (see, e.g. Baquero et al., 2002). For instance, Gregoriou (2002) showed that certain hedge funds classifications experience higher survival times than others, and Baquero et al. (2002) attributed the fall of funds to higher leverage. To the best of the authors' knowledge, there is no evidence relating human knowledge factors to hedge fund performance with some exceptions (Do et al., 2005; Agarwal and Naik, 2000; Brown et al., 2001).This study is motivated by lack of empirical evidence relating human factors, in particular, the impact of social capital (SC), on the subsequent hedge fund performance. For hedge funds, capital is not the scarce resource - talent is. One of the biggest challenge facing hedge funds is identifying, attracting and retaining talent. With so many different asset classes being managed by a same fund, the \"eat what you kill\" mentality is being replaced with a more team oriented structure that is designated to motivate portfolio managers to play together nicely and share their toys, thereby leveraging a firm's resources or core competences and helping to contribute to the greater good of the firm (Brille and Gerson, 2006).Our study contributes to the recent literature on the role of human capital on organisational performance. In the finance literature, previous work such of Golec (1996) and Gottesman and Morey (2006) relating manager education to mutual fund performance found that manager with MBAs have better performance than fund managers without MBAs. Chevalier and Ellison (1999) found that fund manager performance is significantly and positively related to the mean composite SAT scores of the institutions at which fund managers completed their undergraduate degree. We argue that hedge funds require team members to develop skill-based investment strategies for the clients that lead ultimately to fund performance. Since hedge fund managers cannot market easily and widely market their services due to the regulatory constraints on the number of accounts, which means that managers must focus on the number of selected potential client. This further underscores the importance of attracting right people to deliver the performance. The strategies require that hedge funds to capture, harvest, share, and utilise their employees' competencies.\nSC perspective: In the recent years, many authors expressing the view that the modern world is becoming a \"knowledge society\" have discussed the importance of organisational-level human and SC for organisational performance (Pennings et al., 1998). While human capital refers to individual ability, SC refers to opportunity. Some portion of the value a manager adds to a firm is his or her ability to coordinate other people: identifying opportunities to add value within an organisation and getting the right people together to develop the opportunities. Managers with more SC get higher returns to their human capital because they are positioned to identify and develop more rewarding opportunities (Burt, 1997). Thus, SC is a private good that varies depending on an individual position and positioning strategies, and not befitting the public good status.Chou (2006) indicates three levels of SC. Micro-level SC consists of \"bonding\" and \"bridging\" SC. The former refers to relations between family members, close friends and neighbours, the latter to more distant associates and colleagues who may have different demographic characteristics. At the meso-level SC refers to vertical as well as horizontal association and behaviour within and among other entities, such as firms. Lastly, macro-level includes the most formalised relationships and structures, such as the political regimes, the rule of law, the court system, and civil and political liberties. However, in this article, our focus is on the micro-level and meso-level SC.Many researchers have utilised the concept of SC to help explain organisation performance. For example, in the product innovation literature, close and frequent interaction between teams lead to project effectiveness (e.g. Szulanski, 1996; Eisenhardt and Tabrizi, 1995). Similarly, there has been an explosion in research exploring how the composition or demography of groups may affect group process and performance. For instance, in the entrepreneur research literature, the studies have begun to address how the composition characteristics of founding teams affect performance (Amason et al., 2006; Chandler and Hanks, 2005; Chowdry, 2005). In a similar vein, in the knowledge management literature, focus has been on the organisation structure and processes that predict performance. Nahapiet and Ghoshal (1998) suggested that an organisation can provide an institutional environment conducive for the development of SC. They further recommended that the combination and exchange of knowledge can be facilitated when individuals are connected together (structural capital); individuals have strong and positive relationships with one another (relational capital); and individuals have the capability to understand and apply the knowledge (cognitive capital). We will argue in this article that these dimensions are much closer to first two levels of SC defined by Chou (2006). In the next section, we will discuss these dimensions of SC and investigate how these dimension impact the hedge fund performance.Structural capital\nImpact of SC on hedge fund performance: We conceptualise career and social networking histories of individuals as bonding and bridging devices that are likely to affect the hedge fund growth and performance. The SC of a hedge fund team affects its growth through previous strong ties, a set of experiences and contacts that can benefit the new firm. Hedge fund managers bring a great deal of tacit knowledge with them from their prior employment engagements. We refer to this as the influence of the previous affiliation about how to organise and manage work process, and this knowledge is likely to differ even between two funds in a similar strategy style.Similarly, hedge fund managers develop social networks through their affiliation with other organisations. This networking allows the hedge fund managers to seek new investors in cordial environment[1] and opportunities for networking at social events[2]. These diverse sets of affiliations imply access to high net worth individuals who meet certain financial investment criteria. As a result of networking with potential clients, the hedge fund managers take responsibility of managing their private funds once managed by their own family office or a trust fund. In terms of portfolio diversification theory, accessing hedge fund provides these high net worth individuals potential advantage of risk diversification.Sample and methodology\nResults: Table III provides statistics on the hedge funds performance by using hedge fund originator characteristics such as education, age of fund manager at start of business, previous affiliation (AFF1) and affiliation diversity (AFF2) respectively.Table III shows that managers with education achievement at Master-level have the highest annualised rate of return and Sharpe ratio (0.9836). This results is similar to the findings for a mutual fund manager with Master-level education (see, e.g. Golec, 1996; Gottesman and Morey, 2006) The young hedge fund managers have relatively lower return (23.7876) compared to old managers (25.3214), while, the Sharp ratio is relatively higher for the young managers. The fund managers with one previous career affiliation have lower Sharpe ratio (0.5094) compared to managers with more than three previous career affiliations (0.8745). This result might imply that with more career affiliations, hedge fund manager capitalise from knowledge base and learning about market which affects the risk attitude of the hedge fund manager. On the other hand, in terms of manager's relational capital measured by affiliation diversity, managers with three or more affiliation diversity have higher annualised rate of return compared to managers with only one previous or current affiliation. This result seems to suggest managers with more social networks and affiliation should have access to market niche of wealthy investors to increase their investor base.The hedge funds have been promoted as \"absolute returns\" investments because they do not have direct benchmark. We compared the excess return of hedge funds using two different benchmarks i.e. 30-day US Treasury bill rate and a five-factor model (see equation (1)). The hedge funds' absolute excess returns are larger in magnitude using 30-day US Treasury bill rate as benchmark compared to the excess return measured by five-factor alpha obtained from the five factor model. This finding is similar to previous findings of Edwards and Caglayan (2001). The large differences between alpha reported in this article and in previous studies might be due to the number of factors (five in this article) used in estimation of alpha.Effect of SC on hedge fund performance\nConclusions: This article empirically tested the relationship between hedge fund performance and managerial SC using top 25 hedge funds. We used several indicators of hedge fund performance and related them to hedge funds managers' demographics and SC. Our results show that, career and social networking of managers does influence the hedge fund growth and performance. In general, our results provide new evidence for hedge funds in that, managerial skills matters and their unique networking and cognitive skills unearth mystical performance of these funds in recent years. In order to be competitive and innovative in retaining and targeting wealthy clientele, hedge funds managers need to capture, harvest, share, and utilise their employees' competencies.We propose that examining the influence of the diversity of the hedge fund teams and incentives provided to team members in the future extension of our work. For example, selection of individual hedge fund team member: how does individual member complement one another in a portfolio, how additional managers add diversification in terms of geographic and sector focus. There is also extensive scope of research into the ownership culture of hedge funds. The alignment of interests of hedge fund team members and investors within the hedge fund team has become more important, and some funds such as Greenwich-based Lone Pine Capital, has embarked on creating ownership culture whereby equity stakes in the fund are given to the employees. Furthermore, taking into account the use of diverse set of propriety technologies and marketing strategies by hedge funds, future research can explore how hedge fund managers use marketing strategies to increase their market shares.\n",
"role": "user"
},
{
"content": "Purpose: - This article seeks to explain and empirically test the relationship between managerial social capital and hedge fund performance.\nDesign/methodology/approach: - This article uses a capital asset pricing model (CAPM)-style five factor model to estimate excess returns for the top 25 hedge funds.\nFindings: - The results show that hedge funds managers with more affiliation diversity have higher annualised rate of return. This result seems to suggest managers with more social networks and affiliation have access to market niche of wealthy investors to increase their investor base. Hedge fund managers' prior skills sets and repertoire of knowledge significantly influence their risk taking attitude.\nResearch limitations/implications: - The sample consists of the top 25 hedge funds.\nOriginality/value: - The article discusses the role and implications of managerial social capital in hedge fund marketing and performance.\n",
"role": "assistant"
}
] |
The moderating effects of gender and income between leadership and quality of work life (QWL)
|
[
"Gender",
"Leadership styles",
"Income",
"Quality of work life (QWL)"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: A manager's leadership style has been recognized as a key component of organizational success. Thus, managers, by applying appropriate leadership styles, can enhance employees' job satisfaction, commitment and productivity (Kozak and Uca, 2008; Rani et al., 2011). For example, good leaders are more likely to encourage employees to have ownership of tasks and make decisions to enhance the team and company performance (Bennett, 2009). Accordingly, a number of studies have supported that a manager's use of different leadership styles influences employees' outcomes, such as emotional intelligence (Downey et al., 2006), job satisfaction (Saleem, 2015) and the quality of work life (QWL) (Kara et al., 2013).\n2. Conceptual development: 2.1 Leadership style\n3. Methods: 3.1 Data and sampling\n4. Results: Regarding the demographic composition of the respondents (Table I), 61.2 per cent were single. Nearly three times more men (N = 323) responded to the survey compared to women (N = 120), reflecting the male and female composition ratio of hotel and tourism industry workforces in Turkey. Overall, the respondents are aged 20-54, with a mean 32 years old, and 36.1 per cent of the respondents earned monthly income of 1,501-2,500 Turkish Lira. In total, 62.5 per cent of the respondents were department managers and employees. Most respondents were employed full-time (81.0 per cent). In terms of employees' work departments, 11.7 per cent of the respondents worked in front office, 14.9 per cent worked in food and beverage, 20.8 per cent in housekeeping, 10.2 per cent in accounting, 10.4 per cent in public relations, 7.7 per cent in sales and marketing, 13.1 per cent in human resources and 8.6 per cent in other departments. In total, 47.9 per cent of the respondents had a high school education level. The data obtained from the Turkish Ministry of Culture and Tourism (2009) confirmed that the overall demographics of the respondents properly represent the five-star hotel employee population.\n5. Discussion and conclusions: 5.1 Conclusions and theoretical implications\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to explore the moderating effects of gender and income on the relationship between leadership style and quality of work life (QWL). This study provides meaningful implications for the hospitality industry in terms of gender and income between leadership and QWL.\nDesign/methodology/approach: Data are collected from five-star hotel employees (n = 443) in Turkey. The hypotheses are tested using hierarchical linear regression. The independent and dependent variables used to test the hypotheses involving the dependent variable of QWL are centered prior to the empirical analysis to avoid potential multicollinearity.\nFindings: Transformational and transactional leadership styles were significant predictors of QWL, controlling for the demographic variables (i.e. age, employment statistics and education level), but gender and income were not significant antecedents of QWL. When it comes to the interaction effect of leadership styles and gender, gender showed a statistically significant moderating effect between transformational leadership and QWL, but not between transactional leadership and QWL. Income had a statistically significant moderating effect between both leadership styles and QWL.\nPractical implications: The findings of the study potentially affect hotel management by identifying the moderating effect of gender and income of the employees and demonstrating how quality of life of the employees can be improved by leadership styles of managers.\nOriginality/value: Previous literature has addressed the issue of leadership and its outcomes. However, there has been limited research on examining the relationship among gender, income, leadership style and QWL in the case of hospitality management.\n",
"role": "assistant"
}
] |
The dilemma of internal audit function adaptation: The impact of ERP and corporate governance pressures
|
[
"Institutional theory",
"Corporate governance",
"ERP",
"Internal audit",
"Conceptual framework"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The expanding opportunities for the internal audit function (IAF) have been highlighted through the increased interest in better corporate governance (CG) practices (Ebaid, 2011). This claim is supported by real-life cases which elucidate clearly the implications of underestimating the role that the IAF can and should play (Boyle et al., 2012). Moreover, CG creates heated debate since noticeable failures of accountability have been highlighted by the financial scandals and the economic crisis. Certainly, the IAF has attracted the attention of boards those face external pressure for assurance about CG practices. Internal auditing is receiving attention unlike ever before (Carcello et al., 2005). For over a decade, the global emphasis on the need for sound CG triggered an increased interest in the IAF's legitimacy (Radu, 2012; Ebaid, 2011; Goodwin and Yeo, 2001).\n2. Research background: ERP systems pose new risks and challenges to internal control; triggering changes to the IAF's structure and practices. Throughout the following four sections, the authors explore how ERP systems implementation acts as a catalyst for the IAF change, by affecting the internal control and risks within organisations.\n3. Research gaps: The insights gained to date have a number of limitations that constitute the research gap that the authors will address in the proposed framework:\n4. Conceptual framework: The authors propose a conceptual framework for analysing IAF adaptation for the introduction of ERP systems in the CG context. The framework is intended to explore the changes in the IAF associated with ERP systems implementation and to derive conclusions that support policy makers, auditors, managers and organisations. To achieve these aims, institutional theory (Scott, 2008; Mignerat and Rivard, 2009; Gosain, 2004; Oliver, 1991; Powell and DiMaggio, 1991) was found to be a suitable theoretical lens as the study focuses on translating ERP inscribed logic in the organisational context where governance institutional arrangements related to the IAF exist.\n5. Conclusion: Internal auditing is the outcome of complex conjunction of external and internal associated constituents. It is this constellation of connected elements that justifies a function-level approach to the analysis of the IAF change. The IAF change process can therefore be viewed as a response to the new ERP systems. ERP systems constrain the activities of the organisation and shape the cognitive processes of internal auditors. Consequently, it constrains CG's institutional forces that influence specific trajectories of the configuration of the IAF. Between these two main aspects is the resolution process of institutional misalignments through the strategic response to the institutional pressures.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to provide a conceptual framework that helps to investigate how the internal audit function (IAF) responds to both the introduction of the control logic of Enterprise Resource Planning (ERP) systems, and corporate governance's (CG) institutional pressures. Furthermore, the paper aims to articulate the concurrence between the external pressures of CG and internal control logic of ERP systems.\nDesign/methodology/approach: The paper presents a review of the normative literature pertaining to the increase in significance of CG in the light of the worldwide economic crisis. The paper highlights a literature gap related to the lack of studies focusing on the impact of ERP systems implementation on the IAF practices.\nFindings: The authors articulate institutional theory to formulate a conceptual framework that explains the reciprocal interplay between the macro external governance pressures, micro internal institutional logics inscribed in the ERP systems and their effect on IAF practices and structure within organisations.\nResearch limitations/implications: The paper is conceptual in nature and therefore the proposed framework will be subsequently validated using a qualitative research approach in future research.\nPractical implications: The conceptual framework would offer the internal auditors some strategies for enabling adaptation to the different internal and external pressures. Also the paper provides a platform for research community to investigate the influence of CG and ERP systems implementation on IAF adaptation.\nOriginality/value: The paper provides a clearer articulation of the various constructs that affect the IAF, which has gained great attention for assuring good CG.\n",
"role": "assistant"
}
] |
The impact of emotional intelligence on ethical judgment
|
[
"Emotional intelligence",
"Ethical judgement"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In today's competitive business environment, an ethical workplace culture protects company assets, provides emotional security and promotes effective decision making (Ventimiglio, 2015). Despite the benefits of ethical business practices, 20 years of surveys (Ethics Research Center, 2013) indicate consistent ethical challenges that organizations face covering a range of offenses from falsifying records to conflicts of interest to bribery and theft. The latest National Business Ethics Survey of the US Workforce (Ethics Research Center, 2013) reported that 60 percent of workplace misconduct involved someone with managerial responsibilities. Unethical business practices result in severely damaged company image, increased dysfunctional behavior and decreased value of the company (Meinert, 2017). Given the detrimental consequences of unethical decisions, it is important for organizations to foster an ethical workplace culture and promote ethical decision making (EDM).\nReview of literature: Emotional intelligence\nSummary and research questions: Although there have been a few research studies on the topics of emotions or EI and ethics, the extant literature has not explored the direct relationships between EI and specific moral philosophies resulting in EDM. Prior literature has called for more specific investigations of the relationships between these topics. In addition, previous research on cross-cultural comparisons of ethical perspectives has yielded inconclusive results.\nMethodology: Participants were MBA students in two business schools, one located in the USA and the other in India. The study initially included 109 students pursuing their MBA degree at the US business school and 90 students pursuing their MBA at the Indian business school. The final sample had 100 respondents, 77 from the USA and 23 from India, yielding a response rate of 50 percent. The mean age of the participants was 25 years old. The respondents were coded by number to match their responses to the two surveys used for collecting data on the independent variable of EI and the outcome variable of ethical judgment.\nResults: The model with the total EI score as a predictor of the relativism ethical perspective was significant (R2=0.13, F(3, 96)=4.6, p<0.01). An individual's composite EI score was a significant predictor of the relativism ethical perspective (b=0.21, t=2.012, p<0.05). In addition, the age of the person was significant (b=0.21, t=2.012, p<0.05).\nDiscussion: Individuals in ethical situations require the skills and abilities to make informed judgments in order to ethically address those situations. Emotions are a contributing factor to the individual's conceptualization and judgment in their EDM. Our study identifies novel findings between EI and ethical perspectives.\nImplications: The ethical behavior of leaders has been one of the most concerning and perpetual issues facing business and society. While calls for reform and closer attention to business practices have been made, there remains an urgent need to better understand the ethical behavior of decision makers and to determine the potential facilitators of their ethical judgment.\nLimitations and future research: Although this study showed that EI has an impact on the ethical judgment of individuals, one limitation is the sample group of MBA students, many who are yet to enter into corporate avenues. Future research can explore managers at different organizational levels, across industries, and across countries to understand the interplay between the variables. The moderating effect of national culture and organizational culture on the relation between EI and ethical judgments will be an interesting area for further investigation. Our study showed relationships between age and EI, and age and ethical judgment. Additional research in these domains would also throw light on the intricacies of these influences.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the important relationships between emotional intelligence (EI) and ethical decision making (EDM).\nDesign/methodology/approach: Participants were 100 students from MBA programs in the USA and India who completed two surveys: one measuring their EI, and the second their use of four different ethical perspectives in three scenarios. Multiple regression analyses were performed to discover relationships between overall EI, certain dimensions of EI and their ethical judgments.\nFindings: The authors' results found that the composite EI score as well as the EI subscale of decision making were both significantly related to the relativism ethical perspective. Age was an additional significant factor for EI and EDM.\nResearch limitations/implications: Future research should attempt to replicate these findings among different managerial levels, industries and countries to further understand the distinctive relationships between EI and ethical judgments.\nPractical implications: This study highlights the importance of integrating EI and ethical judgment within corporate training programs and business school curricula, as important is the emphasis on moving the subject of ethical judgment from one of awareness to sustained ethical behavior through accountability.\nOriginality/value: This study contributes to the literature by identifying how EI and its decision-making subscale are significant to EDM.\n",
"role": "assistant"
}
] |
Reconceptualising the principles of Penrose's (1959) theory and the resource based view of the firm: The generation of a new conceptual framework
|
[
"Penrose",
"RBV",
"Framework",
"Growth theory"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: It is generally accepted that small firms that grow contribute greatly to the economy (Acs, 2006; Autio, 2007, 2012; Morris, 2011; Ward and Rhodes, 2014), and the greatest contributors are high growth small firms (Autio, 2005; Bravo-Biosca and Westlake, 2009; Anyadike-Danes et al., 2013; Octopus report, 2015). This coupled with the fact that it has also been found that their impact on employment holds true under times of economic recession (Coad, 2009), means that focus into high growth small- to medium-sized enterprises (SMEs) by academic and policy makers has grown. However, despite this vast increase in research there has been a distinct lack of growth theory development, evidenced by the fact that the most popular and in-depth growth theory was developed over 50 years ago by Edith Penrose (1959). Since this time, development of firm growth theory has been limited, due to the complexity of the growth process itself. Add to this, the failure to adhere to similar methodological approaches and operational definitions and the result is a lack of theoretical development (Delmar et al., 2003) and an increase in fragmented theoretical application (McKelvie and Wiklund, 2010; Leitch et al., 2010). As a result, it has been suggested by Davidsson and Wiklund (2000) and Almarri and Gardiner (2014) that the study of firm growth requires consideration of a number of different theoretical perspectives in order to understand the growth process fully. Key theoretical perspectives include the theory of the growth of the firm (Penrose, 1959) and the resource-based view (RBV) (Wernerfelt, 1984). Coupled with a lack of theoretical development and cohesion, there is also a lack of empirical testing of many theoretical perspectives within the literature. For instance, Lockett et al. (2007) argues that there has been little empirical testing of Penrose's concepts, which may be considered surprising given the influence it has had on current academic literature. There are also many issues and knowledge gaps that have been highlighted with the RBV that requires further research (Priem and Butler, 2001; Newbert, 2007; Baker and Ahmad, 2010; Park, 2010). Thus, this research aims to address the identified gaps in the literature to create an integrated conceptual framework, generated through empirical research that incorporates new concepts whilst developing and testing existing theories. Through this the theories will be reconceptualised in a way in which they can be better used by practitioners, making the theoretical practical and able to be widely disseminated.\nSME growth literature: It is generally accepted that small firms which do grow contribute greatly to the economy (Acs, 2006; Autio, 2007, 2012), and the greatest contributors are high growth small firms (Autio, 2005; Anyadike-Danes et al., 2013). Holzland and Friesenbichler (2007) and Allman et al. (2011) argue that even though high growth firms play an essential role in economic development and employment, knowledge about them is extremely limited while Roper and Hart (2013, p. 11) argue that \"growth remains something of an enigma\". Barringer et al. (2005) suggest that gaining knowledge on the attributes of high growth firms will greatly inform our knowledge of the attributes of normal growth firms. Ultimately despite the recognised value of extant research, there remains a considerable knowledge gap as to the process by which SMEs achieve growth and how policy makers and government bodies can support this (Bennett, 2008, 2012; Lee et al., 2016). In an attempt to overcome this, Yandle (2014) created the growth platform model detailing 15 factors critical to the achievement of firm growth, their development over time and to what level they need to be developed to in order for growth to occur. This built upon Klofstens (2010) business platform model but for the start up to growth phase of development. The current research builds upon the growth platform model (Yandle, 2014) attempting to develop a theoretical framework to sit alongside it. This theoretical framework needs not only to be grounded in existing theories but also to merge and extend them. Forson et al. (2014) argue that a conceptual framework is required that is able to bring together the various silos of work that have taken place in the field of small business. As such, a review is now undertaken of the most dominant theoretical perspectives in the firm growth literature.\nTheoretical perspectives on firm growth: The theory of the growth of the firm (Penrose, 1959)\nMethodology: This paper utilises a data set generated during PhD research and builds upon the growth platform model generated from this (Yandle, 2014). The methods used to generate the model upon which this research builds and to generate the accompanying framework will now be presented. As the aim of this research was to explore theory then a methodological approach fitting of this purpose was required. Eisenhardt (1989) advocates case studies for theory building in areas, where existing theory appears to be inadequate. Due to the need for theory development, it was decided that a multiple case approach (Glaser and Strauss, 1967) would be most fitting. The data were generated through a process of semi-structured interviewing, participant observation (at one case company) and analysis of company documentation.\nData analysis: All interviews were transcribed and then all interviews, observations and company documentation were subjected to open coding, axial coding and hierarchical coding (Strauss, 1987). In total, 15 hierarchical categories emerged to explain the firm growth process and comparative analysis was then undertaken based around these 15 categories in order to ascertain if there were similarities or differences. Framework development occurred along the lines of developing a core category as is suggested by Strauss and Corbin (1998), through theoretical coding. This was enabled through the comparison of all the higher level codes. Three core categories were identified and although Strauss and Corbin advocate only one core category, this research involves so many variables that in order to explain the connection between them three was deemed necessary. For the purposes of this research, only the influences of each factor were analysed in order to ascertain how each combine in order to enable firm growth. Figure 1 summarises the data collection and analysis process.\nReconceptualising Penrose's theory and the RBV: The findings from this research confirm many of Penrose's findings but importantly add new insight to create a fresh perspective. It is confirmed that the firm is a bundle of productive resources which are controlled by the administrative hierarchy and which combine to produce goods for sale (Penrose, 1959), as this is exactly the explanation for the findings that is being posited here. The key new insight that this research provides is detailed information with regard to which resources, management and outputs are vitally important, why this is the case and how they develop over time. Although Penrose's theory has been highly influential in academic circles, it has not been able to be practically disseminated. This stems from the lack of clear identification of process and the lack of a coherent model stemming from the theory. This research, however, has added to this body of knowledge, providing the detail needed to explain and support Penrose's theory, clearly identifying the process of development. Penrose posits that firms can have differing amounts of resources, whereas this research argues that firms must have the correct mix of certain resources, developed to specific levels in order to achieve growth. Thus, this research provides more detail with regard to each factor found to have been of importance.\nDiscussion and conclusion: The aim of this research was to reconceptualise two dominant theoretical perspectives within the firm growth literature. This was accomplished through merging the theories and by conducting empirical research resulting in the generation of a new firm growth framework. As a result, this research addresses a distinct gap in the theoretical firm growth literature and makes a clear contribution to the literature. The results indicate that an explanation of the modern firm growth phenomenon requires a multi-theoretical approach. Crucially, the empirical research showed that the principles of the theory of the growth of the firm, the RBV and the dynamic capabilities perspective can be combined and reconceptualised in a new framework that is capable of providing a more specific explanation of the firm growth process. As has been accepted throughout the paper these findings are exploratory, providing an excellent starting point for further confirmatory research to be conducted on the concepts.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to develop a conceptual framework to explain the firm growth process based on an integration and extension, through empirical research, of Penrose's theory of the growth of the firm and the resource-based view. Theoretical development within the firm growth literature has been noticeably limited. Firm growth studies use different theoretical bases and what is needed is integration of multiple theories and empirical testing of these to form a new conceptual framework capable of explaining the modern growth process fully.\nDesign/methodology/approach: The key perspectives are critically reviewed and integrated and empirical qualitative research is undertaken analysing the process of growth in two firms. Semi-structured interviews, participant observation and analysis of company documentation are utilised.\nFindings: The key insight this research provides is detailed information with regard to which resources, mediators and outputs are vital to firm growth, how they need to be developed and why this is the case. The study shows that these act in a cyclical nature to enable firm growth and development.\nPractical implications: These findings could be used by practitioners to determine which part of the conceptual framework requires the most amount of improvement and which are developed to an acceptable state, enabling them to make plans for the achievement of growth.\nOriginality/value: This research is able to reconceptualise two dominant theoretical perspectives resulting in the generation of a new firm growth framework, thereby addressing a distinct gap in the firm growth literature.\n",
"role": "assistant"
}
] |
Localising the packaging of foreign food brands: a case of Muslim consumers in Pakistan
|
[
"Packaging",
"Hedonic and utilitarian consumption",
"Product standardization and adaptation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this issue.\nIntroduction: An ongoing debate in international marketing is whether marketing approaches should be localised as international brands enter foreign markets (Fan, 2002; Okazaki, 2004; Eckhardt, 2005; Fastoso and Whitelock, 2012; Jiang and Wei, 2012). As international marketers seek to find ways to influence consumer behaviour across cultures, the communicative role of packaging is becoming increasingly important. Yet, while many academics and practitioners have focused on international marketing strategies of foreign brands (Zhou and Hui, 2003; Samiee et al., 2004; Zhou and Wong, 2008; Ozsomer, 2012; De Mooij, 2013), hardly any empirical research has examined the effects of localising or standardising a brands' packaging. This study examines whether localising a foreign product's packaging would enhance its effectiveness in influencing consumer choice. Previous research has offered practical evidence to support international marketers strategising their advertising activities in culturally and linguistically different countries (Akaka and Alden, 2010; De Mooij, 2010). It is important, however, to determine similarly the effectiveness of packaging for foreign brands, as packaging is the last point of contact before actual buying.\nConceptual development: Product packaging\nMethodology: Pre-test\nExperimental design: The experiment used a 2 (product type: hedonic versus utilitarian) x 2 (packaging design: standard versus local) factorial design. To ensure external validity, we used popular brands that were readily available in Pakistani supermarkets. For hedonic products, the foreign brands were Coca-Cola for soft drinks and Lays for chips. Similarly, the utilitarian brands were Nido powdered milk and Nestle bottled water. As the only actual modification in packaging that we can see in the local market is by language, we maintained this manipulation only in the experiment to ensure external validity.\nResults: We first determined whether respondents were able to identify the foreign and local brands correctly. Wrong identification of the brand origin would render subsequent results and analyses meaningless. Results of a non-parametric test clearly indicated that participants were able to correctly identify foreign and local brands (see Table II).\nDiscussion and managerial implications: Past research has provided a good understanding of how local consumers respond to advertising cues. Our study extends this research stream by providing empirical evidence on how standard versus localised packaging of foreign products may influence consumers' perceptions and choice decisions of the foreign brands. To the best of our knowledge, this is the first study that explored the effectiveness of standardisation and localisation in a product-packaging context. It further illuminates the differences for products that differ in hedonic and utilitarian attributes. Hence, this study significantly contributes to theories of standardisation and localisation of brands and marketing communication activities by providing a new theme for research in packaging. A better understanding into how consumers perceive these packaging strategies would also help international marketers operating in local markets.\nLimitations and future research: Although this study offers valuable implications for researchers and practitioners, it has some limitations that merit attention. Firstly, the results that packaging matters for hedonic but not utilitarian products may also be biased by the student sample; hedonic consumption may be particularly important with younger consumers (Park and Lessig, 1977). Also, the sample of graduate students had a moderate level of English proficiency and most likely live in urban Pakistan. This is not representative of the Pakistani population, which includes segments of less educated consumers that live in vast rural areas. This difference in socio-economic status may question whether the attitudes and behavioural patterns observed in the study are generalisable across the whole population (Manrai et al., 2000). Future studies could examine the perception of these societal clusters based on their living standards and social classes.\n",
"role": "user"
},
{
"content": "Purpose: The aim of this study was to investigate the extent that consumers prefer the localised packaging over standard packaging, and how the differences may vary across different product types. An ongoing debate facing marketers is whether marketing approaches should be localised as international brands enter foreign markets. In practice, international brands often localise their packaging when sold in foreign markets. This research questions whether and under what conditions is this practice beneficial to foreign brands.\nDesign/methodology/approach: The experiment used a 2 (product type: hedonic versus utilitarian) x 2 (packaging design: standard versus local) factorial designs. Product type was within-subjects, and packaging design was between-subject to minimise learning effects. For each product type, two product categories were used.\nFindings: Overall, the results show that the role of packaging is more pertinent for hedonic than for utilitarian products. For hedonic products, participants preferred the standard packaging to the local packaging and brand likeability is also rated more positively in their standard package. However, there were generally no significant differences in rating between standard and localised packaging likeability and brand likeability for utilitarian products. The results for the choice decisions were similar to those for the likeability ratings across both product types.\nPractical implications: A better understanding into how consumers perceive these packaging strategies would help international marketers operating in local markets.\nOriginality/value: Although past studies on international marketing communications have investigated standardisation and localisation of messages in the context of advertising using foreign and local cues, none have examined this issue with packaging. This study also extends past research by examining the differential effects of localisation on hedonic versus utilitarian products.\n",
"role": "assistant"
}
] |
Logistics alliance management capabilities: where are they?
|
[
"Logistics management",
"Strategic alliances",
"Logistics (Business)",
"Alliance capabilities",
"Logistics service providers"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Alliances between supply chain entities, such as buyers, suppliers, third-party providers, and customers have grown in popularity and have been promoted as an important component of the corporate and competitive strategy of the involved firms. Logistics involves the management of a wide range of core activities for the spatial and temporal transfer of goods, services, and related information, as associated with order processes, inventory management, transportation, warehousing/handling, and facility network management (Bowersox et al., 2007). Recently, value-added services, supply chain management, and financial services have complemented these traditional logistics functions (Carbone and Stone, 2005; Delfmann et al., 2002). Logistics alliances gained attention due to the outsourcing trend of logistics activities in the 1980s and 1990s, which aimed to reduce costs, respond flexibly to demand fluctuations, and reduce capital investment (Daugherty, 2011; Frankel and Whipple, 1996b; Gunasekaran and Ngai, 2003). The importance of logistics alliances has continued to grow due to ongoing deregulation, globalization, and reduced time for the order fulfillment processes (Hertz, 2001). However, not all logistics alliances are successful. For example, according to Dittmann et al. (2010, p. 135), \"successful collaborative relationships between a firm and its core suppliers are still rare\"; this is reflected in the fact that almost 70 percent of strategic alliances fail (Zineldin and Bredenlow, 2003).\nThe dynamic alliance management capabilities construct: With regard to evolutionary economics (Zollo et al., 2002), and resource-based view reasoning (Eisenhardt and Martin, 2000; Teece, 2007), the strategic management literature has devoted a lot of attention to the construct of dynamic capabilities (Barreto, 2010). Extant studies provide strong conceptual and empirical support for the positive impact that a firm\nA LAMC model: In the general strategy field, the dynamic capabilities construct is specified to functional domains so that they can be more concretely analyzed, and more specific implications can be formulated. For example, dynamic capabilities in purchasing (Schlomer, 2011), and new product development (Eisenhardt and Martin, 2000) are identified and specific routines and adaptation mechanisms are forwarded. In order to leverage this potential and to yield more meaningful results for logistics alliance management, there is a need for a specific LAMC model that builds on and further specifies the general dynamic capabilities frameworks to the peculiarities of the logistics alliance context.\nConclusion and outlook: Alliances are a relatively well-studied phenomenon in the logistics field. However, few studies have attempted to link logistics alliance performance to the existence of specific organizational capabilities of the involved firms. This finding is surprising given the insights that capability-oriented studies have been able to achieve in the strategy field, and it suggests valuable potential for an application in the logistics domain. This paper has proposed an initial model of LAMC that provides a new perspective to the explanation of logistics alliance performance and success.\n",
"role": "user"
},
{
"content": "Purpose: Due to the continued high failure rate of logistics alliances, the purpose of this paper is to propose incorporating the widely neglected dynamic capabilities approach into the explanation of logistics alliance performance.\nDesign/methodology/approach: The authors identify specific characteristics of logistics alliances that are used to argumentatively develop a framework of logistics alliance management capabilities.\nFindings: We propose an initial framework of logistics alliance management capabilities, based on the dynamic capabilities approach of strategic management, with the aim of improving logistics alliance success.\nResearch limitations/implications: The presented framework offers a basis for a wider range of empirical studies. Qualitative researchers are encouraged to further specify and better understand the sub-processes that underlie logistics alliance management capabilities in different alliance settings. Quantitative studies could help to reliably assess the differences and performance implications of these capabilities.\nPractical implications: The authors\nOriginality/value: This paper contributes to the logistics literature by making a more general yet established construct accessible to logistics scholars and practitioners. The roots of the authors\n",
"role": "assistant"
}
] |
How managing the interaction created a system of learning in a wholesale distribution model
|
[
"Learning",
"Development strategies",
"Engagement and learning",
"Team development"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Interaction is a fundamental element of learning organization models. Senge (1990) described the concept of interaction as shared visions in his five disciplines. He described the shared vision as, \"sense of commitment in a group, by developing shared images of the future we seek to create\" (p. 6). Similarly, Watkins and Marsick's (1993) dimensions of a learning organization introduced us to the inquiry and dialogue imperative which promotes interaction as a promoting element of inquiry. In both of these models, the group's interaction fosters the system of learning.\nTraditional interactions in the wholesale channel: The traditional marketing model through wholesale distributors limits the interaction at all levels down the channel. The manufacturer interacts directly with the wholesaler to outline the products, promotions and other elements of the offer. The wholesaler in turn communicates with the selling outlets. And, finally, the selling outlets manage the direct end-consumer interaction. It is difficult for the key knowledge elements to align through all parties in a consistent manner. Envision how the message gets distorted in the game where a long line of people attempt to whisper the same phrase to the next person. However, in this game, at every subsequent level, the numbers of people involved multiply (Figure 1).\nChanging the traditional interactions: Step 1\nBenefits of the change: After opening the registration process for two months prior to the promotion, the manufacturer captured over 3,000 selling outlets linked to over 100 wholesale distributors. During the month prior to the promotional period, the manufacturer was able to create multiple knowledge transfer methods with both the wholesalers and the selling outlets. In all communications, the manufacturer started with general promotional knowledge. Additionally, the training extended into specific aspects related to each level of the channel. To ensure that the double-loop learning occurred as the network adapted, the training provided details for the selling outlets to communicate ideas back to the manufacturer and the wholesale distributors, thereby solidifying the importance of interactions.\nThe new interactions: After the promotion, the manufacturer had gained contact information for the entire channel. Across the channel, interactions were established and solidified to support a continued system of learning. And, the unique benefits for every member of the channel pulled the team together. Beyond the promotional period, the manufacturer continued interactions with members of the channel to promote continued learning in all directions via the new interactions.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to illustrate how a manufacturer improved the interaction within the channel to facilitate a system of learning between three subsequent customer levels.\nDesign/methodology/approach: The paper provides an example of how a manufacturer designed steps to change the traditional interactions between the customer levels and place themselves in the middle of a knowledge-sharing environment.\nFindings: To best ensure consistency in knowledge sharing across the multiple levels, the learning and development (L&D) practitioner should examine the interactions and identify the unique benefits to encourage all members to freely engage in the learning system.\nPractical implications: Managing the learning system allows the manufacturer to manage knowledge sharing and reinforce a consistent message.\nOriginality/value: The paper offers the steps undertaken and the benefits resulting from the improved interactions. The example provides the L&D professionals areas into which they can expand traditional learning environments.\n",
"role": "assistant"
}
] |
Teachers' job satisfaction at well-performing, historically disadvantaged schools: A quantitative research in the Gauteng Province, South Africa
|
[
"Hygiene factors",
"Education",
"Job satisfaction",
"Motivators",
"Historically disadvantaged",
"Workloads"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: One of the major challenges of the education system in South Africa is turning around underperforming secondary schools, which are exit points for learners into the job market and higher learning institutions. In fact, school improvement efforts should attempt to enable these schools to become consistently well-performing. Sadly, most underperforming schools are historically disadvantaged and are located in areas with poor socio-economic conditions (Gauteng Department of Education, 2002/2003). These schools are beset with numerous challenges that not only affect learner performance, but teachers' quality of work-life. However, despite poor socio-economic conditions, some of these schools perform well and consistently achieve National Senior Certificate (NSC) pass rates above the threshold of underperforming schools. This implies that such schools present, among other positive factors, satisfying job conditions for teachers. This supposition arises from numerous expert assertions that link job satisfaction with motivation and good performance of employees in organisations (Iwu et al., 2013; Saiti and Papadopoulos, 2015; Chun et al., 2019). Therefore, teachers' perceptions about their jobs reflect the importance of their job satisfaction.\n2. Literature review: The concept of job satisfaction has been studied widely, and still receives attention in current research. Various studies have put forward several definitions of the concept. There are, however, commonalities among the definitions. Spector (1997) defines job satisfaction as indicative of how people feel about different aspects of their jobs. Aziri (2011) defines job satisfaction as a feeling that comes as a product of perceptions that make material and psychological needs possible. Monyatsi (2012) points out that the subject of job satisfaction in different organisations has been dealt with using different variables, such as demographic characteristics, work, pay, relations with co-workers, promotion and others; and that factors such as education, experience, supervision, work and relations with co-workers may be responsible for the observed differences in job satisfaction. Saiti and Papadopoulos (2015, p. 75) assert that job satisfaction is essential because it is closely linked to working behaviours such as productivity and efficiency and they add an important feature of job satisfaction and state:Although job (professional) satisfaction is a complicated and dynamic social phenomenon, as it mainly has to do with unknown human behavior, it is a very crucial factor in the provision of high quality performance as it enhances the cohesion and the morale of people, and hence has a positive influence on an organization's (school's) output.\n3. Theoretical framework: Job satisfaction entails the perceptions of teachers about their jobs, which may be positive or negative. This study is based on Herzberg's well-known two-factor theory, which has been used extensively in job satisfaction studies. The theory propounds that job satisfaction is a function of two groups of factors, namely motivator and hygiene factors. According to this theory, motivators generate employee job satisfaction, while hygiene factors reduce job dissatisfaction, but do not lead to job satisfaction. In this regard, Herzberg et al. (1959, pp. 113-114) state the following:Among the factors of hygiene, when the factors deteriorate to a level below that which the employee considers acceptable, then job dissatisfaction ensues. However, the reverse does not hold true. When job context can be characterized as optimal, we will not get dissatisfaction but neither will we get much in the way of positive attitudes [...] but it is primarily the \"motivators\" that serve to bring about the kind of job satisfaction, the kind of improvement in performance that industry is seeking from its workforce.\n4. Research method: This was a quantitative study with a positivist paradigmatic orientation that took the view asserted by Check and Schutt (2012, p. 15), that scientific explanations are based on universal laws, and aim to \"measure the social world objectively and to predict human behaviour\". To this end, the study used a survey research design with Spector's (Spector, 1985, 1999) Job Satisfaction Survey (JSS) being used to collect data. Spector (1999) permits the use of the JSS free of charge for non-commercial, educational and research purposes. According to Spector (1999), the JSS is standardised and has been tested for reliability and validity. In addition, a pilot study was conducted with a sample of teachers (n=20), to determine the language suitability of the items. The collected data were statistically tested for reliability by calculating Cronbach's a for internal reliability based on the inter-item correlations (Salkind, 2004).\n5. Results: 5.1 Respondents' demographic profiles\n6. Discussion: From the results of this study, it is clear that conditions engendering job satisfaction go beyond assuming that, because of their good performance, teachers at well-performing schools are satisfied with their jobs. It is clear that various factors complement others while other factors regulate or moderate job dissatisfiers. First, the ambivalent level of teacher job satisfaction at well-performing, historically disadvantaged schools is understandable since teachers indicated different perceptions on individual job satisfaction dimensions. There was, for instance, given the mean scores, satisfaction with supervision, co-workers and the nature of work. This resonates with Saiti and Papadopoulos' (2015) finding that teachers are more satisfied with administration, colleagues and the nature of work and is also consistent with Woulfin and Weiner's (2017) finding that principals as supervisors are the initiators of change at schools. Monyatsi's (2012) findings showed that employees become more productive when they feel appreciated by their supervisors. Monyatsi (2012) also found that teachers cooperate and support each other, are hard-working and not boring, which makes their work environment interesting. This implies that teachers are satisfied with their colleagues. Lloyd (2013) found that most teachers generally have good collegial relationships. The work itself is Herzberg's motivator and, unsurprisingly, promotes the teacher's job satisfaction. Supervision and co-workers are Herzberg's hygiene factors and serve to regulate dissatisfaction. In this study, these two dimensions seem to be strong moderators of teachers' job dissatisfaction and can explain the reason for ambivalent levels of satisfaction rather than complete job dissatisfaction.\n7. Conclusion: Although the job satisfaction of teachers at well-performing, historically disadvantaged schools is ambivalent, there are areas from which they derive job satisfaction. Teachers are satisfied with supervision, co-workers and their work. This could explain why their schools are performing well in that, although supervision can result in the imposition of strict rules and work overloads, teachers enjoy working with learners and leading their schools' outstanding performances. Collegiality and teamwork can be seen as moderating factors regarding other areas for which teachers are dissatisfied. Contingent rewards, though ambivalent, are important for teachers' job satisfaction as seen in the subscale's strong relationship with total satisfaction.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the levels of the job satisfaction of teachers at historically disadvantaged secondary schools and to determine the correlation effects among job satisfaction dimensions as they relate to these teachers.\nDesign/methodology/approach: A quantitative survey design using Spector's Job Satisfaction Survey was used, with 1,035 teachers from 30 secondary schools in the Sedibeng and Johannesburg South districts of the Gauteng Department of Education in the Gauteng Province of South Africa. Overall, 738 usable questionnaires were returned.\nFindings: Teachers at well-performing, historically disadvantaged schools experience ambivalent job satisfaction levels. For this reason, they indicate satisfaction with supervision, co-workers and the nature of work; ambivalence with promotion, contingent rewards and communication; and no satisfaction with pay and operating conditions. Correlations between job satisfaction dimensions are significant. Their correlations indicate relationships that range between moderate and strong. While mostly indicating relationships of no practical effect, most of Herzberg's hygiene factors are projected as strong moderating factors of job dissatisfaction as seen in relationships between dimensions reflecting hygiene factors and total job satisfaction.\nOriginality/value: This study pioneers the discourse on teacher job satisfaction at historically disadvantaged secondary schools, which still experience apartheid legacies: poor socio-economic conditions of their communities in South Africa. Strikingly, they consistently perform well in the National Senior Certificate - the basic education exit point. Lessons for educational management and policy practice can be learnt from these secondary schools, including lessons for underperforming schools' leadership.\n",
"role": "assistant"
}
] |
The moderating role of individual-difference variables in compensation research
|
[
"Pay structures",
"Career development",
"Risk assessment"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: For many organizations, salespeople provide a unique opportunity for a sustainable competitive advantage. As boundary spanners, salespeople link the organization to its customers. They often have access to divergent information coming from important customer groups. Furthermore, they, as unique human resources, are less susceptible to imitation and more durable than other types of organizational resources (Barney, 1991). Given the important role salespeople play, motivation of the sales force is widely recognized as an essential component of organizational strategy (Pullins, 2001; Alonzo, 1998). In response, sales firms have created, implemented, and continue to adjust company policies and procedures in attempts to boost motivation levels among salespeople. Policies related to the compensation of the sales force provide a continued source of discussion surrounding the issue of motivation. In this research, we jointly consider the influence of compensation policy (set at the firm level) and individual-difference variables on salesperson motivation.\nTheoretical framework: In previous organizational research, agency theory has been widely applied when studying the pay-performance relationship (e.g. Bloom and Milkovich, 1998; Gerhart and Milkovich, 1990). A key limitation to the agency theory approach is that the agents are automatically assumed to be risk-averse; therefore, the full-range of employee attitudes and behaviors may not be captured (Bloom and Milkovich, 1998). More recently researchers in this area (e.g. Rynes and Gerhart, 2000) have suggested using motivational theories including expectancy theory to further our understanding of compensation programs.The expectancy theory framework proposes that an individual's motivation is comprised of three components:1. expectancy of performance or estimates of the probability that effort will affect performance;2. instrumentality or estimates of how strongly performance is believed to lead to reward; and3. valence for rewards or measures of an individual's desire for the reward.Ultimately, the three components combine to affect the salesperson's motivation (e.g. Cron et al., 1988; Walker et al., 1977).The majority of research based in the expectancy theory framework has focused on identifying variables that influence valences, expectancies, and/or instrumentalities. Most of this work specifically examines task or organization-level variables, rather than individual-level variables. Of the studies that have focused on individual-level variables, most focus solely on reward valences (e.g. Ingram and Bellenger, 1983; Churchill et al., 1979).We extend the past research by providing an examination of the role individual differences play in determining salesperson motivation across valences, expectancies, and instrumentalities, and by considering how individual differences between salespeople influence reactions to firm policy (in this case pay mix). We expect that individual-level variables, including the career stage and risk attitudes of the salesperson, will have a moderating impact on pay mix. Hypotheses concerning the moderating role these variables play in the compensation - motivation relationship are now discussed (see Figure 1).Career stage\nStatement of hypotheses: In the following discussion, we present study hypotheses. Our hypotheses are broken into two distinct sections. First, we focus on the interplay between the individual-level variables and company pay mix. The individual-level variables are expected to affect the relationship between pay mix and motivation in very different ways. For instance, we argue that career stage will interact with pay mix to affect expectancies and instrumentalities of the salespeople. Second, we argue that risk attitudes are expected to interact with pay mix to affect valences for pay increases.Career stage and pay mix\nEmpirical study: The purpose of this study is to test whether or not career stage and risk moderate the relationship between pay mix and the components of salesperson motivation (valence, expectancy, and instrumentality).Sample and procedure\nResults: Correlations, means, and standard deviations of study variables are provided in Table I. Results of the split group analysis and moderated regression analysis are presented in Table II and Table III respectively. A discussion of key findings follows.Results of the split group analysis (Table II) indicate that career stage moderates the influence of pay mix on expectancy and instrumentality estimates of the salespeople. As expected, findings show that a pay mix reflecting greater variable pay results in a significant negative effect on instrumentality estimates (b=-0.221, p<0.05) and expectancy estimates (b=-0.300, p<0.05) for salespeople in exploration. Conversely, variable pay has a significant positive influence on instrumentality and expectancy estimates for salespeople in establishment (b=0.240, p<0.05 and (b=0.288, p<0.05, respectively), as well as a positive significant influence on instrumentality and expectancy estimates for salespeople in maintenance (b=0.297, p<0.05 and (b=0.314, p<0.05, respectively). In addition, Chow tests measuring the differences between the regression coefficients for expectancy and instrumentality estimates were highly significant for the exploration versus establishment groups (F=0.206, p<0.01, and F=0.199, p<0.01 respectively) and for the exploration versus maintenance groups (F=0.271, p<0.01, and F=0.217, p<0.01 respectively).The plotted interaction effects in Figure 2 provide a consistent picture of the results provided above. The outcome scores were plotted at the mean of the proportion of variable pay (pay mix). One standard deviation below the mean represents less variable pay, while one standard deviation above the mean represents greater variable pay. In general, the slope was steeper for employees in the exploration phase. Analyses of the slopes confirm results of the split group analysis.Results of the moderated regression analysis indicate that the risk attitudes of the salesperson have a significant influence on the relationship between pay mix and valences for pay increases. Consistent with our hypotheses, the interaction between risk and pay mix was significant (b=0.48, p<0.01). The change in R2 when adding the interaction term to the equation was also significant (DR2=0.09, p<0.05).\nDiscussion and implications: Study results suggest that career stage and risk preferences impact the relationship between compensation and the three components of motivation. These results go past existing knowledge suggesting that type of compensation alone does not dictate the effectiveness of the program. Results reported here indicate that organizations need to be more \"strategic\" with regard to compensation (i.e. matching compensation and other variables including those related to the individual). Results suggest that characteristics of the individual salespeople are important when matching compensation. Not all salespeople are created equally. For some motivation will increase when paid based mostly on commission, while for others motivation will increase when paid a fixed salary.Past research concerning antecedents of compensation programs suggests a notable disconnect between current compensation practice and the findings of the present study. For instance, Krafft (1999) found high correlations between companies' compensation strategy and firm and corporate level variables. Krafft (1999) clearly points out that there is no correlation between compensation strategy (along with other control decisions, including extent of monitoring and direction) and individual-level variables. To the contrary, findings of the present study suggest that individual-level factors play an important role in determining the impact of compensation on the components of motivation.Thus, this study raises many issues that may be of interest to both managers and salespeople. At a minimum, managers could use this knowledge to improve their general decisions concerning compensation. Managers must acknowledge that, as individuals, certain salespeople respond positively to fixed salary plans while others respond positively to incentive. In the past, managers might have relied on the salesperson to select the appropriate job for them. That is, managers may have assumed that salespeople who are, for example, risk averse would select the sales jobs that pay mainly fixed salary, while those who are risk seekers would select the jobs that pay mostly incentive. However, this is not necessarily the case. Salespeople are not \"weeding\" themselves out during the recruitment process.As a result, greater importance must be placed on human resource and sales managers during the time of recruitment. Using realistic job previews is one potential avenue for improving the outcome of the recruitment process (Rynes and Gerhart, 2000). This technique should be considered further within team interviews (Rushmer, 1997). Research has shown that team-based interviews may help prospective candidates gain a better understanding of their career expectations.Finally, future research in this area should consider the influence of career advancement on preferences and perceptions of various reward systems. According to Anderson et al. (1992), as sales personnel ascend the organizational hierarchy, managers have increasing amounts of responsibility and expectations of higher and possibly different rewards. Our research supports the need for future research linking career stages to differences in rewards. Future research needs to study this phenomenon over longer periods of time.In conclusion, future research in sales compensation must continue to consider the impact of individual differences on the overall effectiveness of compensation. While this study provides an initial investigation of these relationships, the study had limitations that should be noted. In particular, this study examined only two forms of compensation (salary versus incentive). Future work should expand on this to consider other forms of control commonly used in sales force management. Also, additional individual-difference variables should be evaluated, and valences for additional types of rewards should be solicited. Also, our sample limited us from analyzing the disengagement stage. Research has indicated that disengagement is an important, though often overlooked, stage. Thus, a more studies that include this variable would beneficial to the field. Although more research is certainly warranted, the exploratory findings of the current study do provide some interesting insights into the inevitable interplay that takes place between organizational policy (in this case pay mix) and the individual.\n",
"role": "user"
},
{
"content": "Purpose: - To examine the influence of company-imposed reward systems on the motivation levels of salespeople.\nDesign/methodology/approach: - Data were collected from 214 business-to-business salespeople. In order to assure the adequacy of the survey instrument, several salespeople were recruited to \"pretest\" the questionnaire. To test the potential moderating influence of career stage on pay mix and valence, expectancy, and instrumentality estimates, a split-group analysis was performed. To test the moderating hypotheses for risk, we used two-step regression models in which the dependent measures (i.e. valence, expectancy, and instrumentality) were first regressed on the predictor variables as main effects, and then regressed on the multiplicative interaction term along with the main effects.\nFindings: - Support was found for many of the hypotheses. In particular, individual-level variables such as career stage and risk preferences moderate the relationship between pay mix and valence for the reward, expectancy perceptions, and instrumentality perceptions.\nPractical implications: - Managers must acknowledge that certain salespeople respond positively to fixed salary plans while others respond positively to incentive. In the past, managers might have relied on the salesperson to select the appropriate job for them. Salespeople are not \"weeding\" themselves out during the recruitment process. As a result, greater importance must be placed on human resource and sales managers during the time of recruitment.\nOriginality/value: - There already exists a robust stream of literature on person-organization fit that suggests that employees prefer to work for companies that are compatible with their personalities. There is an increasing amount of work in multi-level research that suggests bridging the macro (organizational) and micro (individual) perspectives will enable researchers to make linkages between research streams that are commonly viewed as unconnected.\n",
"role": "assistant"
}
] |
Flesh-and-blood knowing: Interpreting qualitative data through embodied practice-based research
|
[
"Qualitative research",
"Interpretation",
"Affect",
"Carnal sociology",
"Embodied practice-based research"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Qualitative research is heavily focused on researcher's interpretations upon the phenomena under investigation. The researcher's interpretation process is always mediated by performative judgements (Bispo, 2017) aiming to make meaning in a reasoning way on the data produced in the field. However, the notion of reasoning in organization studies is highly based on a mentalist perspective wherein the mind and the body are treated as two distinct entities (Gherardi & Perrotta, 2014).\nInterpretations, judgments, affects and carnal sociology: Different from the mainstream organizational qualitative research wherein the researcher perception and affective conditions are usually not considered as a relevant resource for data interpretation, we advocate that reasoning is not a state of affairs separated from the senses, feelings, perceptions, and affects. Reasoning is an embodied condition. To reason on any social phenomena is to make sense about them. There is no rational reasoning that is disembodied and not mediated by perception and affect. The qualitative researcher should have the \"capacity to combine knowledge, judgement, understanding, emotion, and intuition to act appropriately\" (Macklin & Whiteford, 2012, p. 93).\nThe researcher as an embodied, embedded and affective practitioner: The embodied practice-based research style opens room to advance in certain aspects of the organizational life that are hard to identify or understand them relying only on classic research methods such as traditional ethnography, hermeneutic, interviewing, historiography, case study, archival research, or statistical techniques. These methods employed parochially are not able to apprehend organizational aspects related to the tacit texture of social action that involves \"doxic categories, phronetic abilities, and ordinary ways of being, feeling, and acting\" (Wacquant, 2015, p. 4).\nA story from the field: In this paragraph we wish to reflect on the embodied experience as a group of researchers engaged in working on the interpretation of research data, the problems encountered, how they were solved and how, in solving what was puzzling in that moment, they arrived to theorizing what was named \"affective ethnography\".\nFinal remarks: Our goal in this article was to offer a perspective to interpret qualitative data drawing on the notion of \"embodied practice-based research\". Instead of proposing a rigid set of procedures to interpret and report organizational qualitative data, we advocate that the researcher embodied knowing and experience is relevant to apprehend and understand subjective phenomena - especially those overlooked by organizational scholars on the tacit texture of social action and cognition.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to offer a perspective to interpret qualitative data drawing on the introduction of the notion of \"embodied practice-based research\".\nDesign/methodology/approach: Drawing on a comprehensive literature review to support a meta-theoretical approach, we developed a theoretical essay.\nFindings: The body is not only a field of studies but a mean of study as well. The embodied practice-based research is an inquiry style to access the tacit texture of social action and cognition.\nPractical implications: Embodied practice-based research may impact qualitative researchers' education and the way to report methodological proceedings and data report.\nOriginality/value: The core contribution of the paper is the introduction of a new research style able to change how researchers' bodies may be used in qualitative management research.\n",
"role": "assistant"
}
] |
Market analytics of the rice wine market in Japan: an exploratory study
|
[
"Japan",
"Marketing analytics",
"Wines",
"Purchasing behaviour",
"Asian wine consumption"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Wine is a global and highly fragmented industry with multiple producers and product-styles. Wine making and drinking has a rich tradition in many cultures. For Europe and the Western world, wine is made of grapes. While Asian countries have been embracing wine made of grapes in recent years, Asian countries also consume wine made out of fermented rice: rice wine is known as sake in Japan. Sake is considered a spirit from some perspectives (Camillo, 2012) but sake is part of the wine category in China, Japan and Korea (Anderson et al., 2017). While the production of sake has been declining over many years, from 513 (2006) to 444 (2015) million L (Statista, 2015), the importance of sake is growing in different markets because of the globalisation of Japanese food, increasing number of tourists arriving at Japan, news coverage in western countries, implementation of premium sake category and internationalisation of sake brewers. For example, Japanese sake brewers have established and expanded capacity in the USA in recent years followed the increasing recognition of sake and Japanese food (Otsuka, 2015). Sake consumption is becoming global and competing with wine in food pairing and cooking, especially among millennials. However, there is no enough information about the market for sake. This paper presents an exploratory study of the market for sake with three objectives. Firstly, a qualitative description of the market in Japan. Secondly, a quantitative description of the dynamics of purchasing sake in an off-trade channel using a large data set of point of sale (POS) transactions. Finally, a quantitative analysis of purchasing behaviour using point of sales data and a behavioural theory with simulation.\nLiterature review: The literature review comprises the following three areas: recent studies on Asian wine consumption, consumer behaviour models and empirical studies on wine purchasing behaviour.\nResearch method: The research consists of mixing three methods, namely, qualitative and quantitative analysis to described the dynamics of the market for sake and a simulation to explore the proposition related with anchoring-and-adjustment theory that a loyalty loop exists for some brands.\nResults: This section contains two parts. Firstly, there is a general description of the market based on established concepts in the wine consumption literature. The following section presents the evaluation of the anchoring-and-adjustment theory for selected products.\nDiscussion: I discuss the results in the following two areas: the characteristics of the market for sake; and the use of anchoring-and-adjustment theory for purchasing behaviour.\nConclusions, implications, limitations and future research: This study empirically investigates Japanese consumer behaviour in the rice wine market, sake, through the application of theories used on previous studies on wine purchasing behaviour. This research adds to the literature on purchasing and consumer behaviour new perspectives based on the behaviour observed in an Asian country for a different type of wine. For example, the use of large packaging for embedding wine in everyday cooking. Specifically, the study suggests the existence of a \"loyalty\" loop as a process of anchoring-and-adjustment behaviour. The evidence indicates the existence of an anchor in purchasing activities that adjusts with seasons and annual events. This investigation provides the basis for studying consumer behaviour observing anchoring-and-adjustment processes through long-term behavioural data.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyse consumer buying behaviour in the Japanese rice wine, also known as sake market.\nDesign/methodology/approach: The study applies a novel qualitative and quantitative analytical methodology to an off-license channel in Japan. The methodology involves the use of anchoring-and-adjustment theory and simulation to a large set of point of sale data. The selection of the brands used for the study are more than 230 brands and more than 150 sake breweries.\nFindings: Age and gender are important factors determining recurrent patterns of purchasing behaviour. Small size packaging, e.g. one cup, has the highest volume in sales, for example, convenience shopping, but it depends on exogenous factors, e.g. summer season or festive events.\nResearch limitations/implications: Limitations are related with the lack of specific personal data from consumers that impedes to test behavioural attitudes driving loyalty to brands. Anchoring-and-adjustment theory can be a valid approach to evaluate large longitudinal data sets of purchasing behaviour.\nPractical implications: Results indicate that fragmented markets tend to over-expand the assortment affecting volume stability. However, this dynamics is difficult to avoid when all participants are engaged in this behaviour and the market is strongly segmented by age and gender.\nOriginality/value: The paper contributes to the body of knowledge of buyer behaviour in relation to purchasing and consumption for other types of wine. It is the first application in alcoholic beverages of anchor-and-adjustment theory.\n",
"role": "assistant"
}
] |
Emotional intelligence, emotional labor strategies and satisfaction of secondary teachers in Pakistan
|
[
"Emotional intelligence",
"Emotional exhaustion",
"Emotional labour strategies",
"Private schools"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: School is a challenging workplace and teaching is a demanding job (Yao et al., 2015). Teachers are constantly challenged by the working environment like disruptive students, heavy workloads, insensitive administrators, professional standards and high expectations of parents (Chan, 2006; Leithwood and Beatty, 2008). Like some other human service jobs (police, nursing and customer service), teaching profession requires teachers to suppress certain negative emotions - anger and frustration - and express other more positive emotions while interacting with students and colleagues (Isenbarger and Zembylas, 2006). Hence, emotional labor (EL) is an integral part of teachers' work in school, classroom and parent-teacher meeting (Brennan, 2006). Educational leaders need to better understand teachers' emotions to be directly and intentionally helpful to their teaching staff for the purpose of improving student learning (Leithwood and Beatty, 2008).\n2. Literature review: Emotions and EI are an imperative part of teaching and teachers' lives (Hargreaves, 2001; Isenbarger and Zembylas, 2006; Yao et al., 2015). Individuals with high EI are capable of understanding and expressing their own emotions, recognizing emotions in others, regulating affect and using emotions to engage in adaptive behaviors (Salovey and Mayer, 1990; Wong and Law, 2002). Teachers with high scores on EI experience high motivation, performance, efficiency, commitment and effectiveness (Anari, 2012; Kim et al., 2012; Sy et al., 2006). EL is perceived as the process by which teachers make efforts to inhibit, generate and manage their feelings and emotions according to the normative beliefs and expectations held about the teaching profession (Yin and Lee, 2012).\n3. Research methodology: The participants were secondary school (Class 6-10) teachers employed by private schools in two northern cities of Pakistan. Principals of these schools were briefed about the purpose of the study and requested for the approval of data collection. Upon approval, researchers contacted the teachers and shared the objectives of the study, guided them on how to fill in the questionnaire and assured full confidentiality of their responses. A total of 400 self-administered questionnaires (200 in each city) were distributed conveniently to the teachers in 28 schools. Out of which, 322 completely filled questionnaires were returned with a response rate of 80.5 percent. Participants' mean age was 29 years with average teaching experience of five years. Majority of the respondents were female (78 percent), unmarried (57 percent) and contract-based/visiting employees (81 percent).\n4. Results and discussion: Table I shows descriptive statistics, Cronbach's a value and correlation coefficents among the variables. Cronbach's a values for all the scales are above 0.70 which sufficiently fulfill the conditions for internal consistency reliability (Nunally, 1978).\n5. Conclusion: The current study examines the relationship of EI with EL strategies, emotional exhaustion and TS of secondary school teachers working in private schools. Surprisingly, the study reveals some interesting results that differ from existing literature and sheds light on the role of EI, EL, emotional exhaustion and TS in teachers' work.\n6. Recommendations: This study recommends training and mentoring of school teachers on how to select proper strategies of EL to deal with emotional situations (Truta, 2012). Higher education institutes may introduce such modules or courses in their teachers training programs. The study recommends teachers to make a greater use of surface acting and ENFE that produce the most TS and the least emotional exhaustion. Moreover, Ye and Chen (2015) suggest the use of training programs to develop and enhance EI among teachers. This helps school teachers to improve their TS and to make a wise use of EL strategies (Chan, 2006). TS could also be increased via the improvement of EI or the adjustment of EL strategies (Yao et al., 2015).\n7. Limitations and directions for future research: Being quantitative in nature, current study employs a cross-sectional research design. First limitation of the study is common method variance because self-reported data were collected from a single source. Social desirability and impression management by teachers may have produced somewhat biased results. Conveniently selected small sample size (322) constitutes second limitation which calls for careful generalization of the results. Third limitation is the gender distribution of sample (78 percent female). To a considerable extent, this reflects the dominance of female teachers in schools while to some extent it reveals a biased selection of the respondents.\n",
"role": "user"
},
{
"content": "Purpose: Emotional intelligence (EI) is crucial to determine the emotional labor (EL) strategies and satisfaction of teachers. The purpose of this paper is to investigate the relationships of EI with EL strategies, emotional exhaustion and teaching satisfaction (TS) of secondary teachers in private schools.\nDesign/methodology/approach: A self-administered questionnaire was used to collect data conveniently from 322 secondary teachers working in 22 private schools.\nFindings: Results indicate that teachers' EI positively affect surface acting, expression of naturally felt emotions (ENFE) and TS. Surface acting and ENFE are positively related with TS. Deep acting and ENFE exhibit positive relations with emotional exhaustion of teachers. The study also confirms direct as well as indirect effects of EI on TS through surface acting and ENFE.\nPractical implications: This study recommends teachers to make a higher use of surface acting and ENFE because they produce the most TS and the least emotional exhaustion.\nOriginality/value: This is the first study that examines the mediating role of EL strategies for EI and TS.\n",
"role": "assistant"
}
] |
The interaction of intellectual capital assets and knowledge management practices in organizational value creation
|
[
"Knowledge management",
"Value creation",
"Intellectual capital",
"Organizational performance",
"Survey",
"Knowledge management practices"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Organizational performance is increasingly a knowledge-related issue. The two key academic discussions addressing knowledge in organizations are the literatures of intellectual capital (IC) and knowledge management (KM). While the first focuses on intangible resources that contribute to value creation (e.g. Edvinsson and Malone, 1997; Sullivan, 1998; Spender et al., 2013), typically in terms of human, structural and relational capital assets governed by an organization (e.g. Bontis, 2001; Guthrie, 2001), the latter concentrates on the knowledge-related processes and management activities in firms (e.g. Gold et al., 2001; Lee and Choi, 2003; Heisig, 2010).\n2. IC assets and KM practices: The basic tenet of the knowledge-based view of the firm is that performance differences among firms accrue due to their differing assets and management mechanisms of knowledge. For example, Kogut and Zander (1992), Grant (1996), and Eisenhardt and Santos (2002) distinguish between knowledge and the capabilities to manage it as the key bases of organizational value creation. Ultimately, this distinction goes back to the separation between structure and process or parts and motions, underlying much of the current management and organizational research.\n3. Interaction of IC assets and KM practices: To conceptually analyse organizational value creation with both static and dynamic perspectives, several possibilities concerning the nature of interaction between IC assets and KM practices are overviewed. For the sake of simplicity, throughout the models \"organizational performance\" is discussed as an outcome variable, while \"value creation\" can be seen more as a meta-level concept discussing the whole process. As for organizational performance, a generic perspective is taken here, without distinguishing between different aspects of it. In fact, both financial and innovation-related outcomes of organizational performance have been found to result from the possession IC assets (Bontis and Fitz-Enz, 2002; Youndt and Snell, 2004; Subramanian and Youndt, 2005; Hermans and Kauranen, 2005; Wang and Chang, 2005; Wu et al., 2007; Cabrita and Bontis, 2008; Mention and Bontis, 2013) or utilization of and KM practices (e.g. Gold et al., 2001; Lee and Choi, 2003; Chuang, 2004; Gloet and Terziovski, 2004; Darroch, 2005; Zack et al., 2009; Andreeva and Kianto, 2012). Kogut and Zander (1992) propose that value creation through innovation takes place when various types of existing knowledge is combined to generate new applications, and thereby it is the capabilities for combining knowledge that produce and replenish the IC assets of a firm. This can be - and has been - interpreted in various ways in terms of the nature of interaction between IC assets and KM practices. Next all the possible interactions are discussed, and also their feasibility from both theoretical and empirical perspectives is assessed. In doing this, a division is made between two types of interaction processes: moderating effect and mediating effects. The moderating effect is defined as a third variable's effect that changes the relationship between two related variables, whereas mediation means the effect of third variable intervening between two other related variables (see e.g. Hair et al., 2006).\n4. Discussion and implications: This paper addressed an important issue which so far has been relatively overlooked in the IC and KM literature: how different types of intangible assets can be better managed for creating value and organizational performance. The study is an attempt to bring together the two main literatures within the knowledge-based view of organizations: the IC and KM traditions and to build an conceptual-theoretical bridge between those traditions.\n",
"role": "user"
},
{
"content": "Purpose: Organizational performance is increasingly grounded on knowledge-related issues. The two key academic discussions addressing knowledge in organizations are the intellectual capital (IC) and knowledge management (KM) literatures. However, there are very few earlier studies systematically combining these approaches and demonstrating how IC assets and their management mechanisms might interact in organizational value creation. Therefore, the purpose of this paper is to develop and argue a theoretical model depicting the connections between IC, KM practices and organizational performance outcomes.\nDesign/methodology/approach: The paper draws on IC and KM literatures to build a theoretical model on how intellectual asset assets and their management practices interact in producing organizational performance. Several conceptual models and related discussion on the interaction of IC and KM practices are put forth.\nFindings: Organizational value creation is based on both static (IC assets) and dynamic (KM practices) aspects of organizational knowledge, in various combinations. In this paper, potential interaction effects between IC assets and KM practices in terms of moderation and mediation were conceptually analysed, and four alternative models were proposed on how the knowledge-based issues affect organizational performance.\nResearch limitations/implications: The paper is purely theoretical without empirical evidence.\nPractical implications: The paper suggests that organizational value creation is a function of both possessing valuable intangible assets as well as being able to manage these assets systematically. The four models concerning the interaction of IC assets and KM practices in value creation presented in the paper provide managers with tools to reflect about their own thinking model concerning how knowledge produces value in their own firms.\nOriginality/value: By addressing both the \"static\" asset aspect of IC as well as the \"dynamic\" perspective of how leveraging IC assets can be enabled by systematic managerial activities, the paper combines the key issues in IC and KM literatures and demonstrates how intangible resources should be managed to produce value. The authors are not aware of any previous studies explicitly combining and distinguishing IC and KM fields to this extent. The paper therefore contributes to the literature on knowledge-based issues in organizations at large and potentially offers a theoretical grounding for many empirical and theoretical future studies.\n",
"role": "assistant"
}
] |
Post-merger processes cause problems for Nordea's HR team: Financial-services firm struggles to give a strategic role to human resources
|
[
"Acquisitions and mergers",
"Human resource management",
"Financial services",
"Organizational change"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: A cross-border merger of financial organizations from Denmark, Sweden, Finland and Norway led to difficulties for the newly-formed company's HR function.\nComplex cross-border integration processes: When the new organization was announced, the official organizational chart did not contain an HR function or a head of HR. Eventually it was decided that a Swede would be in charge of the new group HR function and report directly to the chief executive of the new bank, but in fact little interaction took place between the two.After long discussions between national HR executives it was decided that a country-based HR organization would be established with a three-person HR management group. This meant, initially, there was little integration of the HR functions across the country borders. This arrangement continued until 2002, when a new cross-border HR organization was implemented.\nThe HR administrative role at Nordea: The administrative role of HR covers how it designs and delivers processes for staffing, training, appraisal, compensation and reward, promotion and so on. One of Nordea's challenges in the transition, from four domestic financial-services institutions to a single multinational one, was how to deal with the task of developing an HRM infrastructure that would ensure efficient practices while simultaneously paying enough attention to the national and business-unit differences within the complex organization. The merging country-based organizations had different HR systems and concepts and the contrasts between the three business areas were also considerable.Inevitably, in most areas there was less-than-perfect agreement about what constituted \"best\" HRM practices and, consequently, only a few changes were made at the group level in the administrative HRM processes during 2000 and 2001. A decision was finally made in 2002 to invest in a group-wide IT-based platform for HRM data that was intended mainly to provide cross-border HR data and also contribute to HRM alignment across borders and business areas.In 2001, Nordea's chief executive, Thorleif Krarup, hired an international consulting firm to examine the functioning of the HR organization. One result was a reorganization of the HR function into three main units:1. Group HR administration, responsible for administrative HRM processes.2. Group HR partner, which saw a responsible head of HR appointed within each business area.3. Group HR center, responsible for developing corporate tools for staffing, training, leadership development and remuneration.\nHR as an employee champion: HR also has a role as employee champion to enhance staff contributions - such as commitment - by taking notice of employees' needs. However, in most parts of Nordea, the HR units had a tradition of operating as a centralized and quite bureaucratic organization that most employees would not approach naturally with grievances and concerns.The potential strategic-partner role of HR focuses on the alignment of HR strategies with business strategy, but top management at Nordea seemed to lack interest in the HR function. The head of HR had little access to top management, whose attention was more focused on branding the firm. Although the HR function gradually received more attention from top management, HR specialists were still not involved in corporate strategy discussions.Another strategic role is that of HR as a change agent, which could be expected to be particularly important in the integration of cultures in the merged corporation during the post-merger period. However, according to HR managers and top executives at Nordea, the HR organization had limited influence on the transformation and integration process following the mergers and acquisitions of 2000. The corporate HR function played only a secondary role in the phase of cultural integration, which was driven forward primarily by Nordea's identity and communications department.The organizational roles played by the HR function influence how workforces are managed in cross-border mergers, but the low expectations of top management concerning strategic contributions from Nordea's HR seemed to contribute to the limited attention to people management and cultural change in the integration process.From the initial merger negotiations onwards, there is little evidence of top management having high expectations of HR possessing any strategic function - so HR was not invited to play any important part in post-merger processes. It was also difficult to identify one way in which to organize the HR function to satisfy the demands of the cross-border and different business areas.The merger itself appears to have diverted the attention of HR managers from the potential contribution to the strategic development of the rest of the organization and top management seemed to be more focused on financial outcomes.The case of Nordea illustrates how communication of role expectations both to and from the HR function may be especially important in clarifying the roles the function is to play in the post-merger process.\nComment: This review is based on \"The HR function in large-scale mergers and acquisitions: the case study of Nordea\", by Bjorkman and Soderberg (2006), in Personnel Review. It discusses clearly the potential pitfalls for the HR function following a cross-border merger, and considers the scope for human-resource contribution after mergers and acquisitions.\n",
"role": "user"
},
{
"content": "Purpose: - This article explores the roles played by HR in post-merger change processes. Helps to uncover the inherent problems in managing HR issues in the post-merger organization and highlights difficulties in giving HR a strategic status.\nDesign/methodology/approach: - This briefing is prepared by an independent writer who adds his own impartial comments and places the article in context.\nFindings: - Highlights typical problems in organizing and managing HR issues and reveals how the HR function is easily left with a secondary, non-strategic role in these processes.\nPractical implications: - More exploration is needed of both the enablers and constraints of the organizational roles played by the HR function and how these relate to the way people are managed in large-scale cross-border mergers.\nOriginality/value: - The briefing saves busy executives time by selecting only the most pertinent information and presenting it in a condensed, easy-to-digest format.\n",
"role": "assistant"
}
] |
Teachers' professional development through teachers' professional activities
|
[
"Faculty development",
"Creative assignments",
"Faculty commitment",
"Faculty Self-Assessment",
"Inter-disciplinary collaborations"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Faculty development (FD) involves a range of activities for teachers' improvement in their professional skills to carry out teaching, research and administrative activities in education (Kwan et al., 2009). Professional educators continuously reflect on their practices; observe and practice new ways of teaching, learning and networking to develop and articulate their beliefs (Edmonds, 2007). Teachers' professional activities (TPAs) include experimenting, collecting new and up-to-date knowledge, attending courses and conferences with new insights, self-reflection (Thoonen et al., 2011), developing commitment to lifelong professional development and enhancing students learning (Ciampa, 2015); collaboration with the members of university and professional community (Edmonds, 2007); and becoming creative in the class (Butler and Reddy, 2010). These activities enhance the professional learning of teachers and prepare them to meet the challenges of students' development suitable to the society (Ciampa, 2015). This study aims at identifying and assessing the relative importance of various TPAs for FD with special reference to technical education. Every year, around 1.5 million students graduate engineering in India as against 95000 in US (Bansal, 2017). About 80 per cent of such engineers are unemployable (Aspiring Minds, 2016) due to lack of adequate skills for employment (Chakrabarthy, 2016). Among other reasons, chronic shortage of quality faculty is one of the reasons behind it (ET Online, 2018).\n2. Literature review: Education is a process that helps in continuous teachers' progress by improving their abilities and expansion of interests (Bonstingl, 1992).Teachers are required to fulfill their responsibilities by keeping themselves abreast of best practices (Rhodes, 1992).\n3. Methodology: 3.1 Objectives\n4. Discussion: The very first \"most important\" TPA required by faculty is their regular self-assessment and self-reflection. Self-assessment and self-reflection are important for teachers to find what is inside them. This should understand what they are capable of doing and what they are not. Self-assessment is important for life-long learning (Ramani and Leinster, 2008). It is one of the pre-requisites for an effective and inspiring teaching (Sammons et al., 2016).\n5. Conclusions: To conclude, the present study highlights that teachers should regularly undergo self-assessment and self-reflection. This requires a deep commitment toward their profession and a strong sense of making the difference. They must understand that there is a difference between just doing and making the difference. The participation in globally acclaimed courses and conferences is very important because these provide teachers with opportunities and stimuli for self-assessment and self-reflection. Another important professional activity for teachers is developing a creative problem-solving approach among students by giving challenging assignments. This professional activity is highly important in technical education because design thinking in engineering students and strategy crafting in management students will provide a sustainable competitive advantage for the firms they serve.\n6. Implications: The study highlights different teachers' professional practices that they must establish, raise, promote, encourage and organize for their development. The study further classifies different practices according to their relative importance. The institute can evaluate their resources, budgets and efforts according to the relative importance of such practices. The classification of TPAs would help faculty to increase their efficacy.\n7. Limitations and future scope: Although the present research makes an important contribution to the existing literature on TPAs for teachers' professional development, it has certain limitations. First, the influence of demographic characteristics such as age, gender, income and education has not been studied in the present study. Second, the sample was specific to one state. There may be the chances of response bias in such situations. Therefore, there are few reservations in generalizing the findings. These limitations leave scope for future research related to TPAs by considering the influence of educators' demographics on such activities. Further, the study can be performed in other streams and geographic areas to look for further generalization of the findings.\n",
"role": "user"
},
{
"content": "Purpose: This study aims to identify various teachers' professional activities (TPAs) and classify these TPAs according to their relative importance for the professional development of teachers.\nDesign/methodology/approach: The systematic review of the literature had been conducted to identify various TPAs in the institutions. Later, an empirical research had been conducted through confirmatory factor analysis using AMOS 20.0 to classify these TPAs according to their relative importance using the natural gap in standardized beta (b) values. In total, 96 administrators of 35 technical institutions of Punjab (India) offering engineering and management programs and 93 veteran educational experts had responded in a field survey.\nFindings: The results of the study identified eight TPAs and further revealed that \"regular self-assessment,\" \"adopting a creative problem-solving approach\" and \"developing deep commitment to make the difference\" qualify among the \"most important\" activities for the professional enhancement of the faculty.\nOriginality/value: The study highlights different TPAs that they must establish, raise, promote, encourage and organize for their development. The study further classifies different activities according to their relative importance. The institute can evaluate their resources, budgets and efforts according to the relative importance of such activities. The classification of TPAs would help faculty to increase their efficacy.\n",
"role": "assistant"
}
] |
The willingness to disclose personal information: Trade-off between privacy concerns and benefits
|
[
"Privacy concerns",
"Perceived disclosure benefits",
"Privacy assurance mechanisms",
"Personal information disclosure"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: With the technological advancement of information and communication technology, the growth rates of electronic commerce and services have been increasing exponentially. A major reason for this growth is the reliance of both customers and firms on electronic channels. In general, customers use the internet for shopping, and firms use it for selling goods and services online. When customers use internet channels, online firms collect extensive personal information, including addresses, phone numbers, age, marital status, credit card information, etc. The online firms or companies use the internet not only for buying and selling but also for providing online services. The terms online firm, online company and e-commerce company are used interchangeably in this study. Online firms claim that they use customer information to personalize their offerings and services to ensure that customers enjoy better services in future purchases. However, this is not always the case; sometimes customer information is subjected to opportunistic behavior. Online firms, without customer consent, could use personal information for advertisement or sell the information to a third party, thereby, making customers very concerned about their privacy. Such a concern reduces customer willingness to share personal information with online firms (Lowry et al., 2012). However, although customers express concerns about their privacy, they still share their personal information to gain certain benefits from online firms.\nTheoretical foundation: Numerous studies have discussed the privacy concern and benefits issue, and their relationship with information disclosure using privacy calculus theory (Culnan and Armstrong, 1999; Dinev and Hart, 2006; Laufer and Wolfe, 1977; Xu et al., 2009). This theory suggests that people's intention to disclose personal information is based on a calculus of behavior where a trade-off takes place between the expected risks and benefits of disclosing personal information. In the context of e-commerce, the privacy calculus means that customers perform a risk-benefit analysis and evaluate the consequences of disclosing personal information (Dinev and Hart, 2006). Using this concept, Culnan and Armstrong (1999) studied the relationship between information disclosure and the existence of procedural fairness. In other words, customers disclose personal information if there is a fair procedure to protect their information. Xu et al. (2009) studied the relationship between information disclosure in location-based services (LBS) and privacy concern using the privacy calculus theory. They found that financial compensation encouraged consumers to disclose their personal information. Xu et al. (2013) studied the factors affecting information disclosure in the context of social networking sites using the privacy calculus and the theory of planned behavior (TPB). Malhotra et al. (2004) used theory of reasoned action (TRA) to propose a causal model based on the trust-risk framework; according to this causal model, the presence of trust and risk beliefs can mediate the impact of privacy concern on behavioral intention to disclose personal information. Another study by Gurung and Raja (2016) used the TRA and TPB to study the privacy and security concerns in the e-commerce context and how it affects individual behavioral intention. Many other studies in the area of online privacy concern have used the privacy calculus, the TRA, and/or the TPB (Li, 2012; Smith et al., 2011; Li, 2011).\nLiterature review and hypotheses: The link among the foundation theories (i.e. privacy calculus theory, protection motivation theory and information boundary theory) is that the WTD is affected by three factors, namely, privacy concern, perceived disclosure benefits and privacy assurances. This section reviews the relationships between privacy concern, perceived disclosure benefits and privacy assurances and their impact on the WTD. The underlying core constructs of the research model are depicted in Figure 1.\nResearch methodology: Survey development\nResults: Hierarchical regression analysis was used to test the hypotheses, as it tests the main effects and the moderating effects separately. In Block 1, we entered the main study variables: privacy concern variables (COL, ERR and IAU), privacy assurance variables (PCN and GTC) and perceived benefits variables (FNR, PRS and CNV). In Block 2, we entered the interaction terms between the privacy concern variables and privacy assurance variables. The interaction terms were PCN x COL, PCN x ERR, PCN x IAU, GTC x COL, GTC x ERR and GTC x IAU. It is important to note that for the interaction terms, we multiplied the mean-centered values of the privacy concern variables by the mean-centered values of the privacy assurance variables. The use of mean-centered values, as opposed to raw values, avoids the possibility of having multicollinearity issues among the predictor variables (Aiken et al., 1991). We also examined the presence of multicollinearity among predictor variables using variance inflation factor (VIF) and tolerance. The VIF scores ranged from 1.183 to 3.765, below the threshold value of 10; and the tolerance scores ranged from 0.321 to 0.845, above the threshold value of 0.10. Hence, this diagnostic showed that there was no evidence of multicollinearity. Table IV presents the results of the hierarchical regression analysis.\nDiscussion and implications: Unlike previous studies that considered privacy concern as an aggregate variable, this study examined the relationships between the dimensions of privacy concern and the WTD. We argued that there were three major sets of factors that positively or negatively affected customer WTD to online firms. Each major factor consisted of sub-factors or dimensions. The first major factor was privacy concern. It consisted of three dimensions: COL, ERR and IAU. The second major factor was privacy assurance and it consisted of two dimensions: PCN and GTC. The third major factor was perceived benefits; it consisted of FNR, PRS and CNV. Of all these factors, only three dimensions (i.e. ERR, FNR and CNV) had significant relationships with WTD. Table V summarizes the results of the research hypotheses.\nLimitations and future research: Although this study contributes to the literature on information privacy concerns and benefits of disclosing personal information, it has some limitations that may open avenues for further research. First, male and female respondents were not equally represented in our study sample. However, we argue that the sample is suitable for this study, as Saudi Arabia represents a male-dominated society. Second, the dependent variable was the willingness rather than the actual disclosure of personal information. People who expressed less willingness to share are more likely to actually disclose personal information. This inconsistency of willingness to disclose private information and the actual more restrictive behavior is often referred to as the \"privacy paradox.\" Third, the sensitivity of the personal information is an important factor that was not considered in the sharing of information. For example, customers may be willing to share their postal address but hesitant in sharing their credit card information because they feel the latter is more sensitive than the former.\nConclusion: When disclosing personal information, customers encounter privacy concerns and receive beneficial outcomes; this captures the concept of the privacy calculus. The most significant dimension of privacy concern is COL. Other dimensions (i.e. ERR and IAU) had a non-significant relationship with disclosing personal information. This research differs from previous studies as the privacy concern dimensions were examined individually rather than collectively. On the disclosure benefits side, customers were willing to disclose personal information in exchange for PRS or for shopping CNV. The assurance mechanisms did not attenuate the privacy concerns. The results suggest that online firms should be cautious when collecting customer data because they have a major concern with the privacy of their information. On the other hand, online firms should capitalize on the factors that motivate customers to disclose personal information.\n",
"role": "user"
},
{
"content": "Purpose: Customer privacy and security are major concerns. Online firms worldwide collect customer data for various reasons. This study aims to investigate factors that motivate and hinder a customer's willingness to disclose personal information (WTD) to online firms on e-commerce websites.\nDesign/methodology/approach: Based on an extensive literature review, three sets of factors have been identified. These sets of factors are privacy concern, perceived disclosure benefits and privacy assurances. It is hypothesized that privacy concerns negatively affect the disclosure of personal information, while the perceived benefits of disclosure have positive effects. Privacy assurances would positively affect information disclosure and attenuate the negative effect of privacy concerns on the disclosure of personal information. The authors gathered data from 253 online customers in Saudi Arabia.\nFindings: The results indicate that perceived disclosure benefits and privacy concerns have a significant positive and negative relationship, respectively, with WTD online. Privacy assurances had neither a direct nor a moderating effect on information disclosure.\nResearch limitations/implications: The findings will inform online firms about the factors that prevent or motivate customers to disclose personal information.\nOriginality/value: The effect of privacy concerns and benefits on personal information disclosure are not fully understood in Saudi Arabia. This study reveals more insights into the specific factors that make online customers reluctant or motivated to disclose their personal information.\n",
"role": "assistant"
}
] |
The role of emotions in online consumer behavior: a comparison of search, experience, and credence services
|
[
"Online consumer behavior",
"Emotions",
"Search attributes",
"Experience attributes",
"Credence attributes",
"Site atmospheric cues",
"Web sites",
"Shopping"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\nIntroduction: In the last decade, service providers have realized the importance of developing an effective web site. Grewal and Levy (2009) stated that the use of the internet by almost all consumers and business segments has tremendously changed the retailing marketplace. Furrer and Sollberger (2007, p. 106) distinguished between the dynamics and evolution of the services marketing literature and found that \"relationship marketing\" and \"technology\" are the two themes \"with recent popularity and a healthy growth rate\". Rajamma et al. (2007) reported that consumers associate services with the online shopping mode and physical goods with the traditional bricks-and-mortar outlets. These emphasize the importance of investigating online consumer behavior in services marketing. Consistently, Hackman et al. (2006) stated that there is still much to explore about consumer behavior in the emerging field of online services.In traditional services marketing, the interpersonal relationship between a consumer and service employee was one of the most important elements in forming consumers' attitudes and impacting their behaviors. In online services marketing, however, the interpersonal interactions have been replaced by technology and limited to electronic interaction (Holloway and Beatty, 2008), and consumers' attitudes and behaviors are influenced by web site characteristics such as ease of navigation (Holloway and Beatty, 2003, 2008). Instead of interacting directly with service employees, consumers surf and navigate the company's web site to evaluate the service.Over the last decade researchers and practitioners attempted to explain how service providers \"can best design and deliver a quality encounter\" in online contexts (Cho and Menor, 2010). Store environments and atmospheric cues are known as important marketing input to impact consumers' purchasing decision (Baker, 1987; Mattila and Wirtz, 2008). In a review of the literature, Turley and Milliman (2000) reported over 60 published articles that found significant associations between store atmospheric cues and consumer behavior variables (Dailey, 2004). In online contexts, studies reported the influence of site atmospheric cues such as site informativeness, effectiveness, and entertainment (e.g. Davis et al., 2008; Hausman and Siekpe, 2009; Richard, 2005) on consumers' attitudes and behaviors. In this study, consistent with Mazaheri et al. (2011), we focus on consumers' perceptions of site atmospheric cues and investigate the antecedents and consequences of these perceptions across the three service categories (i.e. search, experience, and credence).Based on studies in psychology (Clore et al., 1994; Zajonc, 1980) we propose that emotions which arise from the initial exposure to the web site impact consumers' perceptions of site atmospheric cues (Mazaheri et al., 2011). Then, following the Stimulus-Organism-Response (SOR) framework by Mehrabian and Russell (1974) and previous literature in marketing (Eroglu et al., 2001; Hausman and Siekpe, 2009; Mazaheri et al., 2011; Richard, 2005) we hypothesize the impacts of consumers' perceptions of site atmospheric cues on consumer response variables, namely site attitudes, service attitudes, site involvement, and purchase intentions.Most studies in services marketing were conducted in specific areas such as the hotel industry (Voss, 1993), the banking sector (Surprenant and Solomon, 1987), the telephone industry (Bolton and Drew, 1991), and the financial services sector (Ding et al., 2010) which limit the generalizability of the results. Peterson et al. (1997) stated that the characteristics of goods and services play a key role in determining the suitability of the internet for marketing to consumers. The goals in developing a web site for search and experience goods are not the same. A search good web site must focus on providing information regarding the features of the good and the internet could serve significant transaction and communication functions. An experience good web site must offer virtual experience to consumers and provide more reliable information for them. Peterson et al. (1997), then, argued that it is necessary to consider the characteristics of goods and services when evaluating the impact of the internet. Accordingly, the second goal of this study is to validate the proposed model by empirically testing the path invariances across the three types of services: search-, experience-, and credence-based services. The similarities and differences among the three service categories are then discussed.This study contributes to the literature in two ways. First, based on the previous literature, we develop an online consumer behavior model. Second, and more importantly, this study provides an opportunity to observe similarities and differences among the three service categories (i.e. search, experience, and credence) by empirically comparing all the relationships in the proposed model across these three types.The remainder of this article is organized as follows. First, drawing on the literature, the proposed model is developed. We discuss the influences of the three types of emotions (i.e. pleasure, arousal, and dominance) on consumers' perception of web site atmospheric cues (site informativeness, site entertainment, and effectiveness of site information content). Then, the impacts of these atmospheric cues on site attitudes, site involvement, service attitudes, and purchase intentions are discussed. Furthermore, we develop hypotheses that propose path invariances across the three service categories. The subsequent section describes the data and empirical analyses used to test the model and hypotheses. Finally, the results, conclusions and discussions, as well as limitations and future research directions are discussed.\nTheoretical background: In this section, we discuss the major theoretical concepts used in this study and explain in detail the general model presented in Figure 1.Site atmospheric cues\nMethodology: Site selection\nResults: Test of reliability and confirmatory factor analysis (CFA)\nConclusion: In traditional retailing, Baker (1987) and Baker et al. (1994) suggested that atmospheric cues are more influential than other marketing inputs that are not present at the point of purchase. In a review of literature, Turley and Milliman (2000) reported over 60 published articles that reported significant associations between store atmospheric cues and consumer behavior variables (Dailey, 2004). This influence is expected to be more important in online context as the environment is a small screen and the consumers have control over \"which Web pages to browse, for how long, and how much information to obtain\" (Dholakia and Rego, 1998, p. 725). It takes seconds for consumers to switch from one site to another. Therefore, designing the web site in a way that attracts consumers and involves them in the site is a huge competitive advantage for companies. In services marketing, consumers surf the provider's web site to obtain as much information as possible. As a result, having a site that is perceived as informative and entertaining is crucial for service providers.Based on Zajonc's (1980) theory, which states that affective reactions precede cognitive reactions, and the SOR framework, this study developed a model of online consumer behavior. The proposed model was validated in three service categories and all the relationships were compared across the three groups. Using the pooled data, 18 out of 19 hypotheses were found to be significant. The only non-significant path was the link between arousal and site effectiveness. Moreover, the majority of our hypotheses were also found to be significant in each service group (17 out of 19).More importantly, the findings demonstrate that the strengths of the majority of the structural paths (13 out of 19) were non-invariant between at least two groups. Out of nineteen paths in the model, ten were non-invariant between search- and experience-based services, eight between search- and credence-based services, and seven between experience- and credence-based services. Once again, the results illustrate the importance of service categorization in analyzing the data and interpreting the results.Overall, the findings indicate that certain components of emotions impact consumers' cognitions differently in each service category. Dominance is the most influential component in search-based web sites; whereas, pleasure is more influential in experience-based web sites. Arousal and pleasure are both more influential in credence-based web sites.Also, particular variables are more important to online consumers when they are visiting different service categories. Site attitudes and site involvement are important factors in search-based services as they are highly associated with purchase intentions. Entertainment is a key variable in experience-based sites as it impacted site attitudes and site involvement. The site effectiveness was negatively associated with visitors' site involvement in experience-based services. Experience-based service providers should perhaps pay more attention to the entertainment and informativeness of their web site rather than the effectiveness of its information content.On the other hand, site effectiveness is an important variable for credence-based services. It highly impacts site attitudes without reducing the consumer's site involvement. Perhaps, the complex nature of credence-based services explains the finding. Compared to search- and experience-based services, consumers should obtain much more information to evaluate a credence-based service. The site effectiveness is necessary to provide consumers with the needed information. Therefore, it is important for credence-based services to have an effective site since the site effectiveness strongly impacted site attitudes.\nLimitations and future research: In order to interpret the findings, one should consider the limitations of this study. We used real service web sites. The only disadvantage is that we were not able to control for some variables such as the type of information provided in the sites. The type of information might play a moderating role in the association of site informativeness and site attitudes. Future research should investigate the role of information type on consumers' attitudes toward the site.The majority of our respondents were students between 18 and 24 years of old. Given their familiarity with Internet technology, this group is the one who uses the internet the most. However, one should be cautious about generalizing these finding to other segments of the population. Future research might investigate the possible moderating impact of age and/or social status on online consumer behavior.The third limitation is the manipulation of service categories. We did pre-test the services in each category. However, the model can be tested with more services in each category to validate the results. Also, other product classification could be used to expand our understanding of online services marketing.\n",
"role": "user"
},
{
"content": "Purpose: - The main objective of this paper is to compare consumers' online shopping behavior across three types of services (i.e. search, experience, and credence). Reviewing the marketing and psychology literatures, this study aims to propose that consumers' emotions (pleasure, arousal, and dominance) influence their perception of site atmospheric cues (site informativeness, effectiveness, and entertainment), which, in turn, impact consumers' site attitudes, site involvement, and purchase intention. It also aims to test the proposed model for three major types of services (i.e. search, experience, and credence) and to compare the path coefficients of all the relationships in the model across the three groups.\nDesign/methodology/approach: - Lab experiments were conducted for data collection and structural equation modeling was utilized for multi-group analysis.\nFindings: - The results supported the proposed model and revealed several non-invariant structural paths across the three groups.\nResearch limitations/implications: - The student sample may not represent the general population.\nPractical implications: - Search-, experience-, and credence-based services should attempt to evoke the most desired consumer emotional types (pleasure, arousal, and dominance).\nOriginality/value: - Unlike many other studies in services marketing, this paper tests the proposed model across different service types to increase the generalizability of the results.\n",
"role": "assistant"
}
] |
Interpersonal complaints regarding cancer care through a gender lens
|
[
"Patient perception",
"Value co-creation",
"Service quality",
"Patient complaints",
"Cancer patients"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Many of today's healthcare systems face the challenge of an increasing number of older patients who suffer from multiple illnesses that require more resources (Socialstyrelsen, 2013). Hence, costs are rising while the tax-base is decreasing (Lifvergren, 2013). It is estimated that one out of three Swedes will develop cancer during their lifetime (Socialstyrelsen and Cancerfonden, 2013). This number is predicted to double by 2030 due to the aging population, but also because of improved screening programs, refined diagnostic methods, and increased awareness about cancer among the population (SOU, 2009/2011). Out of the Swedish population of 9.7 million (Statistic Sweden, 2014), 55,000 are annually diagnosed with cancer; more than 60 percent are cured and many others can continue to live with cancer (Cancerfonden, 2013). The positive development has also led to an increasing number of treatments; however, it has also been recognized that the understaffing of oncologists and nurses specializing in oncology leads to heavy workloads and time constraints (Henriksson and Bergfeldt, 2011). This staff shortage has been noticed locally in Western Sweden (Lakartidningen, 2013), where 80,000 people are living with cancer and a further 10,000 are diagnosed with the disease annually (Regional Cancer Centre West, 2014).\nTheoretical framing: Normann and Ramirez (1993) referred to the customer as a value creator, who actively integrates resources such as skills and knowledge (Vargo and Lusch, 2008). Only the customer can create value; the provider can only offer value propositions (Vargo and Lusch, 2008). Similarly, value emerges during customer use as value in use (Vargo and Lusch, 2008). In particular, interactional and relational aspects are emphasized in co-creation (Gronroos, 2011; Vargo and Lusch, 2008). Normann (2001) referred to these interactions as moments of truth, because face-to-face interactions are particularly important in co-creating value. Tronvoll (2007) suggested that helpfulness, empathy, care, courtesy, and friendliness are essential elements in such interactions. The customer evaluates the service interaction, as well as the value in use, of the provider's value proposition (Vargo and Lusch, 2008).\nMethods: Setting\nResults: Table II shows the six interpersonal complaint subcategories. Ten of the complaints were sorted under more than one subcategory.\nDiscussion: Despite the emphasis on the importance of proper interaction in healthcare (Berry and Bendapudi, 2007; Gallan et al., 2013; Holman and Lorig, 2000), including Swedish cancer care (SOU, 2009/2011), this paper reveals unsatisfactory service meetings in the provider-patient/relative interface. In particular, the results of the empirical findings revealed a lack of information from healthcare providers, as well as healthcare customers experiencing insufficient consideration. Female patients in particular experienced a lack of healthcare staff civility. For complaints filed by relatives, it was particularly evident that they did not feel included in patient care and were not offered proper support during hard times. Most complaints by relatives were filed by females.\nConclusion: The findings of this paper suggest that a lack of information and consideration from healthcare providers is a common source of dissatisfaction in cancer care. Furthermore, a lack of empathy and civility also caused dissatisfaction. More than half of the complaints were filed by relatives - most commonly a female relative to a male patient - dissatisfied with their lack of inclusion in the care process and the lack of proper support during hard times.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate healthcare customer complaints concerning interpersonal matters in cancer care.\nDesign/methodology/approach: Complaints from cancer patients and their relatives (n=116) that dealt with interpersonal matters registered between 2009 and 2011 at four local Patients' Advisory Committees in Western Sweden were sampled and analyzed using qualitative content analysis.\nFindings: Complaints concerned lack of information and consideration from healthcare providers. Lack of empathy and civility also caused dissatisfaction, the latter particularly for women. Relatives complained that they did not feel included in the care process or were not offered proper support. Most complaints by relatives were filed by a female relative and concerned a male patient.\nResearch limitations/implications: Information about patient demographics other than gender could not be investigated due to database limitations. Hence, factors such as age, country of birth, and geographical residence were not included for analysis. In addition, neither the type nor stage of cancer among the sampled patients was able to be addressed.\nPractical implications: Patient complaints should not only be viewed as a post-consumption judgment, but also as a service interaction activity. This may require healthcare providers to enhance their interpersonal skills, allowing patients and relatives to provide feedback during service interaction to satisfactorily address dissatisfaction. Visualizing gender disparities may help healthcare providers prevent stereotypical encounters. In addition, the provider should be invited to participate in the customer's value creating network, which may also include knowledge and skills from other sources, such as relatives.\nOriginality/value: Value co-creation offers a different view on patient complaints. Incorporating social construction into value co-creation may reveal socially constructed disparities. The paper provides aggregated information on cancer patients' and relatives' complaints concerning interpersonal issues, which can increase knowledge about patient healthcare service perceptions.\n",
"role": "assistant"
}
] |
Effects of 3D vs 2D interfaces and product-coordination methods
|
[
"2D interface",
"3D interface",
"Mockshop",
"Product coordination"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Currently, retailers have taken significant interest in developing three-dimensional (3D) interfaces because of their highly attractive features (Brandon, 2007). Lee and Chen (2011) stated that the realistic layout of a 3D interface accurately parallels the real world physical environment in consumers' mental representation. Consumers could get a tangible brand experience in a 3D interface because it mimics the consumers' engagement in a physical store (Gabisch and Gwebu, 2011). Moreover, a 3D product presentation reduces perceived risk (Shim and Lee, 2011) because consumers could get a better idea about colour, texture and fit when the 3D virtual model is presented instead of the 2D image. Moreover, a 3D product display method allows consumers to examine products from various angles (Khakimdjanova and Park, 2005). Alpcan et al. (2007) claimed that we are moving towards a 3D internet. On the other hand, navigation in a 2D environment is difficult because it presents a lot of information in an abstract way.\nLiterature review: Stimulus-organism-response framework\nMethodology: Experimental design\nResults: Sample characteristics\nDiscussion and conclusion: The results showed that this study partially supported the S-O-R and flow theories. For example, 3D and 2D interfaces showed significant differences on flow, perceived merchandise quality, satisfaction, time spent and patronage intentions. Also, product display methods based on colour, discount and brand in both 3D and 2D interfaces had significant influences on most outcome variables. However, they did not have significant differences on consumers' emotional states: pleasure and arousal (in most cases). This finding is contrary to the Wu et al. (2013) study because the authors found that product display methods based on colour, style and texture had significant differences on consumers' emotional states. We reasoned that perhaps colour, brand and discounts are more commonly used display approaches in the retail industry than style coordination and texture and thus the differences among them are not as noticeable. Though the 3D interface contains richer multi-sensory elements than the 2D interface, it failed to provide consumers with significantly more impactful emotional experience than the 2D interface. This might be because consumer might concentrate more on examining apparel items rather than focusing on the rich, sensory attributes of the 3D interface since this experiment is a task given to the subjects.\nLimitations and future research: One of the limitations of this study is that the 2D interface is not as interactive as the real online shopping sites. For example, real online retailers such as GAP or Forever 21 provide many navigation options online to view and select apparel items such as online pathfinding assistance, search engine, site map and guide bar at the top or bottom of the page. Therefore, future research could develop a more interactive 2D interface to compare it with 3D interface. Another limitation of this study is that the subjects were given a shopping task, which differs from a real shopping task with a budget and perhaps a shopping list in mind. Also, though all our participants are college students in the USA, they came from methods of recruitment, which could potentially skew our results. Further, there are many extraneous variables that we cannot control, such as the environment that participants were in when responding to the 2D scenarios, participants' prior knowledge about online or 3D shopping, etc. These are shared limitations for experiments involving self-selected online participants.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the effects of product display based on colour, discount and brand on retail pleasure, arousal, flow, perceived merchandise quality, time spent, satisfaction and patronage intentions in both 3D and 2D interfaces. This study uses the flow theory and stimulus-organism-response framework to understand the effects of product coordination methods and interface dimensions on consumers' emotional, cognitive and behavioural outcomes.\nDesign/methodology/approach: This study conducted a 2 (interface: 3D and 2D) x 3 (product coordination method: colour, discount and brand) between-subjects experimental design. The coordination of display refers to the way products are arranged based on similar properties, such as similar colour, discount or brand. 3D stores were designed using Mockshop, a 3D virtual retailing software. A total of 144 US female college students from various universities participated in the study.\nFindings: The findings showed that the 3D interface produced higher flow, satisfaction, time spent and patronage intention than the 2D interface. Also, among the six display scenarios, participants who were in the colour-coordinated 3D store showed the highest levels of emotional, cognitive and behavioural outcomes except for retail pleasure and arousal.\nOriginality/value: This study investigates product display based on colour, discount and brand in both 3D and 2D interfaces. Although product coordination methods have been tested in 3D stores in previous studies, they have not been compared with their effects in the 2D interface such as online shopping sites. Therefore, this study fills in this gap in the literature, which can guide retailers in making the right decisions for visuals.\n",
"role": "assistant"
}
] |
The upside of cultural differences: Towards a more balanced treatment of culture in cross-cultural management research
|
[
"Positive organizational scholarship",
"Cross-cultural management",
"Cultural diversity",
"Positive outcomes"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Cultural differences are a nuisance at best and often a disaster(Geert Hofstede (www.geert-hofstede.com)).\nToward a more nuanced understanding of culture in CCM research: Recent research has drawn attention to the potentially positive role of distance and diversity across national, cultural, institutional, and organizational dimensions. Advocating the study of how diversity, foreignness, and distance creates value for global organizations is gaining nascent appreciation (e.g. Brannen, 2004; Edman, 2009; Mezias, 2002; Stahl et al., 2016; Tung and Verbeke, 2010; Zaheer et al., 2012). These scholars argue that the overemphasis on adverse outcomes associated with cultural differences in existing international business research hinders our understanding of the processes and conditions that help organizations leverage the benefits of diversity in a wide range of contexts. These benefits may be exhibited through the development of strategic capabilities, decisions on foreign direct investment, synergy creation in cross-border mergers and acquisitions, learning through cross-border knowledge-sharing, unleashing of creative potential in diverse teams, and designing and implementing corporate social responsibility (CSR) initiatives that meet the needs and demands of a diverse set of stakeholders spread across the globe.\nOverview of the special issue: Our call for papers generated 35 submissions representing authors from 15 countries - an indication that our invitation to explore \"The upside of cultural differences: towards a more balanced treatment of culture in cross-cultural management research\" struck a chord with international business and CCM scholars. Out of this set of submissions, we accepted five for publication in this special issue. These five papers address a wide range of issues, ranging from the micro (individual) to the macro (national) levels of analysis. as well as cross-level phenomena, and they include both theoretical and empirical papers employing quantitative, qualitative, and mixed-methods approaches. In addition to the five papers we include a Distinguished Scholar Essay by Kim Cameron, the Founder of the POS movement. Also included is an Interview Piece with Richard Nisbett, a Pioneer Researcher in culture and cognition. In total, seven papers comprise this special issue.\nConclusion: With the advent of culture clashes and the focus on the downside of culture in international business, often the positive aspects related to CCM are neglected. This special issue seeks to highlight the emergent topic of the upside of culture research and its application in business and education. Based on our review of 35 submitted and five accepted papers, we were able to draw some compelling conclusions. The breadth and depth of the papers received were affirming that a broad interest on the topic of the upside of culture is emerging across many fields and disciplines. The boundaries of a special issue dictate that the main ideas need to focus on the intended topic and that the papers must be developed within the timeframe for publication. As such, some papers were found to be outside of the perimeters of the special issue and other papers simply needed more time to develop than our special issue allowed. The remaining group of papers were high in quality and wide-ranging in scope. This signaled to the editorial team that investigating the upside of CCM was both warranted and well-timed. Working with expert reviewers is an important element of any successful manuscript review process and we found this to be exceptionally true for this special issue. We would like to thank all our manuscript reviewers, as their feedback and expediency were crucial for the authors.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to encourage scholars to look at commonly considered phenomena in international business and cross-cultural research in new ways and to theorize and explore how cultural diversity, distance, and foreignness create value for global organizations. These considerations should result in a more balanced treatment of culture in cross-cultural management (CCM) research.\nDesign/methodology/approach: The idea that there are negative consequences associated with cultural differences is pervasive in hypotheses formulation and empirical testing in international business and CCM literature, as reflected in widely used constructs such as \"cultural distance,\" \"cultural misfit,\" \"foreignness,\" and related concepts. Consistent with a Positive Organizational Scholarship (POS) perspective on culture and cultural differences, the authors emphasize the positive role of distance and diversity across national, cultural, institutional, and organizational dimensions. In addition, they provide an overview of the contributions to the special issue.\nFindings: Examining the positive side of culture is not only beneficial theoretically in terms of filling the existing gaps in the literature, but is also crucial for the practice of international and global business. Accordingly, the contributions to the special issue highlight how explicitly considering positive phenomena can help better understand when and how cultural diversity, distance, and foreignness can enhance organizational effectiveness and performance at multiple levels. They include five research papers, a Distinguished Scholar Essay by Kim Cameron, the Founder of the POS movement, and an interview piece with Richard Nisbett, a Pioneer Researcher in culture and cognition.\nOriginality/value: The overemphasis on adverse outcomes associated with cultural differences in existing research has hindered the understanding of the processes and conditions that help organizations leverage the benefits of cultural differences in a wide range of contexts. This introductory paper together with the contributions included in the special issue showcases the positive dynamics and outcomes associated with cultural differences, distance, and diversity in a wide range of international business contexts.\n",
"role": "assistant"
}
] |
Achieving luxury status through delivering a culture of uncompromising quality
|
[
"Empowerment",
"Quality awareness",
"Change management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. The rationale for embarking on productivity improvement: Whilst quality has always been recognised as important at Glenmorangie, and key to the success of our brands, other important measures had sometimes been given greater focus e.g. OTIF (on time in full) or OEE (overall equipment effectivity). That changed when Glenmorangie plc was purchased by Louis Vuitton Moet Hennessy in February 2005.We moved from being a small independent whisky company into the portfolio of the world's largest luxury brands company. Whereas we had previously sold primarily to retailers and focussed on compliance, now as part of LVMH, we are targeting premium bars, clubs and hotels, and this calls for products of uncompromising quality and visual perfection.Customer feedback showed that we did not consistently meet these exceptionally high standards over the full range of products which are made at Broxburn, and we found that our internal records did not match the list of issues which we knew to have occurred. Operators were not clear on what was/was not acceptable, different people had different perceptions and the goal posts kept changing. It was evident that we needed to have renewed focus on quality as a major deliverable of our production.In order to achieve luxury status our standards needed to be reset against this higher benchmark in all processes and activities that impact on the finished product quality. We needed to develop and embrace a higher quality culture throughout the organisation.\n2. The opportunities for improvement identified and investigated: A quality review was carried out across site, with interviews conducted with staff from all levels of the business. A gap analysis on the results showed that we did not have a culture in which we strove for perfection every time; there were differences in individuals' understanding/awareness and no set standards. Quality training was only carried out when an employee joined the company and it very much focussed on compliance rather than the aesthetics of the product (e.g. whisky strengths, contamination risks, etc.). There was no process for ensuring that operators fully understood this training and no refresher training was carried out.Our existing performance management competency framework was reviewed and we realised that it needed to be simplified and updated to reflect our new principles. The process itself is well established and is based around technical and personal competencies. Previously there were six personal competencies, which applied to everyone, and up to eight technical competencies, which differed across the company according to the job, for example:* personal - pleasing the customer; and* technical - standards.\"Standards\" was a generic technical competency and did not differentiate between the different aspects of compliance, heath and safety, etc.Each of the competencies was split into five levels, with the rationale that we did not all have to develop competencies to the same level. Line leaders carried out reviews annually with all employees and development plans were agreed. The Training & Development Manager collated the training identified and although line leaders had carried out the reviews, they did not take ownership of the process - i.e. they rarely carried out regular reviews with their team to ensure actions were implemented or monitor their progress. As a result quality issues tended to be seen as the responsibility of the Quality Operators (one on each line).In order to overcome this, the operator competencies were restructured to focus on quality and the new Luxury Principles competencies created - Visual Perfection and Quality Compliance. In conjunction with this we realised that two vital areas (health and safety, and the environment) also needed to be reinforced through additional training and then sustained through the introduction of specific competencies. These became the technical competencies for all operators in Bottling. By understanding these competencies and what they involved, we were able to design a training program that would meet their needs, so ensuring that operators were able to achieve the required competence levels.We recognised that the change in culture would not come about by training alone: we needed to develop a process of ongoing assessment to ensure that operators know whether their level of performance is at the correct standard and are encouraged to maintain this.\n3. How productivity improvements were planned and reviewed: Our aim was stated as:This program will drive a step change in quality awareness within the business. All personnel are fully trained and competent in their roles and fully understand why we are embarking on this programme, and what needs to be delivered to achieve visual perfection.The key steps in the programme were:* training identified;* training delivered;* internal quality processes understood;* visual perfection understood across all roles; and* competence of all staff assessed.First we developed the training for our new luxury principles competencies. Uncompromising quality compliance training covered the following:* legal requirements;* health and safety;* environment;* standards; and* training and development.Visual perfection training covered:* understanding what visual perfection means within the Glenmorangie Company;* your role in achieving visual perfection;* visual perfection from Goods Inwards to Despatch;* new finished product specifications; and* training and development.It was vital that those carrying out the training were the recognised champions in their respected areas and were seen as credible. The training itself was applicable to everyone in the Operations area, so all managers (including the Operations Director) would need to participate in it, with the same exacting standards being expected from all. Every training programme would be introduced by the Senior Manager for that area with clear direction as to why this was so vital for the company and for their team.In order to ensure that all staff were competent in these areas, assessments were to be carried out immediately after the training. Quality compliance was assessed using a written test, while visual perfection used a defect spotting exercise. The pass mark was set high (92 per cent), as this had to be at a level that raised the bar and set expectations. Anyone who did not achieve this pass mark would be given further training on a one-to-one basis.We prioritised the bottling teams as critical to us producing both compliant and visually perfect products, and therefore these were the priority area to go through this Quality Culture Luxury Principles training.A process of ongoing assessment was developed whereby line leaders, technical trainers and manufacturing engineers carry out periodic assessments with all bottling staff to ensure that the training is being followed in practice and that the high standards are maintained. Personal competencies such as Motivating Yourself and Others are also included in these assessments, in recognition of the fact that correct attitude is key to out success. Competency checklists were created for these (see Figure 1 for an example), one for each of the technical competencies, showing clearly the standard to be achieved.\n4. The implementation of the improvement plan: Actioning such a large-scale training programme required all areas to work together collaboratively to ensure that releasing operators for training did not impact on production. This also required a huge commitment from senior management to support the programme - both in their attending the training themselves (the Operations Director was on the first Visual Perfection course with the top score of 95 per cent) and in senior management leading the training on each new programme - with an introduction as to why this was so important for the company and the need for change. We needed to get across to operators that we were now expecting them to stop the line, challenge if required and take responsibility for ensuring what they produced met the highest standards. This message needed to be made many times and senior managers had to back this up on a day-to-day basis as operators tested this out.In Bottling, 90 per cent of those requiring training have been completed (with the further 10 per cent being addressed through individual one-to-one training sessions). This amounts to 297 training instances out of the 331 expected. Overall, 159 operators were assessed as competent after training and then those receiving coaching were reassessed and this increased to 267.The time involved was 796 hours, involving 94 hours of trainer time, 594 hours of operator time, 54 hours of operator coaching for re-sits plus 54 hours of trainer time for coaching re-sits.All line leaders, technical trainers and manufacturing engineers were trained as assessors to ensure they were clear as to their role and the importance of gathering ongoing evidence of competence. These ongoing on-the-job assessments ensure that the training is reinforced and is evaluated using the competency checklists, so providing evidence of operators competence, and any actions required are noted for the next review.The overall number assessed as competent at the first on the job competency assessment was 44 per cent. Time involved in assessments was 294 hours.The sustainability of this programme, and how to integrate this training into standard work practices, required creative thinking from the trainers. As other work priorities impinged on their time, we needed to develop another method to deliver this training.This resulted in the use of e-learning, and we are now developing training modules ourselves that will be easily accessed by employees and can ensure that all new operators have the appropriate training at the correct time. This will include online assessments which will be automatically sent to the trainer who will then be able to follow up with a reassessment if required.\n5. The evaluation of improvements implemented: Operators are now picking up on defects and stopping lines for problems, which in the past had been allowed to run. They are demonstrating increased confidence in their authority and competence, and this is reflected in the increase in the number of non-conformance reports raised. This is driving improvements online, which is therefore improving capability and allows us to produce visibly perfect products.The periodic assessments with all bottling staff ensures that operators know whether their level of performance is at the correct standard and are encouraged to maintain this; if it is below the standard expected, actions are identified to support improvements and timescales agreed for further review. Ongoing assessment and feedback has become part of the day-to-day culture, where quality and visual perfection are clearly understood and their achievement is seen as an operator's prime goal. As a result operators feel more empowered and having regular feedback on their performance ensures they are recognised for improving their performance and are clear on what is expected of them.Our success criteria to deliver the expected results for the business were set at the outset for the project and are shown in Table I.The financial benefits to the business are clear: In 2006 the cost of poor quality in Bottling was PS230,155. The cost of training was 1,090 hours at an hourly rate of PS15 (=PS16,350). The cost for the first six months of 2007 was only PS35,000.The programme has clearly met its aim of driving a step change in quality awareness throughout the business and has involved a culture change in the way performance is measured and reviewed. It has resulted in operators having increased confidence in their authority and competence and the quality results reflect this. Quality is now recognised as the number one priority throughout the business.After six months of continuous improvements, the company organised an employee conference, where all employees attended this company-wide communication. Here the company presented its \"New World\" range of luxury products. Every employee was able to see and sample the new products for themselves in a variety of formats, all within a luxury environment. This ensured employees could see what was now being achieved as a result of all their efforts and how this related to the luxury market.This programme was recently selected as one of three finalists in the Institute of Grocery Distribution's John Sainsbury Award for Learning and Development and was runner-up.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to share the learning gained from the experiences of The Glenmorangie Company in driving a step change in quality awareness within the business.\nDesign/methodology/approach: - The approach is taken in terms of raising standards in the organisation through a competency training and development programme, using a top-down approach to deliver against stretching business targets.\nFindings: - The approach used at The Glenmorangie Company echoes the findings documented in the CIPD research paper \"Managing change: the role of the psychological contract\".\nPractical implications: - When considering embarking on any programme requiring a step change in business performance where a change in culture is vital, this approach will enable you to optimise its success, gain employee support to ensure sustainability and improve both employee and business performance.\nOriginality/value: - This case study will be of value to all those involved in improving business performance, providing a clear, structured approach to the process of building effective learning interventions based on a clear understanding of the need to change.\n",
"role": "assistant"
}
] |
Adult third culture kids: adjustment and personal development
|
[
"Adjustment",
"Development",
"Expatriate",
"Adult third culture kids"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The emergence of new types of careers in global organizational contexts means that personal development has become central for individuals (Banai and Harry, 2004; Baluku et al., 2018; Baruch and Rousseau, 2019; Mutter and Thorn, 2019; Redondo et al., forthcoming). This notion suggests that managers value expatriation for the opportunity it brings for personal development, despite that such personal development may not help them advance within their employing organization (Stahl et al., 2002). As such, the international experience has been shown to hold high intrinsic value and can be used by individuals as an asset to further their personal development (Doherty et al., 2011; Lauring et al., 2014; Quer et al., 2019). Many expatriates cite as their reasons for accepting international assignments not their desire to achieve project results, to be good corporate citizens, or to advance their company careers, but their personal development (Inkson et al., 1997). Dickmann et al. (2008) dismissed financial rewards as being the primary source for relocation motivation; instead, development and career considerations impacted the decision the most. Similar results were found by Thorn (2009) and McNulty et al. (2013). Although the cross-cultural context is particularly fertile ground for personal development and transformation, not all expatriates take advantage of the opportunity for their own personal growth.\nBackground: In general, scholars seem to agree that developmental learning experiences gained through interactions with the foreign location in various situations influence expatriate adjustment and thus expatriate success (Caligiuri and Tarique, 2012). The earlier these experiences are introduced to and experienced by the individual, the longer-lasting and deeper the learning and impact are likely to be (Tarique and Weisbord, 2013). Hence, the experience obtained in a foreign culture may trigger \"personal development,\" defined in a work context as a purposeful, specifiable and structured activity, which seeks to develop discrete skills or qualities, whose effects can either be \"positive,\" \"negative\" or \"transitory\" (Irving and Williams, 1999). Others define personal development as having an improved awareness and identity, and the development of talent and potential that enhances the quality of life and contributes to the realization of dreams and aspirations (Aubrey, 2010).\nHypotheses: Interaction adjustment and personal development\nMethod: In this study, we targeted European business expatriates working in Asia, a region that is known to differ substantially in cultural and linguistic contexts from most European countries, which makes Asia especially challenging for expatriates (Shi and Franklin, 2013). The aim was to explore the moderating effect of ATCK vs AMCK status on the direct relationship between interaction adjustment and experienced personal development. Participants were obtained from a list of names and e-mail addresses acquired from embassies and other databases of white-collar foreign workers in Asia. Data were collected using an English language online survey. Responses came mainly from Singapore (23.32 percent), Indonesia (17.31 percent), Malaysia (12.72 percent), China (11.66 percent) and Hong Kong (9.54 percent). Other host countries included Thailand, Philippines, Vietnam, India, Bangladesh, Lao and Pakistan. Expatriates originated from Sweden (13.07 percent), Denmark (13.07 percent), Norway (12.01 percent), Germany (10.25 percent) and the UK (8.83 percent). In total, 455 survey responses were received from an original database of 2007 potential participants (a 22.7 percent response rate), from which a pool of 424 respondents was obtained after screening for incomplete surveys.\nResults: Our two hypotheses were investigated using hierarchical multiple regression analysis with results in Table II. Control variables were included in Step 1 resulting in significant relationships with three of the five control variables. Number of different jobs had a positive association with personal development (b=0.06; p<0.01), while Gender had a negative association with Personal development (b=-0.26; p<0.05). In Step 2, the interaction adjustment variable was added. Interaction adjustment had a positive association with personal development (b=0.08; p<0.05). In Step 3, the moderator variable ATCK/AMCK was included (ATCK=1 and AMCK=0). The moderator variable was weakly and negatively associated with the criterion variable (b=-0.17; p<0.10). In Step 4, the interaction term was entered. Interaction adjustment x ATCK/AMCK was negatively associated with personal development (b=-0.17; p<0.05). All F values were statistically significant, indicating a proper fit between the regression model and the data.\nDiscussion: In this study, we have examined the relationship between interaction adjustment and experienced personal development among ATCKs and AMCKs. As predicted, our findings show a positive association, which may be related to arguments that intercultural experience leads to personal development (Bankston and Zhou, 1995; Portes and Hao, 2002), but the effect of interaction adjustment varies. On the one hand, AMCKs experience personal development almost independently of their interaction adjustment with host nationals solely due to the novelty of the international experience. In contrast, while ATCKs experience less personal development when there is low interaction adjustment and more personal development when interaction adjustment increases, just being in the new country is not enough for them to feel they have developed personally; given their prior international experience as adolescents, they need to engage more deeply with the local population to experience personal development as the foreign setting in itself is not challenging enough. This finding could be explained by goal setting theory (Locke and Latham, 1990) predicting that relatively easy endeavors will not lead to much personal development.\nTheoretical implications: This research contributes to the literature on the relationship between adjustment and personal development and between international experience and adjustment, thus combining the two fields. In the first instance, the study supports the theoretical argument that adjustment leads to personal development (Haslberger et al., 2013), by empirically demonstrating a link between interaction adjustment and personal development. Second, prior studies (e.g. Takeuchi and Chen, 2013) have suggested there is a negative relationship between prior international experience and expatriate adjustment. In our research, while we have not focused on a direct effect of international experience but the moderating effect of having an early international experience indicated by ATCK status, we nonetheless find little direct effect of ATCK status. More interesting, however, is that early international experience in the form of ATCK status has an indirect effect influencing the relationship between interaction adjustment and personal development. In other words, the role of interaction adjustment changes for those with early international experience in that higher levels of interaction are required for ATCKs, than for AMCKs, to achieve a similar impact.\nPractical implications: We still know very little about ATCKs and about how expatriation during their adulthood develops them personally. Our study contributes to the growing literature linking global talent management to individuals' personal characteristics (Caligiuri et al., 2009; Tarique and Weisbord, 2013) by demonstrating differences in perceived personal development among ATCKs and AMCKs. This has implications in practice for the depth of contact expatriates should, or could, have with host nationals in the international work environment. For example, ATCKs may be the \"preferred\" type of expatriate when the nature of the job abroad requires close contact (interaction) with HCNs to fulfill its purpose. As ATCKs' personal development is closely linked to higher levels of HCN interaction, the likelihood of their success abroad increases when personal development correspondingly increases. Alternatively, when close contact with HCNs is not required, the talent pool can be widened to include AMCKs, with no perceived drawback.\nLimitations: As in all studies, ours has several shortcomings that limit generalizability. Cross-sectional self-reports, for example, may have created common method variance (CMV). However, we took a number of measures such as preventing respondents to go back to previous pages of the survey, and assuring confidentiality (Podsakoff et al., 2003). Most importantly it has been argued that moderation effects also cannot be caused by CMV (Chang et al., 2010). As our main finding relates to the moderating effect of ATCK vs AMCKs, it could be argued that CMV is not a serious problem.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore how interaction adjustment influences personal development for expatriates and to examine whether the effect differs between adults that have, and have not, lived abroad during their adolescence.\nDesign/methodology/approach: The authors use survey responses from 424 business expatriates in Asia distinguishing between adult third culture kids (ATCKs) that have lived abroad during their adolescence and adult mono-culture kids (AMCKs) who have not.\nFindings: The results show that while interaction adjustment generally improves the experience of personal development, this effect is stronger for ATCKs. AMCKs will experience personal development almost independently of their interaction adjustment with host nationals solely due to the novelty of the international experience. For ATCKs, just being in the new country is not enough for them to feel they have developed personally; they need to engage more deeply with the local population to achieve this.\nOriginality/value: The authors still know very little about ATCKs and about how expatriation during their adulthood develops them personally, given they have already had international experiences at a young age.\n",
"role": "assistant"
}
] |
New approach towards formulation of the overall equipment effectiveness
|
[
"Analytical hierarchy process",
"Overall equipment effectiveness",
"Six Sigma",
"Maintenance process",
"Time driven activity based costing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Due to increasing importance in measuring productivity machine in an industry, the adequate formulation of the overall equipment effectiveness (OEE) has become a major necessity for increasing the maintenance process performance. OEE is a measurement tool used in identifying and measuring productivity machines in an industry (Huang et al., 2003). OEE is simply made up of three elements known as performance rate, availability rate (Olayinka and Leramo, 2012) and quality rate. It helps identify potential losses and how corrective actions could be used to reduce it (Bulent et al., 2000). OEE provides a world-class status as follows:\n2. Literature review: In this section, the literature review is organised in six parts. The first part deals with the evolution of the OEE formulation, it could be approached through three-dimensional aspects, which are the temporal dimension, the financial dimension and the quality dimension based on the PA. The current OEE formula has two different aspects: the first aspect only emphasises the time variable to establish the comparison between the time spent producing good parts and that theoretically expected, and the second aspect consists of three-component multiplication: availability rate, performance rate and quality rate.\n3. Research methodology: In this study, the methodology followed is based on three major steps (see Figure 2).\n4. Conclusion: In this article, we presented contributions to the problem of measurement and analysis of productivity via a new formulation of the OEE based on the PA in maintenance. The first contribution concerns a new formulation of the OEE; maintenance was approached a set of sub-processes and not as a function. The second contribution deals with the use of TDABC and PERT methods to formulate the temporal and financial components of the OEE-TCQ. The third contribution concerns the quality component formulation of OEE-TCQ by using the AHP method and quality's improvement (Six Sigma).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to propose a new formulation of the overall equipment effectiveness (OEE) relying on the process approach (PA) in maintenance. This new indicator will be designated by OEE-TCQ.\nDesign/methodology/approach: Methodology followed in this paper is based on using two major parts. The first part deals with the use of time-driven activity-based costing (TDABC) and program evaluation and review technique (PERT) methods to formulate the temporal and financial components of the OEE-TCQ. The second part concerns the quality component formulation of OEE-TCQ by using the analytical hierarchy process (AHP) method and quality's improvement (Six Sigma).\nFindings: It is a new formulation of the OEE, based on the PA applied to operational maintenance processes, and also is interesting for improvement maintenance.\nPractical implications: The new OEE-TCQ indicator can be used to support maintenance managers to have a three-dimensional analysis based on time, financial and quality criteria, also to improve time, cost and quality of each activity of maintenance operational processes.\nOriginality/value: This paper is one of few papers in maintenance proposing a new formulation of the OEE by using the combination of several methods especially TDABC, PERT, AHP and Six Sigma. It proposes in part a new quality evaluation approach for each activity of maintenance operational processes. Also, it suggests a new formulation for defects by millions of opportunities.\n",
"role": "assistant"
}
] |
Indirect cost of maternal mortality in the WHO African Region, 2013
|
[
"African region",
"Discounted value of potential non-health GDP loss",
"Maternal mortality"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nBackground: WHO defines maternal death as the death of a woman while pregnant or within 42 days of termination of pregnancy, irrespective of the duration and site of the pregnancy, from any cause related to or aggravated by the pregnancy or its management but not from accidental or incidental causes (World Health Organization (WHO), 2010a, b, c). Out of the 289,000 maternal deaths that occurred globally in 2013, 170,913 (59 per cent) were borne by the WHO African Region (World Health Organization (WHO), 2013a, b). About 28 per cent of the Region's maternal deaths were from pre-existing medical conditions such as diabetes, malaria, HIV and obesity exacerbated by pregnancy, 27 per cent from severe bleeding, 14 per cent from pregnancy-induced high blood pressure, 11 per cent from infections mostly after childbirth, 9 per cent from obstructed labour and other direct causes, 8 per cent from abortion complications, and 3 per cent from blood clots (World Health Organization (WHO), 2014a). Compared with other WHO regions, African Region's average maternal mortality ratio (MMR) of 500 per 100,000 live births was 7.4-fold that of the Americas, 29.4-fold that of Europe, 2.9-fold that of Eastern Mediterranean, 2.6-fold that of South-East Asia and 11.1-fold that of Western Pacific (WHO, 2014b).\nMethods: Conceptual framework\nResults and discussion: Table III presents the WHO African Region' population and maternal deaths by economic group in 2013.\nConclusions: The estimated potential non-health GDP loss attributed to maternal deaths of Int$5.53 billion is about 0.25 per cent of Int$2,226.39 billion, the combined 2013 GDP of the 45 African countries (IMF, 2014). The estimate signifies the loss of potential GDP in the future years of life lost among the 167,913 maternal deaths revalued relative to base year 2013, i.e. present values.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to estimate discounted value of potential non-health gross domestic product (GDP) losses attributable to the 167,913 maternal deaths that occurred among 45 countries in the WHO African Region in 2013.\nDesign/methodology/approach: A cost-of-illness method was used to estimate non-health GDP losses related to maternal deaths. Future non-health GDP losses were discounted at 3 per cent. The analysis was undertaken for countries categorized under three income groups.\nFindings: The discounted value of future non-health GDP loss due to maternal deaths in 2013 is in the order of Int$5.53 billion. About 17.6 per cent of that occurred in countries in the high and upper income group, 45.7 per cent in the middle income group and 36.7 per cent in the lower middle income group, and the average non-health GDP loss per maternal death was Int$136,799, Int$43,304 and Int$19,822, respectively.\nResearch limitations/implications: This study omitted costs related to direct health care, direct non-health care treatment, patient time for treatment, informal caregivers' time, intangible costs such as pain and grief, lost output due to morbidity, and negative externalities on the family and community.\nSocial implications: The study demonstrated that maternal deaths have a sizable negative effect on non-health GDP of the region, implying that maternal mortality is not only a human rights concern but also an economic issue and that universal coverage of maternal health interventions ought to be an imperative goal in all countries.\nOriginality/value: This paper provides new evidence on the impact of maternal deaths on non-health GDP of 45 countries in the WHO African Region.\n",
"role": "assistant"
}
] |
Does idiosyncratic risk matter? Evidence from mergers and acquisitions
|
[
"Acquirer returns",
"Idiosyncratic risk",
"Market imperfections"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Standard asset pricing theory predicts that idiosyncratic risk is associated with zero excess return, as investors are assumed to hold diversified portfolios (Sharpe, 1964; Lintner, 1965)[1]. The only way for them to increase their expected return is by taking greater exposure to systematic risk. However, market frictions can prevent investors from being fully diversified. Merton (1987) argues that investors have incomplete information, and thus, tend to select stocks for which they can easily get information. It follows that investors are exposed to idiosyncratic risk and accordingly, require compensation for holding that risk. Under that hypothesis, higher idiosyncratic risk is expected to be associated with a higher expected return.\n2. Theory and hypotheses: We first articulate the relation between change in idiosyncratic risk and change in firm value under two alternative hypotheses. We then present several ways through which corporate acquisitions can induce a change in idiosyncratic risk.\n3. Data and methodology: 3.1 General approach\n4. Empirical results: 4.1 Main result\n5. Conclusion: In this paper, our research strategy has been to focus on acquisitions to ensure a potentially large change in the acquirer's idiosyncratic risk and examine whether this change in risk is associated with a revaluation in the acquirer's stock price. A significantly negative relation implies that idiosyncratic risk is impounded in a firm's stock price and that higher idiosyncratic risk is associated with the higher expected return. Using a sample of 2,656 acquisitions undertaken by Australian firms, we find evidence that idiosyncratic risk matters. This result appears to withstand a range of robustness checks. It is unaffected by the incidence of the 2007-2008 financial crisis and holds using alternative measures of idiosyncratic risk. It is resilient to the presence of outliers and more significant after excluding the resources sector to which the Australian economy is highly dependent.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to investigate whether idiosyncratic volatility is a priced risk factor in the Australian stock market.\nDesign/methodology/approach: The authors use the change in idiosyncratic volatility around acquisition announcements and the related stock price revaluation to test whether the idiosyncratic risk is priced. If the idiosyncratic risk is priced, increases (decreases) in idiosyncratic volatility should be associated with decreases (increases) in the acquirer's stock price, as the latter's future cash flows are discounted at a higher (lower) rate. The sample consists of 2,656 completed acquisitions by Australian listed firms over the period January 1990 to October 2014 for which deal value represents more than 5 per cent of the acquirer's market value.\nFindings: Increases (decreases) in idiosyncratic risk are associated with significant decreases (increases) in firm value. This negative relationship is robust to the presence of outliers; is unaffected by the incidence of the 2007-2008 financial crisis; holds using alternative measures of idiosyncratic risk; and is more significant after excluding the resources sector. Firms with a higher idiosyncratic risk prior to the acquisition, and firms avoiding stock to pay for the acquisition, experience a more significant stock price increase in relation to a decrease in idiosyncratic risk.\nResearch limitations/implications: Considering the small size of the Australian economy, investors may have less scope to mitigate idiosyncratic risk. As a consequence, idiosyncratic risk is associated with the positive excess return, contrary to what standard asset pricing theory assumes. The results support Merton's (1987) hypothesis that investors are exposed to idiosyncratic risk due to imperfect portfolio diversification and receive compensation for bearing that risk.\nPractical implications: The pricing of idiosyncratic risk may also explain why the Australian stock market has historically offered a high equity risk premium. A practical implication would be for international investors to take advantage of the diversification constraints of local investors to capture higher risk premiums and achieve superior returns.\nOriginality/value: While prior studies demonstrate that stocks with higher idiosyncratic risk are associated with higher subsequent returns, the authors show that an increase in idiosyncratic risk is associated with a decrease in stock prices using acquisition announcements as shocks to a firm's idiosyncratic risk. In other words, the results arise from within-firm variations rather than from cross-sectional differences in stock returns.\n",
"role": "assistant"
}
] |
Cardiac surgery productivity and throughput improvements
|
[
"Productivity rate",
"Heart",
"Surgery",
"Health services",
"Finland"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Healthcare providers in both the public and private sectors are facing increasing pressure to improve their cost efficiency and productivity. While process improvements in the healthcare sector have often been driven by the development of medical technology, the increasing cost of new technological solutions has lead to applying operations management techniques developed for industrial and service processes. Meyer's (2004) literature review shows that, on average, operating theatres operate at only 68 per cent of capacity. This is a rather low capacity utilisation rate, especially when combined with the cost associated with the long wait times for surgical procedures in many countries (OECD, 2001), emphasises the importance of improving management practices in operating rooms. Our primary objective is to analyse the effect that six different process interventions have on the productivity of operating theatres performing open-heart surgery. However, effectively scheduling open-heart surgery is complicated owing to the high variability and inaccurate procedure time estimates. There is a high variability in the length of cardiac surgery and combined with a four and half hour average duration, it makes scheduling two surgeries during a normal eight-hour workday difficult. Having just one surgery a day results in low productivity. Thus, our second objective is to develop a model for estimating the length of procedures. We also consider whether increased estimation accuracy can improve operating room productivity. First, we review literature on management practices applied to improve operating room productivity. We further focus on how estimating procedure times influence operating room productivity. Second, we derive concrete research questions from the objectives and discuss the method. Third, we created a linear regression model to determine how accurately procedure times can be forecast. Fourth, we estimate the implication of more accurate procedure time estimates and five other process interventions on operating room productivity using a discrete event simulation model. Finally, the theoretical and managerial implications of our findings are discussed.\nImproving operating room productivity: One productivity definition is the ratio of a volume measure of output to a volume of input (OECD, 2001). According Schmenner (2004), service process productivity is dependent both on the throughput time and variability in the time it takes for various tasks to be completed. In the operating room, with a fixed number of tables and personnel reserved for a specific specialty in a given time period, productivity can be measured as a throughput volume. Two ways to increase operating room productivity, therefore, are to speed up the actual operation times or improve operating room utilisation rate by minimising idle time. In this paper, we focus only on how to improve operating room utilisation rate by slashing unproductive idle time. The importance of effective management and scheduling of operating theatres is demonstrated by the number of studies focusing on analysing operating room performance from an operations management point of view. In the following, key findings from these studies are classified by the three ways that are used to improve operating room performance: resource allocation for different specialties, optimal scheduling of surgeries and design and coordination of operating room activities. All are used to increase the operational efficiency of the daily use of operating theatres.Resource allocation\nResearch question and method: Kuopio University Hospital (KUH) is responsible for treating severe illnesses that call for special expertise and technology in the hospital district of Northern Savo, Finland. The hospital district population is about 860000. The average yearly number of invasive cardiology procedures is 1,700 and the average number of open-heart surgeries is about 1,000. In 2003, KUH started a development project to increase the quality and efficiency of the coronary disease care processes. One focus was to increase productivity in the operating theatre, which was identified as a resource that limited open-heart surgery throughput. The hospital was not able to meet medically recommended waiting times for cardiac surgeries. For example, unstable angina pectoris patients are forced to wait six days for CABG surgery after an angiogram. Additionally, KUH offers coronary bypass surgeries for external hospital districts and it is loosing patients to its competitors. Improving coronary bypass surgery productivity, therefore, would directly increase hospital turnover and profitability.The specific features of open-heart surgery shape our research objective, as the length of open-heart surgery tends to be so long that accommodating more than one surgery in an eight-hour working day is a challenge. On the other hand, productivity remains low when there is no more than one surgery per working day. Research in operating room management has not focused on improving the productivity on cardiac surgery. Our main research question, therefore, addresses that gap in the existing research and the second research question addresses the challenge of estimating open-heart operation times:1. How can in open-heart surgery productivity be increased? We suggest some concrete proposals, such as increasing workday length, more overtime and shorter setup times between surgeries, which could lead to productivity improvements. These are tested with a discrete-event simulation open-heart surgery patient process model. Discrete-event simulation enables the evaluation of alternative productivity improvement proposals while maintaining the dynamic nature of the open-heart surgery patient queue. One such proposal, accurate procedure time estimates, was not found in the literature and estimates need to be developed in order make reasonable and achievable schedules in the simulation model2. Can a satisfactory model be constructed for forecasting open-heart operation times? We screen potential factors that could influence open-heart surgery operation times and then use such significant factors to develop a linear regression model for predicting operation times. Input data for developing both the operation time forecasting and simulation models were the actual data of all the 2,613 open-heart surgeries performed at KUH during 2001-2003. Ten cases, where the operation time according database records was null, negative or excessively long (more than 20 hours) were excluded from the analysis making the total number of cases included in the analysis 2603.Forecasting model for operation times\nModel validation: The conceptual model validity was assessed by KUH doctors, who also participated in the project. Their view was that the conceptual model and approach appeared valid for evaluating productivity improvements. Assessing validity through comparison to actual KUH performance is not clear-cut, however, because the model does not aim to exactly replicate existing situations and because historical performance is not available in comparable basis. First, actual variation in the amount of operating tables was excluded from the model. This includes, for example, summer reductions. Second, slack and fillbuffer values applied to KUH cannot be explicitly numerically determined. The KUH slack estimate was 60 minutes. Fillbuffer, i.e. the number of patients called on short notice, was merely said to be uncommon. The patient queue length also varied in actual operation (for example, owing to holidays), while in the model it was fixed. Third, overtime and idle time data were not reliable and direct comparisons were not possible. However, for idle time, a surrogate measure was used in the period between ending and start time of an overtime surgery performed in the same room. Of the 461 data pairs, the average time between two surgeries was 2.25 hours. This can be considered as the idle time, although it does not give an indication of the amount of days without any surgery, and on the other hand, it is likely to overestimate the idle time because of the selection of short surgery and/or empty rooms for overtime surgeries. For output, the surrogate validation metric is the amount of 2nd surgeries in the same operating room per day. In the data, there were 1906 surgeries assessed to have started during regular working hours. Of these 130 (7 per cent) were 2nd surgeries. The simulation model results for idle time and percentage of 2nd surgeries with factor values that either correspond or were judged closest to past conditions (slack: 60 min.; forecast accuracy: no improvement; induction: no; four-day week: no; fillbuffer: 10; queue length: 150) are shown in Table IV.Table IV indicates that confidence can be put to model results and the model appears valid for assessing the means to improve productivity.\nConclusions and recommendations: Surgery volume and overtime work were selected as output variables because they have a direct impact on productivity. The effect of idle time on productivity depends on how idle resources can be allocated for other work. For example, a surgeon may allocate idle time for patient visits. However, in the case study organisation, there was agreement that production improvement methods should not lead to significant use of overtime. As a result, the effects on overtime were minor as compared to the effects on surgery output. If any of the production improvement methods have positive effects on surgery volume then productivity should increase. Reducing changeover time between surgeries by the inducing anaesthesia outside the operating theatre and by reducing the slack time at the end of day after a second surgery have the strongest effects on surgery output and productivity. The area that shows the greatest potential is reducing slack time, because it does not involve any investment. The drawback is the increased overtime. On average, however, overtime is only 0.89 per cent of the total working time. Our research confirms Dexter et al.'s (1999b) claim that improving forecast accuracy does not directly lead to large increases in operating room utilisation. Better forecasting accuracy has an indirect effect by enabling slack time reductions without increasing overtime.Increasing the fillbuffer has a considerable positive effect on surgical output in rooms dedicated to emergency surgeries. If more patients willing to be called for surgery at short notice, as was the situation in the case study hospital, then longer fillbuffers is a cost effective way to increase output. Also, an increasing queue length for elective patients has a positive effect on productivity, but needs balancing with additional costs, decreased quality of care and the potential loss of market share related to long waiting times. A four-day work week increases workday length. It makes it possible to schedule the second elective surgery on most days, which compensates missing Friday surgical sessions and increases the total output without any additional resources.Our results demonstrate that although the length of open-heart surgery, as well as difficulties forecasting operation times, decreases operating theatre productivity, cost-effective methods to considerably improve productivity exist. The challenge in scheduling two open-heart surgeries in one day can be partly resolved by increasing the working day length and decreasing the time between two surgeries or by improving patient scheduling procedures so that two short surgeries can be paired. However, the improvement potential is significantly less than proposed in previous research (Karvonen et al. 2004). Nevertheless, there is much potential for productivity improvement. The average idle time for operating theatres fully dedicated to open-heart surgeries across the experimental design was 33 per cent. One method for decreasing idle time is to have more flexible working hours or arrangements, in which an operating theatre and surgical team can be effectively used for minor surgeries. Changes that lead to increased productivity are difficult to implement without the proper incentives, however. Motivating surgical teams to accept flexible working time and to decrease the average operating time are important factors to consider when designing payment systems. Payment systems should be designed to account for surgery volumes and types of surgeries.A forecasting model based on the linear regression of variables available before the surgery explains 46 per cent of the operating time variation. It confirms previous research findings that there is a correlation between surgeon and average operating time, which unfortunately does not lead to an accurate forecasting model (Zhou et al., 1999; Buchanan and Wilson, 1996; Macario and Dexter, 1999). Owing to our simulation design, improved accuracy decreased overtime, but it did not influence surgical output. Further research on forecasting accuracy, overtime work and how surgery output are related could be fruitful.We have not studied a reduction in the operation time itself. However, the forecasting model can be applied to discover which factors are most significant in explaining variation in the length of open-heart surgery. We propose that further research should be done concerning why the head surgeon's role and piecework pay have such an important effect on surgery length. The forecast model, and the simulation results, created discussion and enabled the evaluation of the impact of concrete improvement proposals without testing them in practice. Based on our results and the experience gained from this research project, the case study organisation is looking for alternative methods for improving its forecast model, e.g. Bayesian and other simulation models could be integrated with existing information systems and be used to plan and schedule open-heart surgery.\n",
"role": "user"
},
{
"content": "Purpose: - The high variability in cardiac surgery length - is one of the main challenges for staff managing productivity. This study aims to evaluate the impact of six interventions on open-heart surgery operating theatre productivity.\nDesign/methodology/approach: - A discrete operating theatre event simulation model with empirical operation time input data from 2,603 patients is used to evaluate the effect that these process interventions have on the surgery output and overtime work. A linear regression model was used to get operation time forecasts for surgery scheduling while it also could be used to explain operation time.\nFindings: - A forecasting model based on the linear regression of variables available before the surgery explains 46 per cent operating time variance. The main factors influencing operation length were type of operation, redoing the operation and the head surgeon. Reduction of changeover time between surgeries by inducing anaesthesia outside an operating theatre and by reducing slack time at the end of day after a second surgery have the strongest effects on surgery output and productivity. A more accurate operation time forecast did not have any effect on output, although improved operation time forecast did decrease overtime work.\nResearch limitations/implications: - A reduction in the operation time itself is not studied in this article. However, the forecasting model can also be applied to discover which factors are most significant in explaining variation in the length of open-heart surgery.\nPractical implications: - The challenge in scheduling two open-heart surgeries in one day can be partly resolved by increasing the length of the day, decreasing the time between two surgeries or by improving patient scheduling procedures so that two short surgeries can be paired.\nOriginality/value: - A linear regression model is created in the paper to increase the accuracy of operation time forecasting and to identify factors that have the most influence on operation time. A simulation model is used to analyse the impact of improved surgical length forecasting and five selected process interventions on productivity in cardiac surgery.\n",
"role": "assistant"
}
] |
Impact of perceived value on the satisfaction of supermarket customers: developing country perspective
|
[
"Customer perceived value",
"Customer satisfaction",
"Supermarket customers",
"Developing countries",
"Fiji"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The creation of superior customer perceived value (CPV) is widely recognised in marketing literature as a vital requirement for organisations to differentiate themselves from competitors (Evans and Wurster, 1997) and to generate customer satisfaction (Demirgunes, 2015). However, \"historically, most companies have relied on the supremacy of their products or services to distinguish themselves from competitors\" (Thielemann et al., 2018, p. 26). Consequently, many companies have been investigating customer-centric approaches to discriminate themselves rather than by their products or services (Rao and Kartono, 2009). More recently, the delivery of value has become the paramount objective of businesses (Reichheld and Teal, 1996).Also, in recent years there has been a renewal of interest by marketing researchers in understanding customers' perceptions of value (El-Adly and Eid, 2015) and the effect of CPV on customer satisfaction (El-Adly and Eid, 2016; Nguyen et al., 2018). However, most research that examines the CPV and customer satisfaction relationship considers CPV as a unidimensional construct, concerned only with the aggregate construct (McDougall and Levesque, 2000) and the measure of overall value (Murray and Howat, 2002). Also, an aggregate measure does not allow for an understanding of the individual dimensions that comprise CPV (Patterson and Spreng, 1997), the complexity of the construct and its interrelationships (El-Adly and Eid, 2015). Therefore, this study aims to broaden the knowledge of CPV by studying CPV as both a unidimensional and multidimensional construct and investigating its influence on customer satisfaction, in the supermarket context.Notwithstanding, specific gaps in the literature exist. First, considering the criticality of the consumer to supermarket sector growth, the consumer is often neglected from such studies. Second, a scarcity of country-wide surveys quantifying supermarket customer's perceptions exists (D'Haese et al., 2008). Third, most prior studies of perceived value concentrate on Western societies (Sarabia-Sanchez et al., 2012). Fourth, a dearth of studies of developing countries' supermarket sectors using the multidimensional PERVAL scale prevails. Fifth, a large number of empirical studies investigate the importance and effect of perceived value on customer satisfaction and produced mixed results (Nguyen et al., 2018). Lastly, despite the value's importance, there has been relatively little empirical research to develop an in-depth understanding of the concept (Sweeney and Soutar, 2001). Due to the multidimensional nature of the value concept, it is crucial to understand the factors that are salient in developing value for customers. This is important for businesses to understand as a customer's perception of value is a precursor to customer satisfaction (Kim et al., 2019; Song et al., 2015; Sweeney and Soutar, 2001).This study attempts to fill those gaps by integrating pertinent theoretical foundations and empirical findings through the use of a modified multidimensional PERVAL scale. We adopt an investigative, cross-sectional research design to provide insight into how developing country (Fiji) customers perceive the value of supermarkets from a construct and multidimensional perspective, the inter-relatedness of the CPV dimensions and the impact of CPV and its dimensions on customer satisfaction. It is anticipated this empirical research will answer the call of Humphrey (2007) for further research of consumers in developing markets and in tailoring supermarket strategies to meet the ever-increasing demands of \"value-driven\" retail customers (Levy, 1999).For this study, the term \"supermarket\" is defined as a large, departmentalised retail store (Pintel and Diamond, 1991), \"which operates on a self-serve basis selling primarily grocery products and, occasionally, a small assortment of nongrocery merchandise\" (Harris et al., 1999, p. 36), and it shall apply to both locally and foreign-owned supermarkets.The next section of this paper provides a critical review of the literature on CPV, customer satisfaction and the impact of CPV from a construct and multidimensional perspective on customer satisfaction. This is followed by the conceptual model, research methodology; results and discussion; and conclusions, future research, and limitations.\nLiterature review: Customer perceived value (CPV) conceptualisation\nResearch methodology: Survey instrument\nResults: Demographic profile of the sample\nDiscussion: Considering the reported Cronbach's alpha values for the modified PERVAL questionnaire administered in this study (a = 0.979), and its constructs dimensions and items, it is suggested the questionnaire has an acceptable/good degree of internal consistency and can be considered a reliable measure of supermarket customer's perception of perceived value and customer satisfaction.The mean scores of this study's constructs reveal that customers tended towards agreeing supermarkets provided CPV and customer satisfaction. These results suggest that supermarket management realise delivering superior value is paramount to generating customer satisfaction (Kesari and Atulkar, 2016), and supermarkets have somewhat appropriate strategies in place to deliver such value offerings. However, the mean scores for CPV dimensions and items are more revealing. These results indicate that customers tended towards agreeing/agree supermarkets provide the four studied dimensions of CPV, and the predominant strategies adopted by supermarkets are based on CPV as a multidimensional construct. These results also indicate a heavy emphasis on supermarket strategies towards providing functional value (price/value for money) to customers, some emphasis on emotional value and social value and a lesser emphasis on functional value (performance/quality). It is implied that these supermarket strategies are guided by context-specific knowledge of consumer behaviour.Additionally, consumer behaviour research indicates, \"growing evidence shows food consumers pay particular attention to price factors\" (Nguyen et al., 2015, p. 2). Further, bearing in mind the Fiji context, the overall poverty rate was reported at 28.1% in 2013-2014 (Fiji Bureau of Statistics, 2015), and \"most [consumers] spent between 26 and 50%...or 51-75% [proportion of their household income]...on food or groceries respectively\" (Kim et al., 2017, p. 17). It is suggested that these low-income shoppers would face spending constraints and trade-off various factors such as price, quantity and quality when purchasing their food products (Leibtag and Kaufman, 2003).Statistically significant positive correlations were shown to exist between each of the CPV dimensions. Results have highlighted that emotional value, social value, functional value (price/value for money) and functional value (performance/quality) are interrelated. This result supports H1 to H6 that was formulated as the start of the study. These results are in line with previous findings that indicate CPV dimensions are interrelated in many retail and service organisations (Sweeney and Soutar, 2001), and reinforces the need for supermarkets to provide products and services as an \"amalgam of rational and emotional factors\" (MacKay (1999, p. 182) when deciding on their marketing strategies. Other correlations tested also showed statistically significant positive correlations (between CPV and customer satisfaction, and between the four dimensions of CPV and customer satisfaction). These results support previous research findings that suggest CPV is positively correlated with customer satisfaction (El-Adly and Eid, 2016; Williams and Soutar, 2009).Model 1 linear regression analysis reveals CPV has a significant positive impact on customer satisfaction, and 43.5% of the variance in customer satisfaction could be accounted for by CPV. Hence, hypothesis H7 is accepted. These results reinforce the findings of researchers in disparate contexts that CPV strongly, positively impacts customer satisfaction (Bajs, 2015; Williams and Soutar, 2009).Model 2 multiple linear regression analysis reveals the social value and functional value (price/value for money) have a significant positive impact on customer satisfaction, emotional value has a significant negative impact on customer satisfaction and in combination accounted for 67.7% of the variance in customer satisfaction. Also, functional value (performance/quality) was found not to have a significant positive impact on customer satisfaction. Hence, hypotheses H9 and H10 are accepted, and H8 and H11 are rejected. Further, the magnitude of the t-statistics indicates the functional value (price/value for money) had a more positive impact than the social value on customer satisfaction. Model 2 results (H9 and H10) align with previous research that identified social value (Williams and Soutar, 2009) and functional value (price/value for money) (Wang et al., 2004) respectively have a direct and positive effect on customer satisfaction. Model 2 results (H8 and H10) are not aligned with previous research that suggests emotional value (Hennig-Thurau et al., 2002) and functional value (performance/quality) have a direct and positive effect on customer satisfaction (Williams and Soutar, 2009). These inconsistent findings reinforce the results of researchers that CPV findings are inconsistent due to the context-specific nature of CPV (Woodall, 2003; Zeithaml, 1988).Theoretical contribution\nConclusions, limitations and future research: The developing country context for this study with one-third of the population living below the poverty line and a large portion of the population as \"working poor\" (with at or below minimum wages and household incomes at or below the poverty level) is suggested to be a significant factor affecting CPV, customer satisfaction and associated supermarket strategies. While some findings of this study are consistent and others inconsistent with the results of other research in disparate contexts (Sweeney and Soutar, 2001; Zeithaml, 1988), such inconsistency is to be expected considering the context-specific nature of CPV (Woodall, 2003; Zeithaml, 1988). It is recommended that comparisons of CPV research findings be cautiously undertaken, bearing in mind the impact of disparate contexts on CPV.Descriptive analysis of CPV and customer satisfaction constructs as aggregates superficially suggests supermarkets adopt strategies that are somewhat effective in delivering CPV and generating customer satisfaction. However, these results also suggest supermarket management and marketers need to further refine and endow with the value their strategies and offerings in order to enhance CPV and customer satisfaction. Additionally, descriptive analysis of CPV as a multidimensional construct suggests supermarket strategies are guided by context-specific knowledge of consumer behaviour. Hence, there is a perceived heavy emphasis on supermarket strategies towards providing functional value (price/value for money) to customers, some emphasis on emotional and social value and a lesser emphasis on functional value (performance/quality). The compound effect of socio-economic factors is suggested to be an overriding factor influencing supermarket strategies, consumer behaviour and ultimately CPV and customer satisfaction. Supermarket management and marketers are also recommended to note the inter-related nature of the CPV dimensions and to ensure marketing strategies promote products and services as an \"amalgam of rational and emotional factors\" (MacKay, 1999, p. 182).Inferential analysis of the aggregate constructs reveals CPV has a significant positive impact on customer satisfaction and accounts for 43.5% of the variance in customer satisfaction. This finding is significant as it enables supermarket management and marketers to hone marketing strategies to optimise CPV, and predict customer satisfaction outcomes.Inferential analysis of the multidimensional construct CPV and unidimensional construct customer satisfaction reveals the social value and functional value (price/value for money) have a significant positive impact on customer satisfaction, emotional value has a significant negative impact on customer satisfaction, functional value (performance/quality) has no significant effect on customer satisfaction and in combination accounted for 67.7% of the variance in customer satisfaction. It is suggested that the significance of the impact of functional value (price/value for money) and insignificance of the impact of functional value (performance/quality) on customer satisfaction findings are interrelated and result from the compound effect of socio-economic factors on this study's customers, and supermarket knowledge of, and strategies that, take into account the impact of socio-economic factors on customers. As indicated by Leibtag and Kaufman (2003), it appears customers resort to a trade-off between various factors such as price, quantity and quality when purchasing their food products. While the non-significant impact of functional value (performance/quality) and significant negative impact of emotional value on customer satisfaction may appear to be inconsistent with other disparate research findings, it is suggested these are significant findings of this study and readily explained. Consider the interrelatedness of the four value dimensions (Sweeney and Soutar, 2001), the viewpoint that cognition precedes emotions (Bigne et al., 2008), and customer emotions are also an essential element of customer satisfaction (Cronin et al., 2000). Now, it is reasonable to assert when customers undergo cognitive appraisals (price/value for money and performance/quality) of the consumption process, this leads to judgments of emotional value and, in turn, affects customer satisfaction. Hence, in this study with functional value (performance/quality) suggested to be traded off by customers and also deprioritised by supermarket management and marketers against the functional value (price/value for money), a resultant outcome is negative customer emotions and an adverse impact on customer satisfaction. Additionally, while the social value is considered in disparate contexts to be the critical CPV dimension that influences customer satisfaction (Williams and Soutar, 2009), in this study, social value has been overshadowed by functional value (price/value for money). Notwithstanding, in this study, shopping at supermarkets is perceived to be an essential factor that affects customers' decision-making, social image and social status and for customers to improve how they are perceived by others.This study broadens the knowledge of CPV by studying customer perceptions of CPV as both a unidimensional (aggregate) and multidimensional construct and investigates its influence on customer satisfaction in the understudied supermarket context of a developing country supermarket. The descriptive and inferential findings of this study offer opportunities for supermarket management and marketers to hone strategies to optimise their CPV efforts and, in turn, enhance customer satisfaction. However, we concur with the recommendation of researchers (Meirovich et al., 2013, p. 3) that \"the complex relationship between cognitive appraisals, affective responses and satisfaction warrants further investigation.\"Certain limitations arise from this study. Firstly, the unit nonresponse bias for this study was 38%, which could affect the representativeness of the sample. To mitigate unit nonresponse bias, future research could consider the implementation of guidelines for initial contact with potential respondents, reduction in questionnaire length and use of an alternative sampling procedure. Secondly, this quantitative study focused on quantifying customer perceptions of CPV and its influence on customer satisfaction. Future research could utilise a mixed-methods approach to shed light on the human experience associated with this phenomenon and to gain a deeper understanding. Thirdly, this study measured CPV in a post-purchase situation. Future research, as suggested by Woodruff (1997), could investigate whether CPV and its dimensions change across different scenarios such as pre- and post-purchase.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to provide insight into the impact of customer perceived value and its dimensions on customer satisfaction in a developing country, and practical suggestions for marketing strategies.\nDesign/methodology/approach: A cross-sectional survey collected data from five-hundred supermarket customers in Fiji. SPSS was used to provide descriptive and inferential analysis.\nFindings: Results reveal that customer perceived value (CPV) has a positive impact on customer satisfaction; and functional value (price/value for money) has more positive impact than social value, emotional value has a negative impact and functional value (performance/quality) has no significant impact on customer satisfaction.\nResearch limitations/implications: Considering this research was undertaken in the supermarket sector of only one country, other researchers are urged to replicate this research in Fiji and other developing countries, to yield further insight into the context-specific nature of CPV.\nPractical implications: It is suggested that marketers note these findings (to understand better the conceptualisation and context-dependent nature of CPV, its dimensional interrelationships and its impact on customer satisfaction) in order to enhance CPV and ultimately customer satisfaction.\nOriginality/value: This study makes several contributions to research on CPV by providing insight into how developing country customers perceive the value of supermarkets from a construct and multidimensional perspective, the inter-relatedness of CPV dimensions and the impact of CPV and its dimensions on customer satisfaction.\n",
"role": "assistant"
}
] |
The gender wage gap among Spanish managers
|
[
"Pay",
"Women executives",
"Glass ceilings",
"Sexual discrimination",
"Counterfactual decomposition analysis",
"White collar workers"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: It is largely accepted in current economic policy, particularly by the European Commission (EC), that equality between men and women is a global priority as a fundamental right and a shared principle of the European Union. It is also well known that equality between the two sexes is not just a goal in itself: it is a precondition for fulfilling the overall EU objectives of growth, employment, and social cohesion.\n2. Empirical specification: The classic ordinary least squares method, which is based on the mean of the conditional distribution of the dependent variable, implicitly assumes that the impact of the regressors along the conditional distribution is irrelevant. Thus, it really misleads significant information about the wage distribution of managers in Spain. Therefore, we apply the QR method (Koenker and Bassett, 1978), which is more interesting and more appropriate in this context: the thth quantile of a variable conditional on some covariates can be accounted for and the effect of those covariates at different quantiles of the distribution can be estimated.\n3. Data: The data used in this paper are from the 2007 version - available since March 2009 - of the EU-SILC database, the new homogenized panel survey that has recently replaced the ECHP. EU-SILC has both a personal (p) and a household (h) file, from where we needed to take data. This necessity obliged us to merge these two sources, inserting the code of the household type in the former database for Spanish individuals. Students, people serving in the Armed Forces and the self-employed were also excluded. The analysis is further restricted to Spanish individuals between 25 and 65 years of age who are in managerial positions. EU-SILC identifies such a working condition with a supervisory role where supervisory responsibility includes formal responsibility for coordinating a group of other employees (other than apprentices). It implies that the manager takes charge of the work and sees whether it is properly done. Such a definition closely follows the one adopted in the theoretical models by Acemoglu and Newman (2002) and Beaudry and Francois (2010). After this process, 2,377 individuals remained: 775 females and 1,602 males.\n4. Results: Ordinary least squares and quantile regression\n5. Discussion and conclusions: Using the wave of the EU-SILC survey released in 2009, this paper has employed both OLS and quantile regressions and a semi-parametric decomposition method to examine the GWG at the mean and over the entire conditional wage distribution of the Spanish managerial workforce.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the existence of sticky floor and glass ceiling effects in the gender wage gap (GWG) among Spanish managers. In addition, the paper determines if the pay gap at every quantile is a result of the gender characteristic differences, or the differences in returns to those characteristics.\nDesign/methodology/approach: The paper exploits a counterfactual decomposition analysis, using quantile regression, to decompose the GWG into one component that is based on differences in characteristics and one component that is based on differences in coefficients across the wage distribution.\nFindings: A significant GWG over all the wage distribution is found. Such a gap exhibits a clear U-shaped pattern, thus pointing out both significant sticky floor and glass ceiling effects. Furthermore, the paper shows that such pattern is mainly determined by the coefficient effect, whose relative incidence is almost continuously increasing along the wage distribution.\nResearch limitations/implications: While it is difficult to give a definitive explanation for the significant U-shaped pattern in the GWG and for the bigger incidence of the glass ceiling, the authors suggest two possible explanations that are consistent with these findings. The paper leaves the identification of these explanations to future research.\nPractical implications: The pattern of rising coefficient effects at higher quantiles suggests that the glass ceiling is a more relevant question than the sticky floor. Indeed, at the highest wage quantiles, differences in characteristics make essentially no contribution to the overall wage gap. This suggests that upper-echelon female managers have the same characteristics as their male counterparts, which emphasizes the role of discrimination for these top-level jobs.\nOriginality/value: Despite the general GWG has been largely investigated, the analysis of a wage differential among managerial workers has certainly drawn much less attention. In particular just a few papers have investigated the existence of sticky floors and glass ceiling among managers. In addition, as to Spain, there is no empirical survey investigating and decomposing the gender pay gap among managers.\n",
"role": "assistant"
}
] |
Identifying international market opportunities: the case of Italian companies in Egypt
|
[
"Strategy",
"Benchmark"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Selection of non-domestic markets is one of the most critical decisions in international strategy (Douglas et al., 1972; Goodnow and Hansz, 1972; Kobrin, 1976; Paliwoda, 1993; Root, 1994; Sarkar and Cavusgil, 1996; Simpson and Kujawa, 1974; Wind et al., 1973; Wind and Perlmutter, 1977; Andersen and Strandskov, 1998; Carbonara and Caiazza, 2009a, b, 2010). Few market selection models have been proposed in the literature (Papadopoulos and Denis, 1988). These models differ mainly for both the number of sequential stages and terminology. Cavusgil (1985) terms these stages preliminary screening of markets in attractive countries, assessment of industry market potential and company sales potential analysis, Kumar et al. (1994) refer to screening, identification and selection, Root (1994) refers to preliminary screening, in-depth screening and final selection.\nInstitutional setting: The European Neighbourhood Policy (ENP) has transformed relations between the European Community and its neighbors. The ENP aims to open political dialogue and cooperation with each of European neighbours and to achieve a substantial degree of economic integration (European Commission, 2010a). At this aim, there has been an on-going strengthening of contractual relations. Through the Association Agreements (AAs), European Community has gradually built a deeper relationship with its neighbours. These agreements permit regular exchanges on all issues of common interest, ranging from governance and democracy to transport and health sector reform, and contributing to increased confidence and reciprocal knowledge. The success of this approach has also prompted several southern partners to seek further enhancement of bilateral relations. Specifically, European Community and Egypt first established contractual relations in 1976 by signing a Cooperation Agreement. An European Community-Egypt AA entered into force in June 2004. It identified the specific areas that could be developed bilaterally. On this basis, the European Community-Egypt ENP Action Plan was approved in March 2007 (European Commission, 2010b).\nMethodology and sample: We follow a process perspective in analyzing risks and opportunities for Italian firms that aim to operate in Egypt. At this aim we realized some face-to-face interviews based on semi-structured survey at senior executives of 70 Italian firms operating in Egypt. We interviewed managers of Italians firms that operate in Egypt through industrial productive centres (61 percent) and representative offices (22 percent). Firms' main sectors of activities are agricultural (17 percent), food industries (17 percent), chemical industries (11 percent), electromechanical (11 percent), trading company (6 percent), marine industries (6 percent), automotive (6 percent), transportation (5 percent), banking (5 percent), tourism (5 percent), representative office (5 percent), general contracting (6 percent). Firms are mainly SME with number of employees between 11-50 (45 percent) and 51-250 (33 percent) and local revenues less than 2 million of Euro (44 percent).\nResults: The goal of this part of research is to collect information on managers' views of risks and opportunities for Italian firms that operate in Egypt. Most of respondents do not operate in other South Mediterranean Countries (67 percent) and considers Egypt as main destination of their products (56 percent). Most of them (56 percent) cooperate with local firms in joint projects for generating more products, for a better use of available resources, for using facilities, for sharing information and discussing market problems, for launching common programs, for sharing resources and facilities, and for a common use of available facilities and resources. They have not find difficulties in starting or managing these relationships (83 percent) because Egyptian partners cooperate (94 percent) and share managerial practices with Italian ones (83 percent). Most of respondents does not perceive any difference in the way to cooperate between Italian and Egyptian actors (89 percent). According to them Egyptian partners stat to cooperate with Italian firms mainly for gaining know-how, reputation of a well-known company, experience, opening opportunities for collaboration with other big multinational companies, opening market for products, because of quality of the products and good understanding.\nOpportunities and risks of Egypt: Respondents accepted to give a rank from 1 (less important) to 7 (most important) for evidencing the reasons for which firms could choose to enter in Egyptian market: cost saving (50 percent gave a rank of 6, 11 percent gave a rank of 5, and 6 percent gave a rank of 7) and access to supplying markets (28 percent gave a rank of 6, 22 percent gave a rank of 7 and 17 percent gave a rank of 5) are very important for 67 percent of respondents. 89 percent of respondents (28 percent gave a rank of 5, 33 percent gave a rank of 6 and 28 percent gave a rank of 7) considered bigger sale market and 61 percent fiscal advantages (22 percent gave a rank of 5, 22 percent gave a rank of 6 and 17 percent gave a rank of 7) important reason for which firms could choose to enter in Egypt. For most of respondents (56 percent=28 percent+17 percent+11 percent) facing local competitors is an important reason to enter in Egypt, while access to convenient financial resources (72 percent=11 percent+39 percent+22 percent) is a less important reason. Despite of some opportunities, according to the interviewed executives firms that aim to operate in Egypt have to face a lot of obstacles.\nAlchemi experience: We analyze Alchemi case-study to offer a firm-specific perspective of opportunities and risks in selecting Egypt as final market.\nConclusions: Internationalization of Italian firms in Egypt is a topic of high interest for researchers and practitioners, however studies focused on this topic are still missing. This study addressed the need for a stage process in analyzing internationalization of Italian firms in Egypt. It is proposed that the application of the additional criteria to the traditional one will compensate for the approach taken by the normative market selection methods by providing a matter-of-fact exposure of Egyptian potential and will complement them in preliminary assessment before the immediate measures of macroeconomic variables, country risk and profit conversion potential. The proposed approach to assessing Egypt contributes to and extends traditional methods by incorporating micro analysis at the industry and customer levels into country level macro analysis.\nDiscussion: From a cultural perspective, our analysis evidenced that Italian operate as multicultural organizations based on mutual exploration of diversities, including a mutual rapprochement and active taking advantage of differences while Egyptian operate as pluralistic organizations. As multicultural organizations Italians are used to assume three different approaches: cultural integration, cultural synergy and cultural dialog. In cultural integration, differences are consciously acknowledged and accepted. In cultural synergy diversities are accepted and an attempt is made to approach each other and develop a new culture. The cultural dialog represents a further step toward mutual exchange and dialogue. In this approach, people not only accept that the others are different, but in fact invest time in getting to know each other with their differences and idiosyncrasies. As pluralistic organizations, instead, Egyptian are used to assume cultural separation behavior. Cultural separation between groups lead to accept differences but despite the tolerance there is only little interaction between the different groups.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the differences and relations between Italy and Egypt and to identify main risks and opportunities for Italian firms that aim to operate in Egypt.\nDesign/methodology/approach: The paper presents a qualitative analysis realized through some semi-structured questionnaire and a case study.\nFindings: High cultural difference, legal uncertainty and administrative instability are main problems in internationalization process of Italian firms in Egypt. Despite these problems, Italian firms still invest in Egypt because of economic opportunities and possible interactions.\nOriginality/value: While interest towards Southern Mediterranean countries is increasing, there is little written on this topic. This article aims to fill the gap in literature on internationalization in emerging markets.\n",
"role": "assistant"
}
] |
Integrating research, clinical care, and education in academic health science centers: An organizational model of collaborative workplace learning
|
[
"Collaboration",
"Workplace learning",
"Clinical effectiveness",
"Academic health centers",
"Academic medical centers",
"Evidence-informed decision making"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Academic health science centers (AHSCs) are organizations of the future, embracing a vision of compassionate and state-of-the-art care, informed by the latest research and delivered by well-functioning, interdisciplinary teams of highly competent service providers (Dzau et al., 2010). AHSCs, also known as an academic health centers, have a tripartite academic mission of excellence in research, clinical care, and education, which differentiates them from teaching or university hospitals. AHSCs have various organizational forms, including stand-alone institutions and partnerships between universities and affiliated teaching hospitals or health systems (Ovseiko et al., 2014; Shrader et al., 2015). Thus, the organizational structures of AHSCs vary in the degree to which the academic functions operate under a single administrative and governance structure (Barrett, 2008). In health care, there can be academic silos (i.e. cultural, social, and cognitive boundaries) that lead to isolation and disconnection between research, clinical care, and education (Ferlie et al., 2005). Even in AHSCs with fully integrated governance, academic silos remain and there are challenges in effectively integrating research, clinical care, and education (Ovseiko et al., 2010). In this paper, we focus on a free-standing AHSC with a single management board. Our context is a pediatric rehabilitation hospital, formally designated as an AHSC in the Toronto Association of Academic Health Science Network.\n1. Background literature on AHSCs: To ascertain what is known about workplace learning in the AHSC literature, we retrieved articles in PubMed using \"academic health\" AND \"science(s) center/centre(s)\" or \"center/centre(s),\" OR \"AHSC,\" with \"learn\" or \"team\" as keywords. There were no restrictions on year. The search revealed a growing number of articles on AHSCs, primarily from the USA and UK, since 2002. The literature largely consisted of case study descriptions of interprofessional education (IPE) and faculty development in professional school campuses that are part of AHSCs (e.g. Aston et al., 2012; Shrader et al., 2015; Willgerodt et al., 2015) and empirical articles on interprofessional teams (e.g. Hall et al., 2009; Magrane et al., 2010).\n2. Current problem concerning the structure and functions of AHSCs: 2.1 Lack of integration of the academic functions of AHSCs\n3. A model for integrating cross-functional activities in AHSCs: The literature has called for the development of operational roadmaps to meet the vision of AHSCs (Institute of Medicine, 2007; Grumbach et al., 2014). Thus, we propose an organizational model outlining how the academic functions of AHSCs can be harnessed to enhance client outcomes through cross-functional, mission-integrating activities designed to enhance workplace learning and thus EIDM. The model conceptualizes the major functions, sub-functions, and collaborative learning processes of AHSCs, thereby providing a vision of how to enhance the quality of health services. It is intended to support the development of organizational structures to address the academic functions and support cross-functional activities, which are defined as collaborative activities involving people representing two or more of the academic functions (Table I). These activities involve clinicians, researchers, and educators (and in many cases service recipients).\n4. Literature used to address the problem: The literature on organizational learning in its broadest sense was used to inform the development of our model. Our interest was in identifying structures and processes related to the AHSC vision of being a \"learning health system/organization\" (Grumbach et al., 2014). If this is the vision, what does the literature say about how to get there?\n5. An evidence-informed model of collaborative workplace learning: This model proposes relationships among AHSC academic functions and sub-functions, cross-functional activities, and collaborative learning processes. The model conceptually integrates sub-functions that are typically siloed within an organization's functioning and knowledge base, and highlights the importance of cross-functional activities, which are considered to create synergies leading to expanded knowledge and innovation. Various collaborative learning processes are hypothesized to be involved in these cross-functional activities, leading to enhanced clinician competency, EIDM, and optimal client care.\n6. Conceptual and empirical support for relationships among model elements: 6.1 Academic sub-functions and collaboration\n7. Model contributions and limitations: The model provides a novel contribution by integrating personal learning, learning in teams, and organizational structures and functions. It provides a unified understanding of learning arising from synergistic collaboration, indicating that innovative knowledge may be fostered by setting conditions for cross-functional activities to flourish. It provides a differentiated view of organizational knowledge, going beyond \"collaboration\" and \"cooperation\" to propose collaborative learning processes. The model suggests the importance of supporting both the formal development and spontaneous emergence of cross-functional teams, where knowledge is generated, exchanged, and applied across traditional research, clinical, and educational boundaries. This vision of collaborative learning corresponds to that of \"learning rich\" organizations, where learning is characterized as experiential, team oriented, ongoing, capacity building, integrated, and situational (Garavan et al., 2002). The contribution of the proposed model is in highlighting the potential importance of cross-functional activities and collaborative learning processes. At present, it is not specifically known whether collaborative learning processes actually do play a mediating, integrating role (although collaboration is widely regarded as important), or whether cross-functional activities have greater impact than sub-functions operating alone.\n8. Implications for AHSCs: The model has applicability to AHSCs and other health organizations striving to be more academically oriented. It can be used to enhance organizational culture by providing clarity regarding the nature of desired outcomes and organizational vision. The model can enable people in siloed operational structures to see their role in the aspirational vision as a whole, as it indicates how each sub-function contributes to common goals of EIDM and optimal client care. The model also has implications for the promotion of cross-functional activities through intentional design, as it describes the nature of activities and processes that can be embedded in day-to-day activities to potentially enhance learning, decision making, and clinical practice. AHSCs can provide opportunities, supports, and resources to encourage active boundary spanning, dialogue, and joint learning activities (Ferlie et al., 2000); establish appropriate coordination and communication mechanisms (San Martin-Rodriguez et al., 2005); provide incentives (reward systems) to nurture partnerships (Lozon and Fox, 2002); and establish boundary spanning roles (Ash, 1997). To chart progress toward the vision of collaborative workplace learning, organizations can use indicators of a fully integrated work environment, such as establishment of positions with dedicated time for other functions, availability of lateral organizational moves for individuals across functions, and evidence of increased EIDM.\n9. Conclusion: This model of collaborative workplace learning has utility for AHSCs and other health care organizations interested in aligning their academic functions into a common vision. The model recognizes the importance of cross-functional activities involving five key academic sub-functions commonly used to operationalize an AHSC's mandate to educate, conduct research, and provide excellent clinical services. The sub-functions target different clinician competencies, but their common aim is to develop clinicians' ability to make evidence-informed decisions in providing clinical care. The model indicates the importance of collaborative learning processes in enhancing workplace learning, namely co-constructing meaning, co-learning, and co-producing and co-using knowledge. When operating as intended (good communication, high cohesion, clear and common goals), cross-functional activities can potentially lead to transformative workplace learning, assisting AHSCs to meet their expected outcomes and academic vision. Collaboration is considered to be a key integrative mechanism in the creation of a learning-oriented workplace environment - one that is able to overcome functional silos in the workplace.\n",
"role": "user"
},
{
"content": "Purpose: One of the major issues faced by academic health science centers (AHSCs) is the need for mechanisms to foster the integration of research, clinical, and educational activities to achieve the vision of evidence-informed decision making (EIDM) and optimal client care. The paper aims to discuss this issue.\nDesign/methodology/approach: This paper synthesizes literature on organizational learning and collaboration, evidence-informed organizational decision making, and learning-based organizations to derive insights concerning the nature of effective workplace learning in AHSCs.\nFindings: An evidence-informed model of collaborative workplace learning is proposed to aid the alignment of research, clinical, and educational functions in AHSCs. The model articulates relationships among AHSC academic functions and sub-functions, cross-functional activities, and collaborative learning processes, emphasizing the importance of cross-functional activities in enhancing collaborative learning processes and optimizing EIDM and client care. Cross-functional activities involving clinicians, researchers, and educators are hypothesized to be a primary vehicle for integration, supported by a learning-oriented workplace culture. These activities are distinct from interprofessional teams, which are clinical in nature. Four collaborative learning processes are specified that are enhanced in cross-functional activities or teamwork: co-constructing meaning, co-learning, co-producing knowledge, and co-using knowledge.\nPractical implications: The model provides an aspirational vision and insight into the importance of cross-functional activities in enhancing workplace learning. The paper discusses the conceptual and empirical basis to the model, its contributions and limitations, and implications for AHSCs.\nOriginality/value: The model's potential utility for health care is discussed, with implications for organizational culture and the promotion of cross-functional activities.\n",
"role": "assistant"
}
] |
Office design and organizational change: The influence of communication and organizational culture
|
[
"Organizational change",
"Communication",
"Organizational culture",
"Office space",
"Open-plan"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: There is considerable evidence from academic and practitioner sources that open-plan office spaces are disliked (e.g. Kim and de Dear, 2013; Konnikova, 2014; Morrison and Macky, 2017), yet organizations continue to pursue this form of office design. In contrast to most reports, our investigation of a move by a law firm to a new building and into an open-plan design revealed that most employees were highly pleased with the aesthetic nature of the office space, the way in which the change was managed and the improved communication and organizational culture that resulted.\nLiterature review: Studies on office space reveal fragmented contributions (Taylor and Spicer, 2007), some focusing on ambient conditions and others on psycho-social aspects. We begin the literature review with an analysis of office space and its relationship to organizational culture and communication. We then examine research on changes to office space and how the processes and outcomes of change affect, and are affected by, culture and communication.\nMethod: Theoretical foundation\nFindings: The focus of this paper is to explore, from a variety of internal stakeholder perspectives, the intersecting influences of the processes and outcomes of organizational change (the office move and open-plan design), communication and organizational culture. In addition, regarding reactions to the ambient conditions (aesthetic appeal, privacy and distraction), as well as desk size and storage space, most employees, even partners and senior managers, liked the attractive, modern design, the use of plants and artwork, the central kitchen/dining area and the sunlit atrium. Smaller desk space and reduced storage space were initially sources of irritation to some, but, over time, most participants claimed that they had adjusted to them.\nDiscussion: The purpose of this study was to evaluate how the processes and outcomes of an organizational change were evaluated by a range of internal stakeholders. Given the predominantly negative findings in the literature on working in open-plan spaces (e.g. Kim and de Dear, 2013; Morrison and Macky, 2017), we were expecting poor perceptions of employees moving to a new spatial design. This would have been consistent with views that open-plan offices are frustrating in terms of distraction and lack of privacy (e.g. Bergstrom et al., 2015; Morrison and Macky, 2017), impinge on personal and social identity (Elsbach, 2003) and are overtly or covertly resisted by employees as sites of surveillance and control (Baldry and Barnes, 2012). At best, we were expecting muted or ambivalent responses in light of the equivocal research findings on this issue (Ashkanasy et al., 2014; Hongisto et al., 2016; Taylor and Spicer, 2007).\nLimitations and conclusions: Our study is limited to a small number of participants in one firm and reports outcomes that run counter to those findings that open-plan offices are disliked and undermine co-worker relationships (e.g. Morrison and Macky, 2017). Given the choice of interview locations that were offered (28 onsite meeting rooms and offsite options), we doubt that staff were concerned if others could see that they were being interviewed or if this possible visibility led to more positive reflections on the change. Researchers are usually expected to be distanced and dispassionate observers and interpreters of the subjects (animate and inanimate) that they study. Having been given a tour of the premises by the COO, and having conducted all our interviews onsite, distance was removed in the literal, spatial sense. Despite our admitted preference for our single university offices, adjacent to an open-plan area for part-time staff, our own subjective perceptions and emotions regarding the law firm space were more in tune with the concept of \"space as the manifestation of our imagination\" (Taylor and Spicer, 2007, p. 327), a site of enchantment - \"awe and wonder\" (Burrell and Dale, 2014, p. 692) in the generative building (Kornberger and Clegg, 2004). Professional design, facilitated by an inclusive change process, had created an aesthetically-pleasing and engaging space.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to compare different employee perceptions of the success of one change: a move to new offices and an open-plan design.\nDesign/methodology/approach: In sum, 25 interviews were carried out in a New Zealand law firm that six months earlier had moved to new premises.\nFindings: Contrary to academic and practitioner reports that open-plan offices are disliked, participants appreciated the new office space. A well-planned and highly participative program of change management led to positive perceptions of aesthetic design, open communication, collegiality, egalitarianism and inclusiveness.\nResearch limitations/implications: Given the small sample used in one organization, the study highlights the need for more research into the processes and outcomes of office space changes.\nOriginality/value: The roles of communication and culture, in particular, collegiality and egalitarianism, were salient factors in a complex web of causes and consequences in this context of change.\n",
"role": "assistant"
}
] |
Towards a model of the right-hand person
|
[
"Diversity",
"Agency theory",
"Top management team",
"Congruence",
"Co-leadership",
"Right-hand person"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Powell noted that it was Cook who simplified Apple's product line, cut its inventory and got Apple into the build-to-order business in the 90s, which is \"not glamour stuff\" but turned Apple's billion-dollar loss into a $US600 million profit in a year.\nIntroduction: Imagine what would have happened to Apple if Tim Cook had not been Steve Jobs's right-hand person. Would Apple have become so successful? If Jobs had had to take care of the operations, would he still have had sufficient energy to concentrate on his creative work, innovative ideas, and the design of Apple's products? If Charlie Munger was not able to provide Warren Buffet with critical and valuable comments when being consulted, would Berkshire be as successful as it is today? Why does Melinda Gates play an important role in running the largest private foundation in the world together with her husband, Bill Gates? These are questions that almost all top executives and their right-hand persons want the answers to. Unfortunately, it appears that we know very little about the factors that contribute to the effective relationship between a top executive and his or her right-hand person.\nTheory and the importance of the right-hand person: There are at least five reasons why the relationship between the top executive and his right-hand person is so important for business organizations. First, the phenomenon of the right-hand person is common in business organizations. It is almost impossible to imagine a top executive without one or more chief assistants. Second, as his chief assistant, the right-hand person plays a very important role in the work of the top executive, either by implementing the top executive's decisions or by directly participating in the decision-making process, which in turn affects organizational performance. Third, the relationship between the top executive and the right-hand person has an important impact on the working climate of the top management team (TMT; Bantel and Jackson, 1989; Pitcher and Smith, 2001; Priem et al., 1999) or even the whole organization. Fourth, as the right-hand person acts to some extent as a representative of the top executive, s/he has plenty of opportunities to interact with the middle-level managers and front-line employees on behalf of the top executive. His or her working style has the potential to impact organizational norms regarding how things should be done and whether politics are needed and desired in the organization. Finally, although the successor in a family business was traditionally determined by blood relationship (Lee et al., 2003), this is no longer true for many business organizations, especially in listed companies. Thus, the right-hand person may be an important element of a good succession plan.\nProposed model of a successful right-hand person: Based on the need for high levels of trust and effective collaboration between the chief executive and the right-hand person, we argue that the congruence between the top executive and the right-hand person is the key to successful cooperation between them. Based on studies of co-leadership and drawing on agency theory and the TMT literature, we propose that different roles require different types of congruence. Table I illustrates the hypothesized congruence that will lead to the success of the right-hand person. For a successful implementer, the interests of the right-hand person must be taken care of and s/he must be empowered to the extent that appropriate authority is delegated from the top executive so that s/he can carry out his responsibilities. For a successful joint decision maker, the top executive and his/her right-hand person need to share similar values and commitment to the mission and vision of the company. Their abilities also need to be complementary. We explain the rationale of the model in greater detail in the following paragraphs as illustrated by the cells of Table I.\nMethods: Research design\nDiscussion: To the best of our knowledge, this is the first study to examine the phenomenon of the right-hand person in organizations. Inspired by the Chinese concept of \"left- and right-hand person,\" and drawing from agency theory and TMT research, we propose that a right-hand person may play two roles: implementer and joint decision maker. We further suggest that a right-hand person is likely to be a good implementer if his interests are met. In addition, although compensation works to motivate implementers, if the right-hand person has spent enough time working with the executive and has developed trust and mutual understanding, the influence of short-term compensation may be mitigated. To become a successful joint decision maker, the right-hand person must share similar values and beliefs with and possess complementary capabilities to the top executive.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to propose a model explaining the roles of right-hand person and the factors contributing to the successful relationship between the top executive and the right-hand person.\nDesign/methodology/approach: In-depth qualitative case studies are conducted. Longitudinal observations, interviews with six right-hand persons and the top executives in three organizations are conducted to test the propositions of the model.\nFindings: Results indicate that different types of congruence between the top executive and the right-hand person are required when the right-hand person is performing the roles of an implementer and joint decision maker.\nResearch limitations/implications: This study extends the leadership literature by investigating the phenomenon of right-hand person of the top executive, which has seldom been studied systematically or scientifically. It provides insights and serves as a stepping stone for future research in this area. One key limitation is that it is a qualitative study with limited samples under investigation.\nPractical implications: Practical implications concerning how to build up a successful relationship between the top executive and the right-hand person can be drawn from the proposed model. Insight concerning how to collaborate between the top executive and the right-hand person can be drawn from the in-depth case analyses.\nSocial implications: The phenomenon of right-hand person is not limited to business organizations. The collaboration between the key decision maker and his chief assistant should be applicable to other contexts such as non-government organizations.\nOriginality/value: To the best of the knowledge, this is the first paper that investigates the right-hand person phenomenon in the literature. As the right-hand person of the top executive can have important influence on organizational performance, the study may serve as the stepping stone for further understanding of this important phenomenon.\n",
"role": "assistant"
}
] |
Analysis of new product development research: 1998-2009
|
[
"Innovation",
"Product design"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: The product development process objective is to transform the market opportunity and a set of assumptions about product technology into a product available for sale (Krishnan and Ulrich, 2001). New product development (NPD) is recognized as one of the most critical areas of firm's competence related to business success (Guo, 2008). Product Development and Management Association (PDMA) best practices surveys (Page, 1993; Barczak et al., 2009) concluded that, although firms had implemented a number of new methods and techniques to improve the way new products were developed, new product success rates continue to be remain stable at around 60 percent. The possibility of improving the new product success rate is always a key motivation in conducting research in NPD by developing the new practices. The total number of NPD review articles available in literature is very less. Page and Schirr (2008) have emphasized the need and importance of a systematic multi journal literature review on NPD research.\n2 Literature reviews related to NPD: Table I summarizes the different reviews in the field of NPD in chronological order.\n3 Methodology and analysis: 3.1 Journal and article selection\n4 Discussion: This paper, a systematic review of research literature was conducted on nine different principal NPD information groups like published journal, published year, NPD research stream, type of organization studied (industrial/consumer/service), level of innovation (high/moderate/low) NPD focus on frameworks, Performance perspective (success, failure or both), NPD research design (conceptual/empirical and qualitative/quantitative) and NPD relevant best practice element. The review enables to succinctly present status of NPD research in existing literature and chart an agenda for future research. The significant findings, gaps that were identified and the future directions will be discussed in the subsequent part of this section.\n5 Conclusions: The paper reviewed 1,127 articles on NPD with multiple objectives like growth of NPD research; growth in identified research streams; exploring the NPD research focus towards various issues like of innovation, framework development and performance measurement in NPD research; list and classify the NPD best practices elements and to show the changes and trends over the time period as and when required. In the study period, only ten journals have published more than 20 articles on NPD which shows that there is a certain level of dispersion in the publication of the NPD. The paper also highlights the growth NPD research. This study on NPD research also identifies the sector wise growth in NPD research vis-a-vis industrial, consumer and service sectors. It is observed that many researchers have failed to judge the level of innovativeness in NPD research and in the classified literature more focus was observed towards moderate innovation. An attempt has been made to explore the focus towards framework development in terms of number of articles published and inadequately addressed issues were highlighted. NPD performance issues were dominated by focus towards success factors. Review also revealed important and frequently visited issues in NPD along with highlighting all the inadequately addressed issues into limelight in the form of significant findings, research gaps and future directions.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present a systematic review and analysis of existing research articles on new product development (NPD) published in the 12-year period starting from 1998 to 2009.\nDesign/methodology/approach: To explore the articles related to NPD, four key words namely new product, product design, product development and product innovation were used in combination of title or abstract or keyword of the articles through several knowledge databases. The abstracts of journal papers were read and a decision as to whether article belongs to any NPD research issue or not was made. In total, 1,127 research articles were categorized systematically and then analyzed on various principal NPD information groups.\nFindings: Analysis of selected articles led to a certain level of dispersion in the publication of NPD research in different journals. It is found that more attention needs to be on knowledge and creativity management, communication and information transfer in any NPD process.\nOriginality/value: By observing extended literature from authors reviewing articles from various journals, growth in research, and variety of topics covered in NPD, a broad systematic multi journal review of NPD literature is clearly overdue. The authors have developed a comprehensive listing of publications on NPD where they have classified the surveyed papers according to various principal NPD information groups like: published year, NPD research stream, type of organization studied (industrial/consumer/service), level of innovation (high/moderate/low), NPD focus on frameworks, performance perspective (success, failure or both), NPD research design (conceptual/empirical and qualitative/quantitative) and NPD relevant best practice element. Based on the classification scheme, the issues were analyzed from the system's perspective and their implications to NPD research.\n",
"role": "assistant"
}
] |
Tax audit and tax productivity in Lagos state, Nigeria
|
[
"Logit regression",
"Back duty",
"Co-integration analysis",
"Desk tax audit",
"Field tax audit",
"Tax productivity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Taxation is not a new word in the world as a whole. In Nigeria, taxation has been in existence even before the coming of the colonial men or the British. Tax is one of the major sources of revenue to all levels of government. Taxation can be defined as a system of imposing a compulsory levy on all income, goods, services and properties of individuals, partnership, trustees, executorships and companies by the government (Samuel and Simon, 2011). In Nigeria, it is a factor to be reckoned with in Federal government's budget, the taxes so collected come back to the taxpayer in form of services. This has over the years encouraged or discouraged some activities in the private sector; this depends on whether the policy of the government is toward discouraging or encouraging such companies. Taxation is recognized as a very important tool for national development and growth in most societies. It has been viewed as a major vehicle for the long-term development of infrastructures of the state (Brima and Festus, 2014).\n2. Literature review: 2.1 Tax audit\n3. Research method: Qualitative and quantitative design was used in this study because both secondary and primary data were utilized in the study. The data used in this study are secondary in nature and they are quarterly. Quarterly data adopted in the analysis of this study spanned from 2000 to 2015 and were sourced from FIRS and LIRS audit division in Lagos state. The research instrument was structured questionnaire. The study adopted and modified the questionnaire as stated by Opoku (2015), who studied the role of a tax audit in revenue mobilization in Ghana revenue authority-the case of Ashanti region. The modified questionnaire was structured into five sections. Section A took into consideration the Demographic variables of respondents. Section B considered objective 1, Section C considered objective II, while Section D took into consideration objective III. However, a questionnaire was administered as used by Mutarindwa and Rutikanga (2014) to measure the growth in tax revenue in Rwanda. The questionnaire was open and close-ended, the response on the latter was ranked using the five-point Likert scale which ranged from \"Strongly Agree\" to \"Agree\" to \"Undecided to Disagree\" and to \"Strongly Disagree,\" while the former was on a free opinion of respondents. Both descriptive and inferential statistics were employed in analyzing the relevant data gathered for the study. Descriptive statistics include frequency counts and percentages, while inferential statistics were in the form of ordered logistic regression.\n4. Results and discussion: 4.1 Unit root test\n5. Conclusion and policy recommendations: Based on the findings of the study, it is concluded that tax audit variables enhance the level of productivity in tax administration in Lagos state which implies that adequate utilization of any form of tax audit will positively influence the revenue accruing to the government from taxation as a result of a reduction in the level of tax evasion, avoidance and an increase in voluntary compliance. Tax audit and investigation are critical to causing the taxpayer to be on their toes. This is particularly true for the taxpayer who is marginally complying and can easily be moved to the realm of tax evaders. Tax audit and investigation is also a means to protect government resources and make sure that defaulters do not go free for their offenses. In respect of the research findings, the study recommends that tax audit in the form of field, desk and back duty should be carried out on a routine basis to ensure that actual revenue collected is what the RTA remits to the government. Internal mechanism to check and monitor the staff of the tax audit department should be put in place to minimize the level of corruption and enhance the effectiveness of the tax audit. Tax audit department should be given autonomy to carry out their responsibility effectively as specified in Federal Inland Revenue Service Establishment Act 2007. RTA, tax auditors and FIRS staff members should have full knowledge of modern audit tools like Computer Aided Audit Tools (CAATs) to enhance performance and maximum tax revenue generation. All relevant authority should encourage tax audit for an increase in tax revenue, in order for the government to keep up with its obligation to her citizen.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the impact of the tax audit on tax productivity in Lagos state, Nigeria. Specifically, the study analyzed trends of tax audit and tax productivity, and the impact of Desk audit, Field audit and Back-duty audit on tax productivity in Lagos state.\nDesign/methodology/approach: The study made use of both primary and secondary data. Primary data used in the study were collected with the use of questionnaires administered to 350 randomly selected staffs of Lagos state Internal Revenue Services, while secondary data used in the study were sourced from Federal Inland Revenue Service and Lagos Internal Revenue Service audit division in Lagos state over the period spanning from 2000 to 2015. Data collated in the study were analyzed descriptively using inferential methods such as unit root test, and estimation techniques such as Fully Modified Least Square (FMOLS) co-integration regression and Logit regression analysis.\nFindings: The study revealed that Field tax audit, desk tax audit and Back duty tax audit exert a significant positive impact on tax productivity with reported estimate of 0.530454 (p=0.0044<0.05) for FIDAUD, 0.774450 (p=0.0085< 0.05) for DEKAUD, 1.244317 (p=0.0001<0.05) for BAKAUD.\nResearch limitations/implications: Relevant tax authority (RTA), tax auditors and FIRS staff members should have full knowledge of modern audit tools like Computer Aided Audit Tools (CAATs) to enhance performance and maximum tax revenue generation.\nPractical implications: The study concluded that tax audit enhances the level of productivity of tax administration in Lagos state and that any form of tax audit has the tendency of influencing revenue accruing to the government from taxation positively. Hence, tax audit should be carried out on a routine basis to ensure that actual revenue collected is what the RTA remits to the government. Tax audit department should be given autonomy to carry out their responsibilities effectively.\nSocial implications: Tax audit should be carried out on a routine basis to ensure that actual revenue collected is what the RTA remits to the government. Tax audit department should be given autonomy to carry out their responsibilities effectively.\nOriginality/value: This tax audit and tax productivity in Lagos state, Nigeria, fulfills an identified need to study how brand-supportive behavior can be enabled.\n",
"role": "assistant"
}
] |
Market and internationalization knowledge in entrepreneurial internationalization processes
|
[
"Internationalization",
"International opportunity",
"Market knowledge",
"International knowledge"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The process of knowledge acquisition and development as a mechanism of either uncertainty reduction or opportunity development is critical in understanding internationalization processes (Casillas et al., 2015; Fletcher et al., 2013). Initially, the accumulation of knowledge was perceived as a mechanism for uncertainty reduction, namely, knowledge gained from experience incrementally reduces the lack of information about a given foreign market (Johanson and Vahlne, 1977, 2009). More recently, the role of knowledge in the internationalization process has been associated with international opportunity development (Chandra, 2017), namely, the exploitation of international opportunities that leads to entry in new foreign markets and to new businesses in foreign markets serviced by the firm (Chandra et al., 2009). This conceptual paper subscribes to the latter viewpoint by looking at the internationalization process as an entrepreneurial process related to the development of international opportunities.\nKnowledge and opportunities: The bulk of the IB research on internationalization has looked at knowledge as a mechanism for uncertainty reduction (Petersen et al., 2008). In the internationalization process of the firm conceptualized by Johanson and Vahlne (1977), knowledge is gained incrementally from experience in a foreign market (i.e. experiential knowledge). When a firm starts servicing a foreign market, it faces challenges associated with the new environment, such as cultural (Kogut and Singh, 1988) and behavioral differences (Sanchez-Peinado and Pla-Barber, 2006), which bring uncertainty. By accumulating experiential knowledge, the firm learns how to operate in the new environment and to deal with those differences, reducing such uncertainty (Johanson and Vahlne, 1977). In this context, the most critical type of knowledge is experiential market knowledge (e.g. Figueira-de-Lemos et al., 2011), that is, knowledge gained from prior experience in a given foreign market (e.g. Chandra et al., 2012; Johanson and Vahlne, 1977). Thus, market knowledge is a central element of the internationalization process described by Johanson and Vahlne (1977).\nOpportunities and internationalization: As previously discussed, this paper looks at market knowledge and internationalization knowledge as antecedents of the development of international opportunities, which lead to the internationalization process, comprising both new foreign market entry and sequential moves that happen after entry. Foreign market entry is a central element of internationalization research, and therefore has been extensively studied (Shaver, 2013). Nevertheless, the recognition of international opportunities that leads to entry in new foreign markets is quite recent in IE (Jones et al., 2011). Studies on IE that connect opportunities and internationalization focus on international new ventures following a rapid internationalization process (Jones et al. 2011). However, Dimitratos et al. (2016, p. 1220) suggest that research in the IE literature \"[...] can shift the attention from the study of INVs to that of all opportunity-driven types of internationalized firms.\"\nThe moderating role of the international knowledge stockpile: On P1 and P2, market and internationalization knowledge are analyzed independently, to show how each type of knowledge relates to the development of international opportunities. On P3a and P3b, the development of international opportunities that follows the accumulation of market and internationalization knowledge is connected to two aspects of the internationalization process - new foreign market entry and sequential moves related to mode continuation, respectively. On P3c, the idea of a threshold that affects sequential moves related to modal shifts is introduced.\nConclusions: By combining the IB and IE literatures, this paper presents a novel framework connecting three constructs that are typically analyzed separately in the literature - knowledge, international opportunities, and the internationalization process. Grounded on the idea that the internationalization process is the result of the development of international opportunities, this study looks at the types of knowledge that shape such opportunities as well as how those types interact to moderate the relationship between international opportunities and the internationalization process. On the one hand, research on IE often connects knowledge to the development of international opportunities (e.g. Chandra et al., 2009), but without highlighting how such development affects the internationalization process of the firm, comprising both new foreign market entry and sequential moves in all foreign markets where the firm operates. This literature also focuses on international new ventures, thus disregarding other types of internationalized firms (Dimitratos et al., 2016). On the other hand, research on IB connects knowledge directly to internationalization, usually focusing on new foreign market entry (e.g. Shaver, 2013). It also posits that the main driver of the internationalization process is uncertainty reduction via knowledge development (Figueira-de-Lemos et al., 2011). This study explicitly connects knowledge, international opportunities, and the internationalization process during and after new foreign market entry and provides testable propositions that contribute to advance and bring together both the IB and IE literatures.\n",
"role": "user"
},
{
"content": "Purpose: By bringing together the IB and IE literatures, the purpose of this paper is to examine the internationalization process as an entrepreneurial process related to the development of international opportunities. It explicitly connects different types of knowledge (i.e. market and internationalization), international opportunities and the internationalization process comprising both new foreign market entry and sequential moves that happen after entry.\nDesign/methodology/approach: This is a conceptual paper that reviews the literature on knowledge, opportunities and the internationalization process. Moreover, the paper identifies the current gaps in the literature and builds new theory that sheds light into how these three concepts are related. The paper also presents a model and propositions that should guide future research.\nFindings: The proposed model shows that market and internationalization knowledge combine to form the international knowledge stockpile of the firm, which moderates the relationship between the development of international opportunities and the internationalization process, comprising not only new foreign market entry but also sequential moves that happen after entry using either mode continuation or modal shift. Moreover, it shows that the development of opportunities only leads to modal shifts after a certain threshold is achieved.\nResearch limitations/implications: The propositions suggest that market and internationalization knowledge are positively related to international opportunities, which, in turn, are related to foreign market entry and sequential moves using mode continuation. International opportunities, however, are related to modal shifts only beyond a threshold. Moreover, the international knowledge stockpile of the firm moderates the relationship between international opportunities and the internationalization process. Because this is a conceptual paper, the propositions have not been tested and, therefore, lack empirical validation. Nonetheless, the model is a starting point to new research on internationalization distinguishing different types of knowledge as well as different sequential moves.\nOriginality/value: This study shows that the internationalization process is contingent on the different types of knowledge associated with it. It also introduces the idea of a threshold that shapes the internationalization process. The bulk of the research on internationalization suggests that such a process is gradual and incremental. This study offers a non-linear alternative.\n",
"role": "assistant"
}
] |
The activating profession: coaching and coercing in the welfare services
|
[
"Social welfare",
"Governance",
"Professionalism",
"Norway",
"Activation policy",
"Street-level bureaucrats"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Throughout Europe welfare states are strengthening their efforts at moving non-working people into employment (Bonoli and Natali, 2012). Activation policies are a key element in these efforts. In a Scandinavian context, such programmes typically condition the right to income support on active participation in skill-enhancing activities with the aim of (re)integrating users into the labour market. The current tendency is that activation-style measures and conditions are being directed at ever wider target groups: the unemployed, lone parents, immigrants, social assistance recipients and people who are out of work due to poor health. For frontline workers in local welfare offices this means that a crucial part of their work consists in assessing claimants' ability to work and be activated, and devising plans for the claimants' progression from inactivity via active measures and into the labour market.\nAnalytical perspectives on activation: The shift from passive to active welfare arrangements constitutes one of the most significant changes to European welfare states in the last decades. Assessments vary. Some applaud efforts to enable people to help themselves and achieve self-sufficiency; others question the efficiency or normative sustainability of such policies (Molander and Torsvik, 2015). Such evaluations aside, the rise of activation policies can be interpreted in a wider context of citizenship restructuring, where active citizens are expected to work but also to manage their own health and well-being, participate in local communities, make active choices a consumers of welfare services and so on - thus being both liberated from the state and responsibilised for the welfare state's future (Clarke, 2005).\nThe empirical context, data and methods: Scandinavian activation policies are often distinguished from Anglo-American neo-liberal workfare policies in being about enabling employment through human capital enhancement, rather than benefit cuts (Dean, 2003). Torfing (1999) uses the term neo-statist as opposed to neo-liberal strategies in his analysis of Danish activation regimes. Also in Norway, the aim of freeing individuals from state dependency through activation is to happen by the means of intensive state activity in the shape of organisational reform and reconfigured frontline work.\nProfessionalism as governance - organisational professionalism: The entire NAV-reform can be read as an attempt to professionalise efforts to move people from social security schemes and into the labour market, and it was launched (among other things) as a competence reform. The reform further aimed to make users' encounters with the welfare services easier, to better adapt services to their needs and to coordinate all services and expertise relevant to help people into work. Two problems were thus addressed: the fact that too many working age individuals relied on welfare benefits instead of being employed and the lack of coordination between the different agencies that should be mobilised into achieving a more work focussed welfare policy (Andreassen and Fossestol, 2011).\nCompetence at the street-level - occupational professionalism: Two characteristics of professional work are that it is knowledge based and involves a significant amount of discretion. Activation workers have considerable space for discretion in their work, but the knowledge base on which these discretionary decisions are made is diffuse and heterogeneous. In this section I will look more closely on the discourse of the SLBs themselves. How do they speak about the competence they have and need in their work?\nConclusion: In this paper I have attempted to go beyond the observation that there is no generally accepted knowledge base for activation work (van Berkel et al., 2010), and instead looked at how this apparent competence void is being filled in the context of street-level activation work. A key finding is that professional competence in activation work is construed as knowledge about how to work with people, more than what to do with them. Rather than being orientated towards labour market expertise there is a tendency to psychologise ideas of a professional activation competence.\n",
"role": "user"
},
{
"content": "Purpose: Activation policies are key elements of contemporary welfare reform throughout Europe. The purpose of this paper is to explore the consequences of more active and individualised welfare policies for conceptualisations of professionalism and competence in the welfare services.\nDesign/methodology/approach: The primary data are 25 qualitative interviews with street-level bureaucrats (SLBs) conducted in two local offices in The Norwegian Labour and Welfare Administration (NAV). These data were supplemented by relevant policy documents. A distinction between the authorities discourse of organisational professionalism and the SLBs discourse of occupational professionalism is applied to structure the analysis.\nFindings: Efforts to professionalise activation work takes place in the absence of a specific professional knowledge base to guide daily work. The paper explores how relevant competence and skills are defined in such a context, both from the perspective of the authorities and from the front-level workers themselves. A key finding is that such competence tends to be defined in terms of the ability to manage communicative processes and relations. Paradoxically, the active turn in social policy with its emphasis on work and activity seems to entail a competence ideal that is inward looking and psychologised.\nResearch limitations/implications: The qualitative approach implies limited generalisability in terms of statistical representativity. Furthermore, the results invite closer studies of the practical effects for social security users of the identified patterns.\nPractical implications: Policy makers who aim to make welfare services more work orientated should look for ways of increasing SLBs concrete relations with and practical experience from collaboration with employers. This may entail reviewing the practice of outsourcing the implementation of active measures to private actors.\nOriginality/value: The paper adds to a small literature on the implementation of activation policy in contemporary welfare states.\n",
"role": "assistant"
}
] |
Makeup or mask: makeup's effect on salesperson trustworthiness
|
[
"Trustworthiness",
"Salesperson",
"Sales effectiveness",
"Service personnel"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: We all want to look and feel our best, and the cosmetic industry offers up easy, daily solutions to boost our perceived attractiveness and self-esteem. Probably the most common, short-term way to increase attractiveness is through the use of cosmetic products. Many women (and men) may attempt to increase their attractiveness by using every day cosmetic products; for example, 106.24 million Americans use foundation/concealer (over 1/3 of the US population), and this number is projected to reach 108.99 million in 2020 (Statista, 2018). The desire to improve attractiveness and the benefits of doing so are well established in the literature, especially the literature examining the world of sales and service personnel. People who score higher on perceived attractiveness in these roles benefit from higher sales performance, increased trustworthiness, and enhanced customer satisfaction ratings (Ahearne et al., 1999; DeShields et al., 1996; Jin and Merkebu, 2015; Reingen and Kernan, 1993; Soderlund and Julander, 2009) - appropriately dubbed the \"beauty premium.\"\nThe beauty premium: Can we take a salesperson or service worker at \"face value?\" Many researchers in the past have endeavored to answer this question by exploring the attractiveness ratings of salespersons and service people and the effect of perceived attractiveness on performance. The idea that beauty is currency, termed the \"beauty premium,\" is not merely a social myth but is also supported by numerous studies that investigate the \"plainness penalty\" and how beauty is truly the gift that keeps on giving - demonstrating the halo effect across various life outcomes (Fletcher, 2009; Hamermesh and Biddle, 1994; Mobius and Rosenblat, 2006; Van Leeuwen and Neil Macrae, 2004). Physically attractive individuals are viewed as being more friendly, intelligent, happy and socially adept (Commisso and Finkelstein, 2012; Eagly et al., 1991; Keh et al., 2013).\nMakeup or mask: conceptions and misconceptions of beauty: One commonly accepted way to enhance beauty is through the use of cosmetic products. The United States is home to the largest cosmetic market in the world; revenue in 2018 exceeded $51 billion, and the cosmetics industry experienced 3.9 per cent growth in 2018 alone (Diehl, 2018). In per capita terms, the average woman uses about five cosmetic products on a daily basis (Stowaway Cosmetics, 2015) and over 50 per cent of consumers in all age groups use makeup at least several times per week (Statista, 2017). However, what happens when the pursuit of beauty goes too far? Can too much of a good thing, like makeup, have counteractive effects?\nImportance of trust in salesperson-customer relationships: Social order in our society is contingent on the existence of stable relationships that are inherently dependent on trust built between individuals (Lewis and Weigert, 1985). All social exchanges require some level of trust, but when risks are involved in such exchanges, the importance of trust is heightened (e.g. in consumption situations where the consumer wants to choose the right product/service for them and make the best decision possible) (Barber, 1983; Blau, 1964). In a consumption context, buyer-seller relationships are built around expectations of cooperation and the forming of mutually beneficial relational contracts (Dwyer et al., 1987); specifically, trust is defined as \"the belief that a party's word or promise is reliable and a party will fulfill her/her obligations in an exchange relationship\" (Blau, 1964, p. 940). In the more specific context of sales or customer service encounters, the key to trust is that the customer is confident that the salesperson can be relied upon to have the customer's long-term best interests at heart. However, this emotional component of trust is paired with a cognitive component such that the service provider must also be \"able\" to provide the information or perform the task needed (Crosby et al., 1990; Lewis and Weigert, 1985).\nStudy 1: Procedure\nStudy 2: To enhance the generalizability of our findings and acknowledge the growing demand for male cosmetic products, study two examines perceptions of a male salesperson with varying levels of makeup usage.\nGeneral discussion: Although prior research finds effects for salesperson attractiveness in terms of enhanced performance (Ahearne et al., 1999; DeShields et al., 1996; Jin and Merkebu, 2015; McColl and Truong, 2013), these results indicate that attractiveness and perceptions of beauty may be specific to perceptions of natural beauty. Our study manipulates makeup usage level across several levels and finds significant negative effects on perceived trustworthiness, purchase intentions and sales effectiveness.\nLimitations and future research: In the current work, we find negative effects for overuse of makeup but fail to replicate a beauty premium with \"natural\" looking makeup. Future work should examine the limitations of makeup use on those who are already perceived as attractive, and conversely, the optimal use of makeup on those who may be well-served by moderate levels of makeup usage (based on starting attractiveness levels). For example, the female model used in Study 1 has a fairly high starting level of attractiveness (Mno makeup = 2.9, where 1 = most attractive on a seven-point scale), so the amount of makeup needed for her to reach a natural beauty level may differ from others that need a bit more, while still achieving a \"natural look.\" The competing effects of true natural beauty and makeup usage to achieve \"attractiveness\" are under explored in the literature. Future work may further examine the counter-intuitive effect of makeup over usage uding individuals with a variety of starting attractiveness levels, to better understand the nuances of optimal makeup usage and its effect on perceived trustworthiness. If the salesperson benefits from a high level of natural beauty to start, does the use of makeup lead to differing results from one that may have a fairly low starting rating of natural beauty?\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to investigate how the use of varying amounts of makeup by sales personnel influences perceived salesperson trustworthiness and downstream purchase intentions/sales effectiveness.\nDesign/methodology/approach: Two studies were run with female and male makeup usage examined separately. In each study, a between-subjects, scenario-based experiment was run on Qualtrics in which participants were randomly assigned to conditions representing a salesperson with varying levels of makeup usage (none to glamorous/extreme).\nFindings: Using ANOVA, the authors find that the glamorous/extreme makeup condition led to significantly lower perceptions of trustworthiness. Furthermore, mediation models for both studies found that increased salesperson makeup predicted lower perceived trustworthiness, resulting in lower purchase intentions and sales effectiveness.\nPractical implications: Managers and salespersons can benefit from these findings through increased awareness of the \"masking\" effect of wearing too much makeup and the insight that more is not always better when it comes to the use of makeup to increase salesperson attractiveness and effectiveness.\nOriginality/value: Much work has been done in the past on attractiveness of salespersons in general and the \"beauty premium.\" The present work expands on this literature by examining a specific strategy, yet to be rigorously examined, that sales personnel may use to increase their attractiveness - makeup, and by identifying an \"optimal\" level of makeup usage.\n",
"role": "assistant"
}
] |
How do emotional intelligence and behavioral biases of investors determine their investment decisions?
|
[
"Behavioral biases",
"Emotional intelligence",
"Investment decisions",
"Multiple regression",
"Relationship"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In the present situation, money has a significant influence on everybody's life. So as to maintain a strategic distance from various kinds of issues throughout everyday life, the individual ought to put the cash in various sorts of roads. The Indian budgetary framework supports the reserve funds among the financial specialists and channels them to their ideal and powerful use. In the present aggressive time, the different speculation roads are accessible to the speculators. Financial specialists when all is said in done have the energy to put resources into those specific roads, which will deliver the most extreme comes back with the least dangers. The speculators settle on the venture choices based on the various variables. The venture choices remember for different speculation procedures, recurrence of the venture, timeframe, targets of speculation, factors influencing speculation choices and some more. The conduct predispositions are the most significant factor while settling on venture choices (Camerer, 1997; Bailey, 2012; Breuer et al., 2014). This segment manages the presentation of the components of the passionate insight and the elements of the conduct predispositions of the financial specialists.\n2. Review of literature: Kunnanatt (2004) discussed the emotional intelligence. He observed that people with high emotional intelligence would produce a win-win relationship and people with low emotional intelligence would produce a win-lose relationship. The emotional intelligence training programs would change both the inside and outside relationships of participants and better attitudes, clearer perceptions and productive affiliations in life. Avsec et al. (2009) analyzed the connection between the emotional intelligence of Croatian and Slovenian college understudies and their identity attributes. The information was gathered from 257 college understudies and 171 college understudies. The information was dissected by regression, relationship and two way ANOVA. They reasoned that extraversion trait and conscientiousness trait were vital indicators and the neuroticism trait was the most grounded indicator.\n3. Objectives of the study: To study the relation between the dimensions of emotional intelligence and investment decisions; and\n4. Research methodology: Test size: the information was gathered from 500 financial specialists.\n5. Analysis: This section represents the relationship among the investors' emotional intelligence, behavioral biases and investment decisions. This section is divided into three parts - first part describes the relation between the dimensions of the behavioral biases and the investment decisions of the investors; the second part describes the relation between the dimensions of the emotional intelligence and the investment decisions of the investors; and third part describes the overall relation among the behavioral biases, emotional intelligence and the investment decisions of the investors. Multiple regression techniques were applied to test the relationship between these variables.\n6. Findings and suggestions: The investigation found that the positive connection between the conduct predispositions of the financial specialists and venture choices of the speculators and positive connection between enthusiastic insight of the financial specialists and their venture choices. Yet, we found that the enthusiastic insight better foresees the venture choices of the financial specialists than the conduct predispositions of the speculators. Among the different elements of conduct inclinations of the speculator's lament and carelessness are identified with the financial specialist's venture choices. Among the various estimations of eager understanding - care, dealing with emotions, motivation, empathy and social aptitudes are related to the hypothesis decisions of the monetary pros. The examination shows that out of these two passionate insights and social predispositions of the speculators, enthusiastic knowledge impact venture choices of the financial specialists more than the conduct inclinations of the financial specialists. The financial specialists who consider their past experience put more and put safely in securities exchange and go out on a limb. Everyone will have the experience of feeling mourn for the duration of regular daily existence and it is common. The money related pros feel regret when they buy at a significant expense and offer to require little to no effort. They will in general hold the stocks excessively long until the moment that the expense accomplishes lower than the buying cost. The theorists who reliably do attempts toward the pined for targets or vanquishing negative sentiments put more in protections trade. The speculators who are submitted and are constantly started toward their work put more in the securities exchange. The theorists should have been full data of the endeavor streets. In this way, the associations should give complete consideration to the money related pros, which help them to pick the right theory at the perfect time. The hypothesis associations ought to deal with the theorists according to their social inclinations and the energetic understanding property while placing assets into protections trade. The individual money related master should not take after the theorists who consider their past experience or who contribute basically in light of the fact that other budgetary experts put assets into that particular street. Thus, the associations need to base on the individual examiner for getting the whole market. The reserve associations should design the endeavor streets according to the conduct inclinations and the enthusiastic information on the monetary authorities. The associations must offer some motivator added organizations to examiners. The financial specialists must not pursue most of the individuals. He ought to know about his ventures while putting resources into any stock. The financial specialists ought not to put resources into stock on the guidance of the others. The financial specialist ought not to put together their venture depending on respect to the exhibition of the stock previously. The money related counsels should concentrate on the arrogance and lament conduct predispositions of speculators while settling on venture choices with the goal that they can guide them precisely so as to reduce such sort of inclinations. Speculators ought to show restraint when taking venture choices. Financial specialists ought to painstakingly recognize and examine the social element, which influences the venture choices of the speculators. The financial specialist needs to build their insight with respect to social accounts. The speculators of all age bunches must have information with respect to the venture, so they can choose the best speculation road in like manner. The organizations ought to give data to the financial specialists with respect to the advantages of making interest in securities exchange.\n7. Conclusion: The financial specialists decide their venture choices as per their behavioral predispositions and enthusiastic knowledge. The financial specialists with various enthusiastic knowledge put resources into various roads. The organizations ought to consider the distinctive behavioral inclinations and enthusiastic knowledge of the speculators keeping in mind the end goal to give them the best venture avenues. Thus, the companies should give the best advice to their clients on the basis of the different factors affecting the investment decisions of the investors.\n",
"role": "user"
},
{
"content": "Purpose: In earlier studies, research has shown that EI is the only element, which influences the ways in which people develop in their lives, jobs and social skills control their emotions and get along with other people. It is EI that dictates the way people deal with one another and understand emotions. The research gap is to explore the impact of behavioral factors and investors psychology on their investment decision-making.\nDesign/methodology/approach: The information was gathered from 500 financial specialists. The region of research was the financial specialists who contribute through LSC Securities Ltd. in Punjab State. The purposive testing system was used in this examination.\nFindings: The investigation found that the positive connection between the conduct predispositions of the financial specialists and venture choices of the speculators and positive connection between enthusiastic insight of the financial specialists and their venture choices. Yet, the authors found that the enthusiastic insight better foresees the venture choices of the financial specialists than the conduct predispositions of the speculators. Among the different elements of conduct inclinations of the speculator's lament and carelessness are identified with the financial specialist's venture choices. Among the various estimations of eager understanding - care, dealing with emotions, motivation, empathy and social aptitudes are related to the hypothesis decisions of the monetary pros.\nResearch limitations/implications: The sample selection was based on purposive sampling, rather than a random probability sample. The sample was area specific, restricted only to Ludhiana Stock Exchange in Punjab state. Therefore, the results of the study cannot be generalized with certainty to all the investors investing through other exchanges in other states. The inferences are based on the assumption that the data provided by the investors are true and correct. The findings may be relevant for other stock exchanges as that of the Ludhiana Stock Exchange. However, the authors do not claim the generalization of the results.\nPractical implications: This study also helps to understand the relationship between investment decision-making and risk tolerance of investors. It will helpful for the financial advisors to know the behavioral biases of investors while making an investment decision, and therefore, they can advise investors properly to mitigate such biases. It may help the investors in understanding the subjective part of their behavior and control their emotions while taking decisions for their investment in stock market options.\nSocial implications: This research will help investment advisors and finance professionals to judge investors' attitudes toward risk in a better way, which leads to better investment decisions.\nOriginality/value: This study is my own study and it is original and has not been published anywhere.\n",
"role": "assistant"
}
] |
Hallowed be thy brand: measuring perceived brand sacredness
|
[
"Brand sacredness",
"Extraordinary experience",
"Religion like properties",
"Brand evangelism",
"Scale development and validation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The fanatics who camp for days on end in order to be first in line for the latest iPhone or iPad. They weep with joy when Apple announces something new, and cry hot tears of frustration whenever Fandroids dare to post a negative opinion about Apple. For them, Apple is a religion, loyal allegiance, a way of life. Apple attracts love (and the other stuff) more than any other tech company [...] It inspires devotion amongst legions of fans, around the world. Heck, I'm writing this on Macworld, because Dell world and HP world just couldn't get off the ground (Egan, 2015).\n2. Conceptual background and construct domains: Our conceptualization of the brand sacredness construct is built upon the extant literature that distinguishes between \"sacred\" and \"profane.\" According to Belk et al. (1989, p. 2), while profane pertains to the everyday mundane reality, sacred \"fulfills a need to believe in something significantly more powerful and extraordinary than the self - a need to transcend existence as a mere biological being coping with the everyday world\". As the sacredness concept suggests gaining extraordinary experiences beyond daily ordinary life experiences after encountering the sacred (Roberts, 1984), it can also be used in contemporary consumption realm when an ordinary consumption object like a brand may be revered as extraordinary. Durkheim (2001) posits that even everyday mundane entities like lizards and plants could take on the sacredness of the 'totem' or become a symbol of the tribe. Hence, a brand can also be sacralized when it is imbued with certain extraordinary properties that would make some consumers perceive the focal brand as sacred over all other profane brands. When a brand attains such a sacred status, it becomes a potential substitute for a religion, thus fulfilling human needs for deriving life meanings and self-expression (Cutright et al., 2014; Shachar et al., 2011; Wang et al., 2018).\n3. Scale development and validation: As brand sacredness involves ascribing certain properties of scared to an otherwise secular/profane brand (Belk et al., 1989; Beckford and Demerath, 2007), Brand Sacredness Scale items should be able to measure the degree to which sacredness is perceived in the focal brand by the brand believers and should be able to capture varying extents of brand sacredness across brand fans and non-fans. Leveraging on prior research that describes the concept of sacred and delineates its properties from the domain of religiosity (Belk et al., 1989; Mathras et al., 2016), our first goal was to develop a set of items that can tap the content domains of brand sacredness construct. The initial item generation was done by referring to the extant literature regarding sacred and sacralization in the fields of theology, sociology and consumption.\n4. General discussion: Previous research has identified pervasiveness of sacred in consumption domain and has reported findings of extraordinary and at times transcendental meanings in mundane consumption objects and situations (Belk et al., 1989; Kozinets, 2001). The primary objectives of this article are to operationalize the brand sacredness construct and then develop a valid measure reflecting the strength of the perceived sacredness of a brand.\n5. Theoretical implications: This research contributes to the extant literature that explores and describes the sacred dimensions of secular consumption from the hermeneutic approach. The existing studies on sacralization of consumption are mostly based on in-depth interviews or case analyses of the phenomenon. However, to better understand the phenomenon, there is a need to clearly define the construct and delineate the construct domain. In this regard, this manuscript introduces the study validating a scale to measure Brand Sacredness, and this is the first work to our knowledge to engage in development of a measure for Brand Sacredness. As Monson (1985) pointed out, performance measure and report would accelerate the understanding of the concept and enhance the performance. Therefore, the development and validation of a scale for Brand Sacredness will improve the understanding of consumer-brand-consumer triad by providing a measurement tool that can help marketers capture and report the strength of consumers' perception of brands as sacred entities.\n6. Managerial applications: The Brand Sacredness Scale captures brand sacredness on a continuum, suggesting that a secular brand has a potential to be sacralized though management and consumer interactions. Brands can be evaluated along the sacredness continuum (with low to high extent of perceived sacredness) in which some brands are highly regarded as sacred brands by their ardent fans. However, brands without global recognition and with a relatively small niche of fans, like Mazda Miatas (Wells, 2001), Apple Newton (Muniz and Schau, 2005) and Manolo Blahnik shoes (Wintour, 2003) also have the potential to attain sacred status. As brand sacralization is a brand co-building process, brand managers can imbue sacred properties into the brand through marketing campaigns and interact with consumers and brand fans to create a sacred brand though fandom management.\n7. Limitations and future research directions: This study is based on cross-sectional survey data obtained from respondents belonging to well-established brand communities. A longitudinal study involving recent and emerging brand communities could provide an enhanced understanding of the evolution of brand sacredness with time, including brand sacralization process as well as a possible de-sacralization process.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to capture the strength of consumer's perceived brand sacredness. The authors developed and validated a measurement scale composed of three related dimensions: supremacy, mesmerization and communitas.\nDesign/methodology/approach: Six empirical studies were conducted to identify the brand sacredness construct domains, develop and validate the measurement and test the nomological network between brand sacredness and it antecedent and outcome variables.\nFindings: Results from a series of studies provided robust supports for the scale structure and demarcated the construct domains from other consumer-brand relationship measures. Testing of nomological validity of the scale further showed that brand sacredness is influenced by brand love, emotional brand attachment and brand loyalty and, meanwhile, provides explanatory power to predict theoretically related outcome variables, including transcendent consumer experience, defense of brand, incorporation brand in extended-self, brand ritualism and brand evangelism.\nResearch limitations/implications: This study is based on cross-sectional survey data obtained from respondents belonging to well-established brand communities. A longitudinal study involving recent and emerging brand communities could provide an enhanced understanding of the evolution of brand sacredness with time, including brand sacralizaton process as well as possible de-sacralization process.\nPractical implications: The study provides significant insights for brand managers to create an enduring brand and ascertain that consumers find their affiliations with the brand and make it the sacred core of their lives by fandom management through brand evangelism.\nOriginality/value: This study adds to the theory on consumer-brand relationship realm by delineating the domains of brand sacredness with its defining feature of extraordinary experience transcending an ordinary brand. It contributes to the existing body of branding and customer-based brand equity literature by incorporating the spiritual aspects of faith, passion and devotion into measuring the value of a brand.\n",
"role": "assistant"
}
] |
An examination of firms' strategic orientations, innovativeness and performance with large Korean companies
|
[
"Firm innovativeness",
"Customer orientation",
"Competitor orientation",
"Technology orientation",
"Internal/cost orientation",
"Performance outcomes of innovativeness"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Firm innovativeness (hereafter, INNO) has been regarded as a critical component for the success of firms in many academic fields, such as marketing, strategic and organizational management, international business, etc. (Hauser et al., 2006; Rubera and Kirca, 2012). It has been conceptualized in various ways, such as INNO inputs (i.e. efforts made toward innovation, e.g. R&D expenditure and patents), INNO outputs (i.e. consequences of innovation activities, e.g. incremental and radical innovation products) and INNO culture (i.e. the firm's ability to constantly introduce new products) (Rubera and Kirca, 2012). We are interested in the INNO as culture (i.e. the process view of innovation) for multiple reasons. First, it was found that firm INNO culture is positively related to the organization's ultimate performance outcomes such as financial position (e.g. profits) and firm value more than the other INNO constructs (Hurley and Hult, 1998; Rubera and Kirca, 2012). Second, it has been less examined in terms of its important drivers and consequences. Last, we are interested in the firm's strategic issues of the higher-order level. Likewise, the stream of INNO research found the importance of and called for further examination of corporate culture in driving INNO and its performance outcomes (Hauser et al., 2006; Shane and Ulrich, 2004; Tellis et al., 2009; Walker, 2004).\n2. Conceptualization and hypotheses: We start by presenting the result of our conceptualization in advance and then explain our reasoning behind the hypotheses. Figure 1 summarizes our research model. We propose the relationships among four different strategic orientations and INNO with three definite directional associations. More specifically, we derive and test the layered relationships among strategic antecedents, INNO and firms' performance outcomes, such as CS, market adaptability (ADAPT) and firm profitability.\n3. The research method: The location context of the study is the metropolitan area of Seoul, Korea. Given its global reputation as an advanced emerging country, we think that Seoul is a reasonable context in that it possesses unique mixtures of modern companies' orientations such as CO, PO, TO and IO. More importantly, as argued in the Introduction, it is very important to accumulate the empirical evidence on the antecedents and consequences of INNO in Eastern countries, given the sheer number of studies accomplished in Western countries. Compared to Zhou et al.'s (2005) analysis of Chinese companies and Rhee et al.'s (2010) inquiry of Korean small- and medium-sized companies, our study aims to test our hypotheses in the context of relatively large Korean companies, contributing to enhancing the generalizability of findings about the relationships among firms' strategic orientations, INNO and performances.\n4. Analyses and results: 4.1 Basic analysis\n5. Discussion and conclusion: 5.1 Summary and conceptual implications\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to provide an examination of firms' strategic orientations, innovativeness and performance with large Korean companies.\nDesign/methodology/approach: The authors investigated the impacts of firms' major key strategic orientations (customer orientation [CO], competitor orientation [PO], technology orientation [TO] and internal/cost orientation [IO]) on firm innovativeness (INNO) and performance outcomes with large Korean companies.\nFindings: The results of the analysis showed that CO, PO and TO positively influence the innovativeness, which contributes to firm performance.\nOriginality/value: The authors provide some managerial implications on the multiple roles of strategic orientations on firm INNO and performances, along with limitations of this study and future research directions.\n",
"role": "assistant"
}
] |
What's best for whom? The effect of product gender depends on positioning
|
[
"Product positioning",
"Conceptual congruity",
"Product credibility",
"Product gender"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Companies are increasingly using visual design elements, such as color and shape, for promoting a product's gender to appeal to desired target customers. Indeed, recent research suggests that strongly gendered products outperform weakly gendered products (van Tilburg et al., 2015). For example, the company Jack Black, which is the number one seller of prestige skincare products exclusively for men, consistently uses dark colors, such as blue and black, in their packaging to communicate the products' masculinity. In contrast, the company Avon, a top seller of skincare products exclusively for women, predominantly uses light colors, such as white, or the colors pink and purple to communicate the products' femininity. At the same time, product positioning is an essential marketing strategy. Jack Black promotes how its shower gel \"gets rid of dirt and sweat without stripping skin\" (functional benefit), while Avon promotes how its shower gel \"leaves skin feeling soft and conditioned\" (hedonic benefit). However, despite the implications for marketing strategy, whether the communicated product benefits influence consumers' evaluations of gendered products has yet to be explored.\nTheoretical background: Gendering brands and products\nSelection of product categories: As the aim of this research is to investigate the effects of product gender (evoked through the product's design) on product evaluation, it was critical to find product categories that are not characterized as either feminine or masculine by nature. Thus, similar to Debevec and Iyer (1986), a pretest was conducted to select gender-neutral product categories that are used by both female and male consumers in order not to confound the effects of product gender with product category gender. For this purpose, a total of 51 US consumers (Mage = 33, 43 per cent female), recruited via Mechanical Turk, rated several product categories in terms of their femininity (not feminine at all/very feminine) and masculinity (not masculine at all/very masculine) on seven-point scales (Fugate and Phillips, 2010). Participants were not exposed to any product pictures, but indicated their perceptions of product category femininity and masculinity based on the following preamble: \"How would you rate [product category]?\"\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to investigate how product positioning affects the influence of product gender on consumers' product evaluations.\nDesign/methodology/approach: Using experimental designs, this research investigates how hedonic versus functional consumption goals affect consumers' choice between feminine and masculine products (Study 1) and how positioning products as either hedonic or functional influences consumers' evaluations of feminine and masculine products (Studies 2 and 3).\nFindings: When pursuing hedonic consumption goals, consumers are more likely to choose feminine (vs masculine) products, whereas when pursuing functional consumption goals, consumers are more likely to choose masculine (vs feminine) products. Further, consumers evaluate feminine products more favorably when the products are hedonically (vs functionally) positioned, whereas they evaluate masculine products more favorably when the products are functionally (vs hedonically) positioned. Perceptions of product credibility mediate this effect.\nResearch limitations/implications: Connecting theories of gender identity, product positioning and congruity, this study extends previous literature by demonstrating that the effects of product gender are context-dependent.\nPractical implications: Many companies use visual design cues (e.g. shape, color) to promote their products' gender. The findings of this study suggest that companies promoting their products as feminine should highlight the products' hedonic benefits, whereas companies promoting their products as masculine should highlight the products' functional benefits.\nOriginality/value: Applying a conceptual congruity approach, this research is the first to demonstrate that the effects of product gender on consumers' product evaluations depend on the product's positioning.\n",
"role": "assistant"
}
] |
The integration of shop floor control in wafer fabrication
|
[
"Shopfloor",
"Production planning and control",
"Despatch",
"Semiconductors",
"Taiwan"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Semiconductor chip is a complex device that consists of miniaturized electronic components and their connections. There are five steps in semiconductor manufacturing: wafer fabrication, wafer probe, device assembly, class test, and final test. Wafer fabrication is most technologically complex and capital intensive. Since, required capital investment is extremely large, the implementation of an improved shop floor control (SFC) strategy could result in a considerable amount of increase in profit. It is challenging to develop the SFC strategy for a wafer fab due to the long flow time, ever-changing products yielded, re-entrant feature of the production sequence, and stochastic aspects of the wafer fab including machine failures. There are basically two types of SFC strategies at wafer fabrication stage. The first and the most familiar one is dispatching. Each time a workstation is ready to commence processing another order and there is a queue of work-in-process (WIP) waiting to be processed at the workstation, the dispatching strategy selects an order or orders to start processing. The second type of SFC strategy, which is referred to as the order review and release (ORR) strategy, determines the type, amount, and time point of release of raw wafers so as to maximize the efficiency of the wafer fab.In the past two decades, ORR has been the most important decision strategy in production control. The mean and variance of the workload directly depend on its ORR strategy in the system even though dispatching strategy will impact some performance indicators, such as WIP, flow time, and throughput rate. In practice, ORR and dispatching strategies should be complementary to each other. Besides, defective wafers occur often in wafer fab, wasting raw materials and lowering utilization rate of manufacturing machines. Owing to the high cost involved, finding the best strategy to reprocess defective wafers is critical to the goal of overall manufacturing efficiency. The decision on rework strategy is affected by the dispatching strategy in photolithography. Hence, we will consider ORR, dispatching, and rework strategies simultaneously in this research.The purpose of this paper is to find the best combination of SFC strategies for improving the system performance. Considering the trade-off among some system performance indicators, such as WIP, flow time, tardiness, and delay cost, we attempt to discuss the impact of these SFC strategies on specific system performance indicators.The remainder of this paper is organized as follows. Section 2 will summarize the relevant literature on SFC strategies. Section 3 will discuss classification and selection of SFC strategies. Section 4 will describe the simulation model and the experiment design. In Section 5, the simulation results will be presented, and the proposed SFC strategies will be discussed. The conclusions and suggestions for future study will be included in Section 6.\n2 Literature review: In wafer fabrication, the integrated circuit is constructed on a silicone wafer layer by layer. The manufacturing procedure for each layer includes washing, oxidizing/thin film deposition, masks manufacturing, photolithography, etching, diffusion/ion implantation, inspection and measurement. The way these steps are executed depends on the design of the integrated circuit. Wafer manufacturing has several characteristics (Sha et al., 2001):* reentry is permitted;* longer production cycle time;* possibility of rework; and* problem of waiting time.Owing to the above factors, the many types of wafers to be processed, and also the differing machine malfunctions, it is important to exercise reasonable control over the production environment.2.1 Order review/release strategies\n3 Selection of SFC strategies: This paper is focused on the interaction of SFC strategies. Firstly, we should find the representative SFC strategies. We try to define a classified structure for each kind of SFC strategies and find the most popular strategies by survey of the researches related to SFC in wafer fab since 1990. Furthermore, the opinions of both practitioners and researchers are considered in selecting SFC strategies. We will include the representative strategies in our simulation model.The model will be very complex to include all the algorithms in our research. Furthermore, some algorithms have the similar nature and someone will be more excellent than other, for instance, SPT and TSPT. These algorithms included our research are representative and outstanding at each type of strategies. It will be obvious that the result of statistical test will be limited in those algorithms. However, the result of our test will be stable and representative, if the algorithms included in the model has not the significant different than those one we used.3.1 ORR strategy\n4 Simulation experiments: To compare the integrated strategies in this study, a series of simulation experiments is performed. Performance measures used for the comparison are WIP, flow time (FT), standard derivation of flow time (FT_STD), tardiness, percent of tardy jobs (tardy rate), and delay cost, which are often used in mass production environment. The due dates are determined internally as: Equation 1 di - due date of lot i, ai - arrival time, TPT - total processing time.The function of delay cost is defined as follows: Equation 2 Equation 3 4.1 The wafer fab model\n5 Suggestions on SFC strategies: The interactions of SFC strategies are quite significant (as shown in Table IV), especially the interaction between ORR and dispatching, ORR and rework. The interaction between dispatching and rework is significant in terms of FT-STD and tardy rate.5.1 The performance of SFC strategies\n6 Conclusions and future research: ORR, dispatching, and rework strategies are very important to production control in wafer fab. In our research, we examined the interaction among three kinds of SFC strategies. The interaction among ORR, dispatching, and rework strategies is significant. Although ORR has the largest impact on the system performance, the selection of properly integrated strategies will significantly improve the system performance. In our research, we can make some suggestions about these SFC strategies based on simulation and statistics analysis. Load-oriented strategies, TB/WR, are outstanding among the seven ORR strategies. In eight dispatching strategies, NexQL is better than the others as indicated by WIP, FT, and tardy rate. COVERT is the best one in terms of FT-STD and delay cost. Same as the traditional opinion, EDD has the minimum average tardiness. In rework strategies, lot-spit is the best one except for FT-STD and delay cost. The difference between lock-step and lot-spit is not significant in Duncan's test. However, significant difference can be found when the interaction between dispatching and rework strategies is considered. NexQL x lot-spit will obviously improve WIP, FT, and tardy rate. EDD x lot-spit is outstanding in improving tardiness and delay cost. In terms of FT-STD, COVERT x lock-step is better than the others. The suggested combinations of ORR, dispatching, and rework strategies include TB x SRPT x lock-step (WIP, FT), TB x EDD x lock-step (tardiness, tardy rate), and WR x COVERT x lot-spit (FT-STD, delay cost). The performances of most ORR strategies will be improved when combined with suitable dispatching and rework strategies. No single strategy can satisfy all performance indicators. In practice, we should choose SFC strategies carefully based on the system conditions. Besides, ORR, dispatching, and rework strategies cannot be separately considered. Instead, they should be combined and integrated for system performance improvement.In the future, there are some topics can be discussed. The integral strategy of SFC can be developed. ORR or dispatching strategies considering the demand for rework will improve the system performance. On the other hand, the integration of SFC strategies can be extended to due date assignment strategies, especially, for improving the performance related due date.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to consider the interaction among shop floor control (SFC) strategies (order review/release, dispatching, and rework strategies) and its impact on the performance of wafer fabrication. How to select a suitable set of SFC strategies in wafer fabrication is the main topic.\nDesign/methodology/approach: - The paper tried to find the better combination of these strategies by specific performance indicators. A virtual wafer fab is built based on a real case in Taiwan and used for simulation test. Seven ORR rules, eight dispatching rules, and three rework rules are considered in the simulation test.\nFindings: - The paper finds that the performance of most ORR strategies will be improved when combined with suitable dispatching and rework strategies. But no single strategy can satisfy all performance indicators.\nPractical implication: - The performance of most ORR strategies will be improved when combined with suitable dispatching and rework strategies. No single strategy can satisfy all performance indicators. In practice, we should choose SFC strategies carefully based on the system conditions. Furthermore, ORR, dispatching, and rework strategies cannot be separately considered. Instead, they should be combined and integrated for improving the system performance.\nOriginality/value: - Simulation tests and statistical analysis are done for three kinds of SFC strategies on a virtual wafer fab. The suitable combinations of SFC strategies for different performance indicators are suggested in this paper. The selection of ORR, dispatching and rework rules in wafer fabrication are discussed in the paper.\n",
"role": "assistant"
}
] |
Drivers of customers' switching behaviour in Indian banking industry
|
[
"Banking industry",
"Customer retention",
"Switching behaviour"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Indian banking space is occupied with a large number of small banks. The increasing use of technology along with liberalization and deregulation has brought a revolution in the way of serving customers. Technology is unsettling the earlier business processes and customer behaviour is undergoing change. Nevertheless, the fundamental aspects of banking, i.e. trust and confidence of the people on the institution remain the same. This scenario of banking environment has presented many opportunities and threats in front of the management.\n2. Background: The economic reforms initiated by the government of India in early 1990s have changed the landscape of Indian banking sector. The country has 87 scheduled commercial banks with deposits worth Rs.71.6 trillion (US$1.21 trillion) as on 31 May 2013. Of this, 26 are public sector banks, which control over 70 per cent of India's banking sector, 20 are private banks and 41 are foreign banks (Figure 1). The contribution in economic growth is also rising significantly as deposits increased to 71.3 per cent of gross domestic products (GDP) in FY11 from 37.4 per cent in 1991, lending grew to 53.9 per cent of GDP from 22.6 per cent and bank credit to commercial sector increased to 57.6 per cent of GDP from 30.3 per cent in same year. Despite the remarkable growth, only 35 per cent of all adults in India had a bank account with a formal banking institution (Demirguc-Kunt and Klapper, 2012). This massive opening has shifted the bank's focus to Indian consumer. The emerging competition has generated new expectations from the existing and the new customers. The competitive strategy of banks requires recognition of the importance of consumers.\n3. Literature review: 3.1 Switching behaviour\n4. Research methods: 4.1 Data collection\n5. Results and discussions: 5.1 Descriptive statistics\n6. Conclusions and implications: 6.1 Major conclusions\n",
"role": "user"
},
{
"content": "Purpose: The price war and intense competition in Indian banking industry have exposed banks to one of the major threat of switching. Consumers are now more price and service conscious in their financial services purchasing behaviour. They are more prone to change their banking behaviour as banking products and services are nearly identical in nature. The purpose of this paper is to provide an insight of the drivers that lead a customer switch from one service provider to another in Indian banking industry using exploratory design.\nDesign/methodology/approach: The impacts of the influencing factors have been studied and tested empirically using exploratory factor analysis. Quantitative data have been collected by means of questionnaire employed from Clemes et al. and administered to 296 banking customers of Rajasthan utilizing convenience sampling.\nFindings: Results reported that price, reputation, responses to service failure, customer satisfaction, service quality, service products, competition, customer commitment and involuntary switching have their significant effect on customers' switching behaviour.\nResearch limitations/implications: The findings of present study can be used by the Indian banks for their product and service designing strategies, marketing strategies and customer services practices in order to reduce customer switching. It would help them in improving their service operations and also in increasing customer satisfaction and loyalty by understanding the banking behaviour of their customers.\nOriginality/value: The originality lies in the fact that this study is one of few which have focused on the drivers leading to the switching intentions of Indian banking customers.\n",
"role": "assistant"
}
] |
Bank institutional setting and risk-taking: the missing role of directors' education and turnover
|
[
"Corporate governance",
"Cooperative banks",
"Bank ownership",
"Board of directors",
"Bank risk"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In the years before the financial crisis, bank risk came under greater scrutiny by regulators, reinvigorating the debate among policymakers and academics regarding best practices in bank risk management and governance. In the banking industry, a crucial role in managing the relationship between risk and governance is assigned to the board of directors. As stated in the corporate governance principles for banks introduced by the Basel Committee on Banking Supervision (2015), \"The board is responsible for overseeing a strong risk governance framework, including review of key policies and controls, and should be active when it comes to defining the risk appetite and ensuring alignment thereof within the bank. It should ensure the efficacy of the risk management, compliance and internal audit functions\". Since the Basel Committee introduced the prudential capital framework in 1988, the literature on this topic has grown rapidly and in different directions (Anderson and Fraser, 2000; Boyd and De Nicolo, 2005; Jimenez et al., 2014; Saunders et al., 1990). More recently, the literature has emphasized the critical role of good corporate governance in banking and revealed how existing regulatory failures could severely impair the stability of the financial system. To reduce those loopholes revealed by the 2007-2008 crisis, regulators at the global level have revised their corporate governance standards in areas related to the board of directors and management compensation. The responses of banks to these initiatives have differed. The most divergence in these reactions is evident between joint-stock banks and cooperative banks. Joint-stock banks and their governance were at the center of the financial crisis, and they are therefore more willing to comply with new standards, whereas cooperative banks argued that they performed better in terms of having lower volatility and more stable returns, as they took less risk during the crisis than joint-stock banks did because of their specific corporate governance characteristics (European Association of Cooperative Banks [EACB], 2015).\n2. An overview of the Italian banking system: Italy is a bank-oriented system in which banks are the key providers of loans to non-financial companies and the main collector of household savings. On the one hand, credit provided by banks is almost of 166 per cent of Italy's GDP, which is relatively higher compared to Germany (127 per cent) and France (157 per cent). On the other hand, Italian banks have deposits amounting to almost 69 per cent of total liabilities, which is higher than the levels for French (54 per cent) and German banks (62 per cent). Overall, banks account for almost 85 per cent of Italy's financial sector, and their total assets represent approximately 220 per cent of Italian GDP.\n3. Related literature and hypotheses development: 3.1 Risk propensity and bank institutional settings\n4. Research design: sample, variable and estimation framework: 4.1 Sample and data collection\n5. Results: In Table V, we present the results of our mediation analyses. We measure bank risk with the Z-index, s(ROA) and the NPL/gross loans ratio.\n6. Discussion and conclusions: Although bank governance is a subject of wide debate in the literature, to the authors' knowledge, no empirical study to date has focused on the relationship between bank institutional setting (by distinguishing between cooperative and joint-stock banks) and risk-taking behavior by exploring the underlying board-level mechanisms (education and turnover). In this paper, we analyze how board characteristics in terms of education and turnover affect decision-making about risk.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to analyze the relationship between bank institutional setting and risk-taking by exploring whether board education and turnover are drivers of the risk propensity of cooperative banks compared to joint-stock banks.\nDesign/methodology/approach: Based on a comprehensive data set of Italian banks over the 2011-2017 period, this paper examines whether these board characteristics affect the risk propensity of cooperative and joint-stock banks. Bank risk is measured by the Z-index, profit volatility and the ratio of non-performing loans to total gross loans.\nFindings: The findings show that cooperatives take less risk than joint-stock banks and have lower board turnover and education. Furthermore, this study finds that while board education mediates the relationship between the cooperative model and bank risk-taking, there is no evidence for board turnover. Thus, the lower educational level of cooperative directors contributes to explaining the lower risk-taking of cooperative banks.\nImplications: The findings have several implications. In terms of the more general policy debate, the results point to the need to strengthen the governance model for both joint-stock and cooperative banks while supporting the view that a more ad hoc perspective on the best models and practices for each type of institutional setting would be preferable. In particular, the study reveals how board education's effects on bank risk-taking should be carefully monitored.\nOriginality/value: Through a mediation framework, this study provides empirical evidence on the relationship between bank institutional setting (by distinguishing between cooperative and joint-stock banks) and risk-taking behavior by exploring the underlying mechanisms at the board level, which is novel in the literature.\n",
"role": "assistant"
}
] |
Corporate governance practices and comprehensive income
|
[
"Financial performance",
"Corporate governance",
"Board of directors",
"Financial reporting"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Research interest in the concept of \"corporate governance\" has grown considerably over the past decade (Calderon and Villabon, 2017), particularly as a result of recent economic crises. Some studies suggest that high levels of corporate governance may reduce managers' earnings manipulations and tendency to commit fraud; in addition, it may achieve a higher quality of information transparency (Prior et al., 2008; Scholtens and Kang, 2012; Shi et al., 2017). Incidents of frauds are related to the boards of directors' decisions, as boards of directors are seen as the instrument used by shareholders to monitor the behavior of top managers (Fama and Jensen, 1983). However, actions intended to improve corporate governance have information-processing and opportunity costs for firms; the questions then become, do these strategies have a positive impact on companies' results, and do corporate governance strategies create value for companies? If the answers to these questions are \"yes\", managers would invest in corporate governance, because this strategy will impact positively on firms' future performance (Bozec and Dia, 2015).\n2. Corporate governance and firm performance: From the agency theory perspective, guidelines for the relationships between managers and owners are warranted; they should protect shareholders' interests through implementing codes of good governance. The agency theory suggests that a better-governed firm should demonstrate superior performance due to lower agency costs (Rani et al., 2014; Sami et al., 2011). However, there is no consensus in the existing literature with regard to how corporate governance or performance outcomes should be measured. This topic is a central issue confronting corporate governance scholars, and some areas of research into corporate governance still draw intense interest (Bozec et al., 2010; Kumar and Zattoni, 2013). Some previous empirical studies, such as those of Gompers et al. (2003), Brown and Caylor (2006, 2009) and Dittmar and Mahrt-Smith (2007) and others, support the link between corporate governance and firm performance, although some papers show conflicting results (Bozec and Dia, 2015). This could be due to the measures of both variables.\n3. The study: 3.1 Sample\n4. Results and discussion: 4.1 Descriptive statistics and correlation analysis\n5. Conclusions and practical implications: The objective of this paper was to analyze the effect of corporate governance strategies on a measure of firms' financial performance, the CI. The previous literature has not conclusively determined the effects of corporate governance on financial performance, but, in the majority of cases, the results are significant and positive. Our corporate governance index is more concrete and precise than others because we selected the most intrinsic corporate governance variables and excluded external variables, such as auditing committees, ownership structure and others, that do not have any relationship to firms governed by board. The process by which financial performance is measured is also more robust because it has an impact on value creation.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze the effect of corporate governance practices on firms' financial performance, as measured by comprehensive income (CI).\nDesign/methodology/approach: Using a sample of 237 firms from the Standards & Poor (S&P) 500 index during the years 2004-2009, multivariate statistical analyses are conducted to confirm the authors' main hypothesis.\nFindings: The results indicate that having high levels of corporate governance culture has a positive impact on a measure of firms' financial performance, namely, CI. Furthermore, they indicate a positive correlation between a higher percentage of external directors and financial performance, and a negative relationship between number of board meetings and financial performance.\nOriginality/value: The main contribution of this research is that good corporate governance strategies deliver superior financial performance for businesses in terms of CI. This serves as a method of value creation, which is the ultimate goal of a business. In addition to the use of CI as an indicator of financial performance, a unique measure of corporate governance level is tested.\n",
"role": "assistant"
}
] |
Managing the knowledge paradox in product development
|
[
"Design and development",
"Knowledge management",
"Social action",
"Product development"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Today, there is no doubt that knowledge is one of the most strategic weapons that can lead to achieve competitive success (Grant, 1996; Spender, 1996). Successful organizations are characterized by their ability to consistently create new knowledge, share it, and employ it in the development of new products and services (Ramesh and Tiwana, 1999). Actually, product development is a knowledge intensive work (Clark and Fujimoto, 1991; Hong et al., 2005) aimed to create new knowledge or recombine existing knowledge in order to innovate and match with the changing market conditions (Katila and Ahuja, 2002). The development of a steady stream of new products is often seen as the only way to ensure survival and success (Mallick and Schroeder, 2005). It is thus fundamental for the continual prosperity of the firm (Fliess and Becker, 2006; Handfield and Nichols, 2002).Developing highly successful new products involve a tension between the old and the new - a struggle between the comfort of the past and the uncertainty of the future needs-. This tension is also known as the need to reconcile knowledge exploitation and knowledge exploration (March, 1991). Accordingly, it has been argued that successful product development demands firms to exploit their existing competencies while at the same time avoiding their dysfunctional rigidity effects by renewing and replacing these competences with new ones (Leonard Barton, 1995; Knott, 2002; Atuahene-Gima, 2005). Product development efforts need thus to improve the conflicting demands of knowledge exploration and knowledge exploitation (Adams et al., 1998; Becker and Zirpoli, 2003).The product development literature has tried to analyse how organizational characteristics and/or market orientation may help firms to face the paradox implicit in managing both exploration and exploitation (Leonard Barton, 1995; Danneels, 2002; Kyriakopoulous and Moorman, 2004; Atuahene-Gima, 2005). The activities associated with this effort are often identified as knowledge management. Knowledge management requires the introduction of criteria to decide which knowledge is most critical for the organization and to govern the factors and conditions that guide the activities of knowledge creation, integration and use (DiBella and Nevis, 1998). Since knowledge management (KM) emerges as a discipline for studying what needs to be done to get the most of organizational knowledge, linking product development and knowledge management becomes a necessity (Fedor et al., 2003; Madhavan and Grover, 1998). That means that increasing the firm's competence to derive competitive value from knowledge requires the understanding of those aspects that affect product development's potential to acquire, create, and apply knowledge.Although the importance of the association, it is surprising to find relatively few empirical investigations focusing on the position of KM in product development. Accordingly, this paper focuses on a knowledge paradox: the need to achieve exploration and exploitation in the context of product development. In order to reach this goal, it is examined the different KM practices that can lead to harmony and balance in their conflicting demands. The study thus develops a paradoxical view of knowledge management as a driver to promote the balance of exploration and exploitation by remarking a dual role that should go from the transaction, storing, and processing of knowledge to the facilitation of people networks, coordinated interface, and collaboration (Van den Brink, 2003).In order to execute the analysis, the authors first give entrance to knowledge-based capabilities in product development and then discuss their relation to dual KM initiatives: cultural and structural. Next, the authors test the proposed hypothesis on the basis of data generated from a questionnaire survey accomplished in a sample of product development efforts. Such test can give a snapshot of where differences exist and how the paradoxical view of knowledge management can promote the balance of exploration and exploitation activities. A discussion of the implications, limitations, and future research directions concludes this research paper.\n2. Theoretical foundations and hypotheses: 2.1. Product development and the knowledge paradox\n3. Research methodology: 3.1. Sample characteristics and data collection\n4. Analysis and results: Data analysis has involved several steps. First, since the research variables are measured through multiple-item constructs, the authors need to verify that items tapped into their stipulated construct. Thus, the authors conducted two independent factorial analyses by using SPSS 13.0 for Windows: one for variables corresponding to KM dimensions and other one for variables of knowledge activities in product development. Table II summarizes results. The internal consistency measures (Cronbach's alpha) were obtained in order to assess the reliability of the measurement instruments.Second, the authors applied a cluster analysis to the KM dimensions. This cluster analysis leads the authors to define different KM configurations in terms of the cultural and structural dimensions. A major issue of the clustering technique is determining the number of clusters. In this case, the authors have applied a Ward's hierarchical method using the Euclidean distance and an agglomeration schedule to determine the number of clusters and the initial seeds (centres of the groups) that have been used in a second K-means no hierarchical analysis that has provided the final categorization of the firms. The decision on the number of clusters was guided by an agglomeration coefficient, which displayed the squared Euclidean distance between each case or group of cases (see Table III). Result shows clear differences between three of the clusters.The characterization of the three final clusters based on the final centres is as follows. Cluster 1, including 23 product developments with low structural and high cultural KM dimensions, represents a cultural KM configuration. Cluster 2, comprising 44 product developments characterized by high cultural and structural KM dimensions represents a balanced KM configuration. Finally, cluster 3, formed only by 15 firms, differs from the other two groups in its high structural KM dimension and low cultural KM dimension, so represents a structural KM configuration. Table IV shows these results and the non-existence of product development with low emphasis on both cultural and structural KM dimensions. So, KM initiatives, whether cultural or structural, seems to have a role to play in product development.Next, the relationship between the KM configurations and knowledge activities, exploration and exploitation, is analyzed within clusters/configurations. Table V shows descriptive statistics (mean and deviation values) and ANOVA test for the segmented configurations. As indicated by the ANOVA test, it is possible to observe that significant differences in knowledge exploitation as a result of variations in KM configurations (p< 0.05). On the other hand, differences among clusters in terms of knowledge exploration are also significant, but at a 10 percent level (p < 0.10). Therefore, results provide partial significant evidence about the differences that may exist on the levels of knowledge exploration and knowledge exploitation as a result of the differences existing in KM configurations. The hypothesis H1 is thus partially supported.Tukey and Duncan tests were also used for further analysis. Results in Table V reveal that differences in terms of exploitation are especially significant between clusters 1-2 and clusters 2-3, while clusters 1 and 3 can be considered homogeneous in this respect. This means that the balanced KM configuration, this is, the combination of both the structural and cultural dimension of KM, is the one with the best potential to retrieve and use existing knowledge assets in order to innovate. In terms of knowledge exploration, differences between the cultural and the balanced KM configurations are significant at a 10 percent level (p < 0.10), which give an idea of the significant impact of the addition of structural aspects to the cultural aspects of KM. It is also possible to observe that cluster 2, which is the one with a balanced KM configuration, is the one that exhibits the best levels in terms of both knowledge exploration and knowledge exploitation. As previously said, both cluster 1 and two and clusters 2 and 3 significantly differs in terms of knowledge exploitation. In terms of knowledge exploration, differences are moderately significant between clusters 2 and 1. According to these findings, it is thus possible to consider that the hypothesis H2 is also partially supported.As shown in Figure 1, although results do not provide main differences in terms of exploitation between cultural KM configurations and structural KM configurations in product development, the cultural KM configuration significantly overcomes the structural KM configuration. In terms of knowledge exploration, the structural KM configuration is the one that prevails over the cultural KM configuration. Finally, it is the balanced KM configuration the one that presents the highest levels in both knowledge exploration and exploitation, and thus the best to address the paradox of exploration and exploitation.\n5. Conclusions: The advantages of KM as a major issue for knowledge work in product development have begun to receive increasing attention in recent years. Several studies have examined the relationships between KM and product development capabilities, but most of them have provided findings that tend to be either mixed or exploratory. This study provides additional insights about the benefits provided by KM in managing the paradox of knowledge exploration and knowledge exploitation during product development. Specifically, the authors analyse the link between different configurations of KM initiatives and knowledge exploration and knowledge exploitation in product development. The results offer only partial evidence to the main hypothesis, which propose that differences in KM configurations, in terms of dual dimensions, structural and cultural, may lead to differences in the knowledge activities associated to product development, in terms of the intensity of knowledge exploration and knowledge exploitation. The study has found support to the fact that a balanced KM configuration, which properly combines both cultural and structural KM enablers, works the best for knowledge activities in product development. This result is particularly significant for knowledge exploitation. This confirms the duality of KM in product development, in such a way that KM initiatives must properly align both structural and cultural enablers so as to create new knowledge and apply it in concert with existing knowledge during product development processes.Specially, a balanced KM configuration, where structural and cultural dimensions complement each other, is a key instrument for the exploitation of knowledge, and thus for the articulation, codification, storage, systematisation, diffusion, and retrieval of existing knowledge. This is particularly relevant in product development, when the need for quickly replicating and reapplying workable solutions in the form of standards and procedures is critical. Conversely, the study shows a moderately significant relationship between KM configurations and knowledge exploration in product development. As argued by Katila and Ahuja (2002), it is necessary to have in mind that firms able to exploit their knowledge will reach fine-tuning and efficiency but, in some degree, they also reach the creation of new knowledge by recombining the existing knowledge to generate new solutions. Nevertheless, it seems that a more careful analysis of how KM affects the exploration of knowledge in product development must be carried out to better understand how KM initiatives may be conductive to knowledge-based capabilities. This study has just developed a first estimate that clearly suggests that product development managers seeking to manage the paradox of exploration and exploitation must establish effective KM efforts aligning both structural and cultural enablers.Results must be viewed in light of the study's limitations. First, sample size is not large. As a second limitation, it is necessary to say that the authors have tried to define the constructs as precisely as possible by drawing on relevant literature and to closely link the measures to the theoretical underpinnings through a careful process of item generation and refinement. Evidently, this measurement effort represents an advance for research but, nonetheless, the research items are far from being perfect as long as they measure facts that are neither fully nor easily measurable. Another limitation concerns the fact that although the findings may help to explain certain relationships between variables, the authors are aware that replies from multiple respondents would have ruled out potential drawbacks. It is also necessary to keep in mind that both the external environment and the organization's internal characteristics naturally interferes with product development efforts, therefore amplifying or attenuating the organization's tendency to explore and/or exploit. Finally, sample limitations in terms of size and national characteristics might have affected the findings, and specially their generalization to other context. This work is thus obviously only a preliminary step towards a better understanding of the impact of KM on managing knowledge paradoxes in product development and, on the basis of previous limitations, it naturally points out avenues for future research.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to obtain from paradox a novel lens to elucidate the connections between knowledge management (KM) and product development as a knowledge intensive process. By focusing on the \"social side\" of KM, it is proposed that different KM orientations can emerge as a result of higher or lower emphasis on both cultural and structural enablers. These KM orientations generate specific potential to manage the tension and paradox of managing both knowledge exploration and knowledge exploitation knowledge during product development.\nDesign/methodology/approach: - Empirical evidence is provided by performing survey research with data collected from 80 product development projects developed in Spain.\nFindings: - Results show that product development efforts can show three different KM orientations that significantly differ in their results in terms of knowledge exploration and, specially, knowledge exploitation. Moreover, product development requires a symbiosis between both cultural and structural enablers to best manage the paradox of exploration and exploitation.\nOriginality/value: - The measurement items used can realistically be thought of as only proxies for an underlying and latent phenomenon that is neither fully nor easily measurable. Together with it, replies from multiple respondents would have ruled out potential drawbacks.\n",
"role": "assistant"
}
] |
Personality and gender differences in revealed risk preference: evidence from Ghana
|
[
"Gender",
"Risk",
"Aversion",
"Binary logit"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In making decision under uncertainty, there is a common belief that, compared to men, women are less likely to take risky decision (i.e. women are more risk averse compared to men). This common belief or stereotype is supported by a large body of literature on gender differences in preferences (see the survey by Croson and Uri Gneezy, 2009). Although prior studies on gender differences regarding risk-taking behavior provide valuable insights to explain differences in decision making across gender types, a critical weakness in the existing empirical studies is that they clandestinely consider men or women as homogeneous groups without examining the intra-gender differences. Meanwhile, there are intra-gender differences relating to personality that may contribute to the understanding on human behavior regarding decision making under uncertainty. In epidemiological studies, Graves and Mortimer (1994) suggested that personality is an important factor that accounts for the association between risk-taking behaviors and Parkinson's disease. Nicholson et al. (2005) also suggested that risk propensity reflects a combination of personality factors. In recent times, a growing body of economic literature has examined the role of personality traits in predicting or understanding individual decision making, productivity, earnings and life outcomes (Bucciol and Zarri, 2017; Caliendo et al., 2016; Rustichini et al., 2016; Brown and Taylor, 2014; Fletcher, 2013). Bucciol and Zarri (2017), for example, found that personality traits such as agreeableness and neuroticism have significant negative correlation with financial risk taking.\nLiterature review: Theoretical foundations\nMethodology: Data\nResults: Socio-psycho-demographic profile of respondents\nDiscussion and implications: The findings from this study are related to extant literature. As shown in this study, gender is associated with risk aversion preference in two ways. First, the descriptive results show that there are significantly more females being risk averse than males. Second, the regression output indicates that being a woman is associated with higher likelihood of being risk averse. Both results are consistent with findings of previous studies that have documented significant association between gender and risk-taking behavior (Croson and Uri Gneezy, 2009; Loewenstein et al., 2001). One of the major explanations that has been given by some of these studies, which is also plausible in this study, is the affective reaction to risk (Loewenstein et al., 2001). According to this explanation, men and women differ in their emotional reaction to uncertain situations and this differential emotional reaction results in differences in risk taking (Croson and Uri Gneezy, 2009). We found this explanation more plausible because the personality differences of females and males in this study show that neuroticism/stability, which describes the tendency to experience negative emotion and related processes in response to perceived threat and punishment, is higher for females than males.\nConclusion: In conclusion, this study has provided empirical evidence in a developing country context to establish the gender differences in risk aversion and to explain how these differences are related to the Big Five personality traits. In sum, the evidence shows that women are more risk averse than men. Gender differences in terms of the Big Five personality traits were also observed. To be precise, the personality traits of extraversion, openness, stability and agreeableness are significantly higher among females than among males. The findings are largely consistent with those of previous studies that have documented significant differences in risk aversion between males and females and the association between gender and personality traits. Beyond the descriptive results of the gender differences and personality traits, the analytical results of the logistic regression also identified gender, age and education as the demographic factors that associated with the likelihood of being risk averse. For the gender variable, being a female is associated with increased likelihood of being risk averse.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate gender differences in risk aversion and determine the effect of personality traits on risk aversion from an African country context.\nDesign/methodology/approach: The study combined both descriptive and analytical designs, and employed t-test, Pearson's kh2 and binary logistic regression as the main analytical techniques. Data for the analysis were obtained from the World Bank's Skills toward Employment and Productivity survey on Ghana.\nFindings: Results of the study revealed systematic differences in personality and gender and their associations with risk aversion. Specifically, women were found to be more risk averse than men. Differences in personality also showed that females reported higher levels of personality in all but one of the Big Five personality traits - extraversion. In addition to gender, age and education, the personality traits of conscientiousness and stability were the main predictors of the likelihood of being risk averse. Although personality differences existed between male and female, the interaction terms between gender and personality factors were not statistically significant.\nOriginality/value: The paper departs from the extant literature on developed countries and western cultures to add to the understanding on how individual differences account for variation in revealed risk preferences.\n",
"role": "assistant"
}
] |
"This class is not for you": An investigation of gendered subject construction in entrepreneurship course descriptions
|
[
"Gender",
"Higher education",
"Entrepreneurship education",
"Course description",
"Course type"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper investigates the social construction of gendered subjects in entrepreneurship education (EEd), through an analysis of course descriptions from higher education institutions (HEIs). This focus is important in light of emerging research on the impact of EEd on women (Gupta et al., 2009) which suggests that they may gain more positive benefits from EEd than men (Bae et al., 2014; Terjesen et al., 2016). Given that those who are highly educated are more likely to start businesses (Schott, 2009), that more women than ever are pursuing higher education (HE) (Eurostat, 2016), and that female students have higher entrepreneurial intentions than male students (Haus et al., 2013), one would expect more women than men to participate in entrepreneurship courses. Yet, this does not appear to be the case (Menzies and Tatroff, 2006; Petridou et al., 2009).\nThe interplay of gender and language: For feminist and other critical researchers, gender refers to socially produced distinctions between men and women (Acker, 1990; Connell, 2009). The term sex represents how people are categorized as male and female, while the term gender represents the meanings that society places on such categorization (Malach-Pines and Schwartz, 2008). Although conceptually this study separates biological sex from gender, we acknowledge that cultural and social beliefs \"cannot be separated from biological 'knowledge'. The meaning associated with the two gender categories (male and female) unavoidably clouds every aspect of thought, perception and behaviour\" (Weatherall, 2002, p. 81).\nThe gendered construction of fictive subjects: the Fictive Student and the Fictive Entrepreneur: Bourdieu and Passeron (1994, p. 17) originally suggested the concept of the Fictive Student and emphasized the interplay of objective and subjective positions, suggesting that the position of the student \"compels them to enter into the game of fictive communication.\" Windle (2009, p. 94) argued that fictive subject constructions are often subconscious on the part of both educators and students, and that the suggested norms alienate real students who do not fit the socially narrow mould of the ideal Fictive Student. Nadesan and Trethewey (2000, p. 228) argued that the \"entrepreneurial subject [...] implicitly emerges as normatively masculine in character.\" However, Bourdieu and Passeron's notion of the Fictive Student is gender neutral, so this paper extends the concept to consider how the Fictive Student may be gendered. It further extends the notion of fictive subjects by analyzing how HEIs and educators also construct the Fictive Entrepreneur within course descriptions.\nEEd courses and course descriptions: In 2005, Hannon summarized the development in EEd, stating that the commonly applied conceptualization of courses is \"about,\" \"for,\" and \"through,\" which is still a widely accepted typology (Warhuus and Basaiawmoit, 2014). About-type courses study entrepreneurship theoretically as a social phenomenon (Laukkanen, 2000). For-type courses are more applied, and prepare students for \"what to do and how to make it happen\" (Laukkanen, 2000, p. 26), and typically focus on the skills required to start new and manage small businesses (O'Connor, 2013). Through-type courses focus on personal involvement and experiential learning through participation in entrepreneurial activities (Heinonen and Hytti, 2010; Thrane et al., 2016) and encourage students to feel the life-world of an entrepreneur (Gibb, 2011; Neck and Greene, 2011). Different types of courses tend to be taught and assessed differently and have different types of teachers (Levie, 1999; Pittaway and Edwards, 2012). It is also likely that these different types of courses may attract different students disproportionally and incentivize students to become entrepreneurs to varying degrees. For example, courses about entrepreneurship carry few expectations that students will become entrepreneurs.\nMethodology: This paper investigates gender and entrepreneurship at a very early stage in the EEd pipeline, before students even enter the classroom. Guided by the three objectives above, course descriptions, the unit of analysis, are defined as any text made publicly available by the university about a specific course, typically provided through online course catalogues.\nFindings: The manifest analysis (see above) yielded a 1 to 9 ratio of feminine to masculine words. This indicates that masculine language dominates and that course descriptions do not deviate from mainstream conceptualizations of entrepreneurship and are not sensitive to ongoing critiques of the masculinization of the field. However, as Brine's process indicates, this is a high-level and inadequate analysis, being only step 5 of her 10-step analytical process (see the appendix). The next step, latent analysis (see above) highlighted that two out of three codes were masculine. These data are as shown in Table I.\nDiscussion: This section discusses the findings in relation to the paper's three main research objectives as presented in the \"Introduction.\" Regarding the use of gendered language, feminine language was less prevalent and relatively more concentrated than masculine language. As institutions lean heavily on popularized descriptions of entrepreneurship, the course description may play a role in perpetuating the barriers and gendered lack of legitimacy identified by Haus et al. (2013), Jones (2015), and Marlow and McAdam (2013).\nConcluding thoughts: Critical and feminist research highlights and analyses entrepreneurship as a masculinized phenomenon and these debates are being brought into the broader entrepreneurship literature (Ahl and Marlow, 2012). However, this is also an important consideration in educational contexts that position entrepreneurship as meritocratic and open to all (Jones, 2015; Siivonen et al., 2016), and where the EEd student is positioned as homogenous (Jones, 2014; Rae et al., 2012). Consequently, this paper asks new questions, from a constructionist feminist perspective, about the degree to which course descriptions use gendered language, how such language constructs fictive subjects, and its implications. Its contribution is fourfold.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the social construction of gendered subjects in entrepreneurship education (EEd), through the analysis of course descriptions. For this purpose, the analytical constructs of the Fictive Student and the Fictive Entrepreneur are developed.\nDesign/methodology/approach: Through analysis of 86 course descriptions from 81 universities in 21 countries, this study examines the degree to which course descriptions use gendered language, how such language constructs gendered subjects, and the resultant implications.\nFindings: This paper finds that course descriptions are predominantly, but not exclusively, masculine in their language. More importantly, the distribution of feminine and masculine language is uneven across course descriptions. Context variables such as regional or national culture differences do not explain this distribution. Instead, the phenomenon is explained by course content/type; whereby practice-based entrepreneurship courses are highly masculine, compared to traditional academic courses, where students learn about entrepreneurship as a social phenomenon.\nPractical implications: Universities and educators have not taken into account recent research about the real and possible negative consequences of positioning entrepreneurship in a stereotypical, masculinized fashion. This may offer an inexpensive opportunity to improve recruitment and description accuracy.\nOriginality/value: The paper's contribution is fourfold. First, it contributes to debates on the gendering of entrepreneurship by extending these into EEd. Second, it extends Sarasvathy's (2004) concern with barriers to, rather than incentives for, entrepreneurship to include EEd. Third, it contributes to the emerging literature on entrepreneurship as practice, by highlighting the masculization of EEd, as it gets closer to practice and the role of language in this. Finally, it highlights the gendered implications of English medium courses.\n",
"role": "assistant"
}
] |
The optimal system for series systems with mixed standby components
|
[
"Reliability management",
"Cost benefit analysis"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Redundancy, repair maintenance, and preventive maintenance are some of the well-known methods by which the reliability of a system can be improved. Two-unit standby redundant systems have been extensively studied by several authors in the past. Achieving high reliability and/or availability is often an essential requisite. In this paper, we consider the manufacturing system or the electric power systems (power plant) to be a series systems with mixed standby (include cold standby, warm standby and hot standby) components. A standby component is called a \"hot standby\" if its failure rate is as primary component. The standby component is referred as \"warm standby\" when the failure rate is not equal zero and is less than the failure rate of a primary or hot standby components. A standby component is called a \"cold standby\" if its failure rate is zero. Primary, hot standby and warm standby components can be considered to be repairable.We study four different system configurations of series and mixed standby components. These four configurations are compared based on their mean time-to-failure, MTTF, their steady-state availability, A([?]), and their cost/benefit ratio C/B. Cost is considered to be a size-proportional cost for the components. Benefit is divided into two categories according to whether the measure used is the system reliability given by MTTF or the system availability given by A([?]).Therefore, Wang and Sivazlian (1997) studied the two different configurations with parallel components are compared based on their overall availability and life cycle costs under uncertainty in systems lifetime. The time-to-repair and the time-to-failure for each of the primary and parallel components are assumed to have the negative exponential distribution. Wang and Pearn (2003) analysed the series systems with cold standby components and warm standby components, respectively, where the repair time distribution of the server is assumed to be exponentially distributed. Wang and Kuo (2000) considered the reliability, the availability, and the cost/benefit analysis of four different series system configurations with mixed standby (include cold standby and warm standby) components. El-Said and El-Sherbeny (2006) studied two systems, each system with two parallel components. The second system differs from the first system due to the additional feature of preventive maintenance. The two systems are analysed under the assumption that the failure, replacement and preventive maintenance times of the units are assumed to be arbitrarily distributed. The stochastic analysis of a non-identical two-unit parallel system with common-cause failure by graphical evaluation and review technique (GERT) considered by Sridharan and Kalyani (2002). Chandrasekhar et al. (2004) studied the two unit standby system and obtain exact confidence limits for the steady-state availability of the system, when the failure rate of an operative unit is constant and the repair time of the failed unit is a two stage Erlang distribution. The stochastic analysis of two-unit outdoor electric power systems in changing weather considered by Dhillon and Rayapati (1985). The purpose of this paper is accomplishes three objectives. The first one is the use of Laplace transform techniques to derive the explicit expressions for the mean time-to-failure, MTTFi, and the steady-state availability, Ai([?]) for configuration i (i = 1,2,3,4) . The second one is to compare these configurations in terms of their MTTFi and their Ai([?]). The third one is to compare the four configurations with their cost/benefit (Ci/Bi) based on assumed numerical values given to the distribution parameters, as well as to the costs for the components.\n2. Description of the problem: The present paper is devoted to consider the requirements of a 10 MW power plant. We assume that generators are available in units of both 10 MW and 5 MW. Standby generators are always necessary in case of failure. Let us assume that all switchover time are instantaneous and switching is perfect, e.g. never fails and never does any damage. We also assume that the switchover time from hot standby component to operative component, from warm standby component to hot standby component, from cold standby component to warm standby component and from failure to repair are instantaneous. Operative components, hot standby components and warm standby components can be considered to be repairable. Each of the operative components fails independently of the state of the others and has an exponential time to failure distribution with parameter l. Whenever one of operative fails, a hot standby moves into operation if any is available, a warm standby is put on hot standby, and a cold standby is put on warm standby state if any is available. We also assume that of a hot standby moves into a operative component state, its failure characteristics will be that of a operative component, when a warm standby moves into a hot standby state, its failure characteristics will be that of a hot standby, and when a cold standby moves into a warm standby state, its failure characteristics will be that of a warm standby. We assume that each of the available hot standby components fails independently of the state of all the others and has an exponential time to failure distribution with parameter l. We also assume that each of the available warm standby components fails independently of the state of all the others and has an exponential time to failure distribution with parameter a where a[?] (0, l). Whenever a operative component or a standby components fails, it is immediately repaired in the order of breakdowns with a time-to-repair which is exponentially distributed with parameter m. Once a component is repaired, it is as good as new.The following configurations are considered. The first configuration is a serial system of one operative 10 MW component, one hot standby 10 MW component, one warm standby 10 MW component, and one cold standby 10 MW component. The second configuration is a serial system of two operative 5 MW components, one hot standby 5 MW component, one warm standby 5 MW component, and one cold standby 5 MW component. The third configuration is a serial system of one operative 10 MW component, two hot standby 10 MW components, one warm standby 10 MW component, and one cold standby 10 MW component. The last configuration is a serial system of one operative 10 MW component, one hot standby 10 MW components, two warm standby 10 MW components, and one cold standby 10 MW component.2.1 Cost-benefit factor\n3. Problem solutions: 3.1 Calculations for configuration 1\n4. Comparison of the four configurations: The purpose of this section is to present specific comparisons for the MTTFi and the Ai([?]). Using an efficient Mathematica and Matlab computer program, all configurations will be compared in terms of their MTTFi and Ai([?]) [?]i = 1,2, 3, 4 with the following values: Equation 81 4.1 Comparison for the MTTFi and Ai([?])\n5. Conclusions: In this paper, we studied the mean time to system failure and the steady-state availability with minimum frequently maintenance of four different series system configurations with mixed standby components. By comparing the MTTFi and Ai([?]) listed in Tables II and III. Numerical results for the cost/benefit measure have been obtained for all configurations as in Table IV. It is interesting to note first that the optimal configuration using the cost/MTTFi measure is configuration 4. Next the optimal configuration using the cost/Ai([?]) measure is configuration 2.\n",
"role": "user"
},
{
"content": "Purpose: - The main objective of this paper is to study the optimal system for series systems with mixed standby (including cold standby, warm standby and hot standby) components.\nDesign/methodology/approach: - The paper deals with the reliability and availability characteristics of four different series system configurations. The failure time of the operative, hot standby and warm standby are assumed to be exponentially distributed with parameters l, l, and a respectively. The repair time distribution of each server is also exponentially distributed with parameter m.\nFindings: - The mean time to failure, MTTFi, and the steady-state availability Ai([?]) for four configurations are examined and comparisons made. For all four configurations, the configurations are ranked based on: MTTFi, Ai([?]), and Ci/Bi where Bi is either MTTFi or Ai([?]). Obviously, the system with height MTTFi and Ai([?]), do not need frequent maintenance, i.e. less maintenance.\nOriginality/value: - Numerical results for the cost/benefit measure have been obtained for all configurations. It is interesting to note first that the optimal configuration using the cost/MTTFi measure is configuration 4. Next the optimal configuration using the cost/Ai([?]) measure is configuration 2.\n",
"role": "assistant"
}
] |
Exploring the mediating role of social capital in the relationship between financial intermediation and financial inclusion in rural Uganda
|
[
"Financial inclusion",
"Social capital",
"Structural equation modelling",
"Mediation effect",
"Financial intermediation",
"Rural Uganda"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Development agencies such as the World Bank, G20, Bill & Melinda Gates foundation, and UNDP among others, suggest that in order to end poverty as stipulated under the new sustainable development goals agenda, there is a great need for full financial inclusion, especially in developing countries.\nLiterature review and hypotheses development: Financial intermediation and financial inclusion: mediating role of social capital\nResearch method: Study design, population, and sample\nResults and discussion: For the purpose of understanding the characteristics of the different households' heads who participated in the study, they were categorized based on age, gender, members in households, and ability to read and write as shown in Tables I and II.\nConclusion: The current study was carried out to establish the mediating effect of social capital in the relationship between financial intermediation and financial inclusion in rural Uganda.\nPolicy implications: Thus, from the study findings, a number of policy and research issues can be taken into consideration in order to address the findings of the study.\nLimitation of the study: The sample used in this study although large enough, may not be generalized to other segments of the population. Data were collected from only poor households located in rural Uganda. Future studies may be conducted using samples selected from other vulnerable groups such as the disabled persons, women, youth, and refugees. Besides, the study was cross-sectional, thus, limiting efforts in investigating certain characteristics of the sample over time. Perhaps, future studies could adopt the use of longitudinal research design.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to establish the mediating role of social capital in the relationship between financial intermediation and financial inclusion in rural Uganda.\nDesign/methodology/approach: The current study used cross-sectional research design and a semi-structured questionnaire was used to collect data for this study. The study applied structural equation modeling through bootstrap approach in AMOS to establish the mediating role of social capital in the relationship between financial intermediation and financial inclusion.\nFindings: The results indicated that social capital significantly mediates the relationship between financial intermediation and financial inclusion in rural Uganda. Therefore, it can be deduced that social capital among the poor play an important role in promoting financial intermediation for improved financial inclusion in rural Uganda.\nResearch limitations/implications: Although the sample was large, it may not be generalized to other segments of the population. Data were collected from only poor households located in rural Uganda. Besides, the study was cross-sectional, thus, limiting efforts in investigating certain characteristics of the sample over time. Perhaps future studies could adopt the use of longitudinal research design.\nPractical implications: Financial institutions such as banks should rely on social capital as a substitute for physical collateral in order to promote financial inclusion, especially among the poor in rural Uganda.\nOriginality/value: This study provides empirical evidence on phenomenon not studied in rural areas in Sub-Saharan Africa where the poor use social capital embedded in customs and norms for doing business. The results highlight the importance of social capital in mediating the relationship between financial intermediation and financial inclusion of the poor in rural Uganda.\n",
"role": "assistant"
}
] |
Be flexible: turning innovativeness into competitive advantage in hospitality firms
|
[
"Competitive advantage",
"Process innovation",
"Product innovation",
"Innovation capability",
"Organizational flexibility",
"Hospitality firms"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Adaptation has become an important firm-level characteristic necessary for helping companies deal with a complex and dynamic marketplace. Today's fast-paced business environment, facilitated by evolving customer needs and technology, requires firms to adapt to its changing dynamics to succeed. Creating flexible organizational structures and systems that allow firms to adapt to environmental dynamics has become a necessary condition for increased control. Hence, in this study, adaptable firms are considered flexible and have flexibility capabilities (Gibson and Birkinshaw, 2004). Flexibility allows the organization to, as Tomlinson (1976) puts it:\n2. Theoretical background and hypotheses: This study relies on the dynamic capability theory to explain how two firm-level capabilities (i.e. flexibility and innovation) create competitive advantages for hospitality firms. Inherent in firms' routines are capabilities and resources that enable them to function and compete. When capabilities are able to integrate, reconfigure, renew and recreate resources, they are then considered dynamic capabilities (Teece et al., 1997). Dynamic capabilities further upgrade and reconstruct core competencies of the firm to adapt to the changing environment to attain and sustain competitive advantage. Innovativeness and OF are considered firm-level capabilities that have the potential to create and sustain competitive advantages. For example, innovation capabilities allow the firm to generate new value to meet the changing needs of the market (Calantone et al., 2002). That means, innovativeness is able to create the needed external flexibility for the firm to meet changing needs of the market. Furthermore, flexibility allows the firm to balance internal operational efficiency and external exploration of market opportunities to be competitive. OF is also seen as type of dynamic capability that has the capacity to facilitate the functioning of other capabilities, such as innovation, to create competitive advantage. We use product and process innovativeness as internal and external focus activities which seek to meet clientele demand. OF is both internally and externally focused to ensure that the firms' actions, activities and strategies appropriately fit the external environment. Hence, when innovativeness enhances OF, the firm external outlook will be positive. Zollo and Winter (2002) described OF as a learned and stable pattern of collective activity through which the organization systematically generates and modifies its operating routines in pursuit of improved effectiveness. This improvement results from the firm's ability to integrate, build, and reconfigure internal and external competences to address environmental pressure (Teece et al., 1997). It is this improved effectiveness that yields the competitive advantage. According to Collis (1994) categorization of capabilities, hospitality firms' OF capabilities reside within the operating routines and allow the firm to perform basic functional activities and to recognize the intrinsic value of other capabilities to meet the environmental conditions better than the competition. This study theorizes that both innovativeness and flexibility are dynamic capabilities that reside in the routines of the hospitality firm, however, when innovativeness creates the needed flexibility for the firm, there is a higher chance to improve competitive advantage. However, OF acts as that kind of capability that is able to recognize the intrinsic value of innovativeness such that it is able to enhance the innovation-competitive advantage relationship.\n3. Methodology: 3.1 Sampling and data collection\n4. Measures: To assess competitive advantage, adapted the measures by Akimova (2000) for measuring the competitive advantage of firms based on the marketing mix elements. These scales have been widely used in the firm level literature and as key assessment criteria for competitiveness (Anning-Dorson, 2018). Eight items, which were measured on a seven-point Likert scale (1 = strongly disagree, 7 = strongly agree), survived the minimum cut-off point of 0.70 for item loading. To measure process and product innovation capability, we followed the works of Anning-Dorson (2018) and Anning-Dorson et al. (2019) with a seven-point Likert scale (1 = strongly disagree, 7 = strongly agree). Four and five items measured process and product innovation capabilities, respectively. The two sets of innovation in this study are measured as dynamic capabilities that reside in the routines of the hospitality firm. Innovation here is therefore measured as a necessary capability that serve as an input into the creation of competitive advantage. The study does not measure process or product in terms of output but the propensity to embark on process and product innovation activities Several studies have measured innovation capabilities in this light (Calantone et al., 2002; Cavusgil et al., 2003) where frequency new idea tries, seeking new ways of doing things, number of new product introduction within a period as used as a measure of innovativeness. The measures are therefore consistent with literature. To measure the organizational flexibility of hospitality firms, how firms manage market changes, promptly respond in a proactive manner to market threats and opportunities; and adapt to external environmental pressures was used. This was measured by six items derived from the literature (Theoharakis and Hooley, 2003; Vorhies and Morgan, 2003, 2005; Tsai and Shih, 2004; Santos-Vijande et al., 2012). Respondents were asked to rate to what extent their firms were able to rapidly respond to new market and competitive conditions compared to their major competitors (1 = hardly at all, 7 = very easily).\n5. Data analysis and results: 5.1 Common method bias assessment\n6. Discussion and implications: While organizational adaptation has been championed as an important characteristic of flexible enterprises, the extent to which this important firm characteristic helps explain how innovation capability translates into competitive advantage has not received adequate empirical support. This study sought to analyze how OF enables hospitality firms to be more mobile, responsive, and agile and, therefore, to transform innovation into competitive advantage. While the sector has been admonished that it must be more innovative to meet the changing needs of its clientele, there is still evidence that innovation efforts and outputs are not yielding the desired benefits. This study explains that this may be due to the absence of a critical firm factor in the form of OF. The study argues that the gap between what hospitality firms do via their innovation capability and the results on the market (innovation failure; see Hon and Lui, 2016) is a result of OF. The study proposed that OF determines the adaptive capacity of the hospitality firm under conditions of uncertainty in the market (Hatum and Pettigrew, 2006). Hence, while the capability to innovate (product and process) may yield some benefits, such activities do not have the flexible capacity in themselves to adequately respond to the environmental changes in the hospitality sector that in turn explain competitiveness in that sector. In this sense, this study provides a better explanation as to why the innovation failure rate is high within the hospitality sector. The findings suggest that OF is an important mediator between innovativeness and competitive advantage, which provides a practical and empirical contribution to the innovation and hospitality literature.\n",
"role": "user"
},
{
"content": "Purpose: While acknowledging the importance of innovation capability and its potential for creating competitive advantage for firms, the purpose of this paper is to further explain how firms can gain the most from this potential. In the face of the low success rate of innovations and innovation activities in the hospitality sector, this study explains - through the dynamic capability perspective - that building a flexible organization is instrumental in creating competitive advantage out of innovation capabilities.\nDesign/methodology/approach: The study uses data from hospitality firms operating in an emerging economy with a fast-growing hospitality sector. As this study uses survey and statistical methods, the possibility of common method bias is addressed by partial least squares structural equation modeling to test the hypothesized relationships.\nFindings: The study finds that organizational flexibility (OF) is an important mediator in the relationship between innovation capabilities and competitive advantage. The study explains that building a flexible organization is instrumental in creating competitive advantage out of innovation capabilities in hospitality firms.\nOriginality/value: This study suggests that relying only on innovation capability offers hospitality firms just a fraction of what would be gained if such capabilities are properly aligned with OF. This is because such alignment generates better response power to changing needs as the firm becomes more mobile, responsive and agile to rapidly identify market trends, adjust internal operations and respond quickly to new market demand.\n",
"role": "assistant"
}
] |
Predictability of HK-REITs returns using artificial neural network
|
[
"Investment",
"Artificial neural network",
"Real estate investment trust",
"Hong Kong REITs",
"Real estate investment",
"Return forecasting"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In finance literature, the predictability of stock returns has remained an important area of research. Several key variables which can influence stock returns have been identified. For instance, market risk, size risk and value risk have been observed to affect stock returns. This outcome is as stipulated by the capital asset pricing model (CAPM) (Sharpe, 1964; Lintner, 1965), and Fama and French's three-factor model (Fama and French, 1992). Previous studies (Gaunt, 2004; Soumare et al., 2013) have used the linear regression method to model the linear relationship among the variables. Due to the complex nature of the equity market, more complex techniques like the ANN have also been used to model stock returns (Cao et al., 2005; Kryzanowski et al., 1993; McNelis, 1996).\nLiterature review: The predictability of stock returns has always been a concern of researchers and practitioners. Hence, it was not surprising to note that the CAPM (Sharpe, 1964; Lintner, 1965), and Fama and French's three-factor model (Fama and French 1992) were commonly being used by researchers (Brooks and Tsolacos, 2003; Gaunt, 2004; Soumare et al., 2013; Cao et al., 2005; Kryzanowski et al., 1993; McNelis, 1996; Serrano and Hoesli, 2007; Tsai et al., 2014) to predict stock returns. The CAPM explained the asset returns through market risks, while Fama and French's (1992) model measured the relationship between asset returns with market risk, size risk and value risk.\nData: The data used in this study comprised of all the REITs which were actively traded on the Hong Kong Stock Exchange as of December 31, 2018. Each REIT was then analyzed based on a set of data which contained excess returns of the REIT, its market capitalization, its book to market ratio and the market excess returns. Table I illustrates further.\nMethodology: Two methods comprising the ANN and the regression technique were applied in this study to forecast the HK-REITs' returns. The CAPM and Fama and French three-factor model were applied. Table III demonstrates the predictor(s) information noted for each of the models employed. The REITs' excess returns served as the dependent variable of this study. They were modeled against the lag one observation of the predictor(s).\nModel evaluation: The model evaluation was used to examine the amount of excess returns generated. This was accomplished by following all the forecasting results of the forecasting method. For instance, it was assumed that investors would purchase an REIT at its closing price on a particular day if the investors anticipated a positive excess returns on the subsequent day. The investors then closed the position at the end of the following day. Following this, all the actual excess returns that were generated from each trade would be totaled up to represent the trading performance for each REIT, within a certain period of time.\nResults: The results gathered from this study are further presented under the different headings.\nConclusion: The proliferation of machine learning research has attracted the attention of researchers and practitioners to the extent of including machine learning techniques within their studies (Cao et al., 2005; Qi and Maddala, 1999; Serrano and Hoesli, 2007). The ANN method, being one of the most popular, gained its popularity due to its strong prediction power.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to determine if artificial neural network (ANN) works better than linear regression in predicting Hong Kong real estate investment trusts' (REITs) excess return.\nDesign/methodology/approach: Both ANN and the regression were applied in this study to forecast the Hong Kong REITs' (HK-REITs) return using the capital asset pricing model and Fama and French's three-factor models. Each result was further split into annual time series as a measure to investigate the consistency of the performance across time.\nFindings: ANN had produced a better forecasting results than the regression based on their trading performance. However, the forecasting performance varied across individual REITs and time periods.\nPractical implications: ANN should be considered for use when one were to attempt forecasting the HK-REITs excess returns. However, the trading performance should be always compared with buy and hold strategy prior to make any investment decisions.\nOriginality/value: This paper tested the predicting power of ANN on the HK-REITs and the consistency of its predicting power.\n",
"role": "assistant"
}
] |
Consumer engagement behaviors: do service convenience and organizational characteristics matter?
|
[
"Customer engagement behavior",
"Service convenience",
"PLSpredict",
"Word-of-mouth",
"Helping"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The dynamics of the service provider-customer relationship have evolved over time, especially since the advent of the networked society (Hollebeek and Andreassen, 2018). Today, customers are aware of competitive offerings and actively interact with service firms as well as other stakeholders (Lariviere et al., 2017). Consequently, service firms have also strategically introduced practices to engage customers, beyond revisits and positive word-of-mouth, to develop a meaningful and profitable relationship with them (Pansari and Kumar, 2017). For example, the Lego IDEAS website invites Lego supporters to submit designs for new playsets. If a design receives more than 10,000 votes, it moves into the review phase wherein Lego designers and a marketing team decide on the feasibility of the project for production. Customers also exhibit engagement behaviors on various social media platforms, for example liking or recommending a firm on Facebook, advocating a firm to their friends and/or relatives, providing feedback to the firm and helping other customers (e.g. BeautyTalk by Sephora.com) (Roy et al., 2018a, b; Verleye et al., 2014). These positive behavioral manifestations of customer engagement (hereafter, CE) directed towards a brand or a firm beyond the actual purchase are referred to as customer engagement behaviors (hereafter, CEBs). Compared to disengaged customers, engaged customers are posited to demonstrate less price sensitivity, resist switching to competitors, actively contribute to the development of new products/services and advocate firms to others. CEBs have received increased academic and practitioner attention because of their significant contribution to firm performance in terms of profitability and sustainable competitive advantage (Hollebeek et al., 2016). Given its importance, the Marketing Science Institute (MSI) lists identifying the drivers of CE as a research priority for 2018-2020 (MSI, 2018).Extant research has examined the valence as well as the dimensions of CEB and proposed antecedents of CEB (Verhoef et al., 2010; Verleye et al. 2014). Based on our literature review, we observe that the earlier CE literature was largely conceptual in nature and conducted in the social media context (see Table I for select studies on the antecedents of CEB/CE). We find that limited attention has gone into empirically investigating the impact of organizational factors on CEB (Grewal et al., 2017a, b; van Doorn et al., 2010) and have identified five major organizational characteristics that retailers (especially supermarkets) use to elicit CEBs. These include the brand name, store atmosphere, store design/layout, customer information and social environment (Verhoef et al., 2009; Berry et al., 2002). Prior studies show that these organizational characteristics help retailers enhance the customer shopping experience, which in turn can lead to customer loyalty and positive word-of-mouth (one form of CEB) (Hocutt et al., 2006; Kim and Lennon, 2008; van Rompay et al., 2012). There is limited research on the effect of these organizational characteristics on other forms of CEB (Grewal et al., 2017a, b) such as service improvement behavior, customers' cooperation and customers helping other customers.CEBs require customers to invest their discretionary time and effort. Since customers are increasingly becoming convenience-oriented due to the demands of modern-day life and its associated stresses (Elizabeth-Lloyd et al., 2014; Seiders et al., 2007), they want to judiciously invest their time and effort to improve their quality of life and overall well-being (Roy et al., 2017). We argue that the time and effort scarcity may limit the effectiveness of the organizational characteristics in eliciting CEBs. Hollebeek and Chen (2014) observe that customers' willingness to invest cognitive, emotional as well as physical resources is based on the perceived benefits derived from their interaction with firms and/or brands. Hence, service convenience is gaining traction in the CEB literature (Pansari and Kumar, 2017; van Doorn et al., 2010).On the one hand, service convenience may simply be another organizational characteristic that retailers can use to enhance CEBs. Service convenience improves the entire customer experience, which in turn, leads to a variety of CEBs. If this is the case, then retailers will need to focus on the design and management of organizational characteristics that enhance the customer experience. On the other hand, service convenience may be the prime driver of CEBs in that consumers are only able to expend their scarce time and effort if service convenience provides them the opportunity to do so. If this is the case, then retailers will need to focus on those organizational characteristics that lead to service convenience. Doing so will provide consumers with the necessary resources (i.e. time and effort) to engage in CEBs. Thus, the twofold objectives of this study are as follows:To examine the impact of organizational characteristics on service convenience; andTo examine the role of service convenience in the relationship between organizational characteristics and CEBs.We use a competing-models approach to examine the role of service convenience in the prediction of CEBs in a grocery store setting. The direct-effects model suggests that service convenience is an additional driver of CEBs alongside many other organizational characteristics (e.g. service design, layout). The mediating model suggests that service convenience is a key mediating variable between organizational characteristics and CEBs. Our results support the mediating model. The main contribution of this study lies in identifying the impact of organizational characteristics in helping customers save their time and effort related to shopping activities (i.e. provide service convenience) and thereby elicit different types of CEBs. Although service convenience has been shown to impact the outcome variables of customer satisfaction and repurchase intention, this study contributes to the literature by examining the impact of service convenience on respective CEBs. This is a welcome addition to the CEB literature in particular and services research in general.The rest of the paper is organized as follows: first, we review the literature on the key constructs in this study and propose the two competing research models and related hypotheses. Next, we detail the research methodology used, data analysis and discussion of results. Finally, we outline the theoretical contribution of this study, managerial implications, limitations, and scope for future research.\n2. Literature review and hypotheses: 2.1 Customer engagement behaviors (CEBs)\n3. Research hypotheses: We propose two competing models to examine the role of service convenience in the prediction of CEBs as follows. The direct-effects model suggests that service convenience is an additional driver of CEBs alongside many other organizational characteristics (e.g. service design, layout). This model is grounded in adaptive cost theory. The mediating model suggests that service convenience is a key mediating variable between organizational characteristics and CEBs. This model is grounded in the social exchange theory. 3.1 Research model A: the direct effects model\n4. Research methodology: 4.1 Research approach and context\n5. Data analysis and results: We considered partial least square based structural equation modeling (PLS-SEM) in our study for a number of reasons. Literature suggests that covariance-based structural equation modeling (CB-SEM) is most suitable when one is testing, or confirming a theory or comparing alternative theories (Hair et al., 2017, 2011). On the other hand, PLS-SEM is most appropriate when research is exploratory and prediction oriented. As per our research objectives, this study is prediction oriented in that we are trying to predict the relationship between the organizational characteristics and service convenience, and then the relationship between service convenience and CEBs. Testing or confirming a theory is not the main objective of our study. Hence, the use of PLS-SEM is deemed most appropriate for our study. We also used PLS-SEM as it is distribution-free, component-based and computationally very simple. Following Hair et al. (2017), we analyzed and interpreted the research model in two stages: evaluating the validity and reliability of the measurement model followed by empirical testing of the proposed hypotheses. 5.1 Assessment of the measurement model\n6. Discussion: 6.1 Theoretical contributions\n",
"role": "user"
},
{
"content": "Purpose: The purpose of the study is to investigate the role of service convenience in the relationship between organizational characteristics (such as brand equity, store ambiance, store layout, customer information and employee responsiveness) on customer engagement behaviors (CEBs), including service improvement, customer cooperation, positive word-of-mouth and customer helping customers. It examines two research models, with service convenience as a separate antecedent of CEBs (model A) and as a mediating variable between organizational characteristics and CEBs (model B).\nDesign/methodology/approach: Using a positivist paradigm, data were collected from 384 respondents representing the existing customers of grocery retailers based in India via a survey instrument. Data were analyzed using partial least squares (PLS) path modeling.\nFindings: Results demonstrate service convenience as a motivational driver of CEBs. Results also show that the organizational characteristics significantly influence service convenience which in turn impacts CEBs.\nPractical implications: The findings have important implications for store managers in effective management of customers' time and effort in terms of saving customers' time and effort as well as motivating customers to elicit their engagement behaviors.\nOriginality/value: The originality of this paper lies in identifying the impact of organizational characteristics in helping customers to save time and effort in their shopping activities and thereby elicit various types of CEBs. The paper also adds to knowledge by examining the role of service convenience in the nexus between organizational characteristics and CEB types.\n",
"role": "assistant"
}
] |
The perceived effectiveness of overt versus covert promotions
|
[
"Advertising",
"Experimental design",
"Brand communication",
"Consumer behaviour",
"Brand awareness",
"Product placement"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The advertising sector is currently facing increasing criticism from the general public and traditional advertising is continuing to lose popularity (Belch and Belch, 2018). ACNielsen (2018) has recently released a report, which shows that 92 per cent of consumers around the world trust word-of-mouth communications or recommendations from peers and family, while only 47 per cent trust television, magazine and newspaper advertisements. The unfavorable attitudes toward advertising together with the advancement in advertisement-skipping techniques make traditional advertising hard to reach audiences. Advertisement blocking also happens in digital media. PageFair and Adobe (2017) reported that worldwide advertising blocking has increased 41 per cent in 2015 causing the advertising industry to lose nearly US$22bn in revenue. It implies that about 14 per cent of global advertising expenditures were wasted.\nLiterature review and hypothesis development: Overt and covert promotions\nMethodology: A web-based experiment with a 2 (form of promotion: overt versus covert) x 2 (brand awareness: low versus high) between-subjects full factorial design was used to test the hypotheses derived from the research model (Figure 1). A web-based experimental setting enabled the manipulation of the research stimuli and maintained a high degree of control over the allocation of stimuli. The random allocation of research stimuli and a between-subjects design were enforced, while participants were unaware of this. This also allowed the participants to be exposed to the research stimuli in a less artificial environment. The research design combined the merits of high internal validity in an experimental setting and high external validity in a naturalistic setting. A between-subjects design was used to rule out any repetition and carryover effects (Malhotra, 2012). It also helped to prevent the participants from recognizing the research purpose.\nResults: The participants were equally distributed across the four research conditions. There were no significant differences in the research conditions with respect to gender [chi-square (kh2)(3) = 1.03 and p = 0.79], age group (kh2(3) = 1.05 and p = 0.79), education (kh2(9) = 8.16 and p = 0.52), income (kh2(18) = 4.64 and p = 1.00) or level of skepticism toward advertising F(3, 533) = 1.89 and p = 0.13. Therefore, these factors were unlikely to be confounding variables in the research model.\nDiscussion and conclusion: In response to earlier calls, this study empirically contrasted the effectiveness of overt and covert promotions of a well-known and a less well-known brand using a public sample (Liljander et al., 2015). A direct contrast between the effect of overt and covert promotions is sparse and inconsistent findings have been recorded in previous studies. For example, studies that have compared the effectiveness between advertising and product placement have primarily focused on well-known brands (Davtyan and Cunningham, 2017; Uribe, 2016). The current study is the first to extend this work to recipes, which are a covert form of promotion and is increasingly adopted by food and grocery marketers (Luth, 2017). Therefore, this represents the first study to empirically examine the mediating role of perceived believability of promotional materials and the moderating roles of advertising skepticism and brand awareness on the evaluation of brands placed in overt and covert promotions. The conceptual model was further validated using research stimuli of a fictitious brand and a similar pattern of results was obtained, which confirmed the robustness of the model. The results have important theoretical implications and they are able to advance the field of knowledge. The study is also able to provide pragmatic recommendations to marketers who wish to develop effective marketing communication strategies for brands.\nLimitations and future studies: This section describes some of the methodological limitations of this study, and makes a number of recommendations for future research. The first limitation is that this study focused on consumers of 18 years old or above. However, it is important to examine the effect of information-coated promotions on younger generations and to check whether the relationships established in this study will be moderated by the individuals' cognitive development. In a recent study, Van Reijmersdal et al. (2015) found that boys of 8-12 years old found toy store catalogs (i.e. overt promotion) more persuasive than a brand-integrated magazine (i.e. covert promotion), even though they were more aware of the persuasive intent of the former. Consequently, future studies may extend this work to the younger cohort or to adolescents.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study contributes to literature on marketing communications by empirically contrasting the effectiveness of a new form of covert promotions (product placement in recipes) and an overt promotion (traditional advertisement). The mediating role of perceived believability of promotional materials and the moderating roles of advertising skepticism and brand awareness were examined based on a conceptual model.\nDesign/methodology/approach: A Web-based experiment with 2 (form of promotion: overt versus covert) x 2 (brand awareness: low versus high) between-subjects full factorial design was used. A public sample of 537 participants had participated in the study. The conceptual model was further tested on 106 participants using stimuli of a fictitious brand.\nFindings: A two-stage moderated mediation analysis shows that the perceived believability of promotional materials was a significant mediator of the form of promotion and brand evaluations. Consumers showed a higher level of believability toward covert promotion, which, in turn, led to more positive evaluations of the promoted brand. Advertising skepticism and brand awareness were found to significantly moderate the relationship between form of promotion and attitudes toward the promoted brands. A similar pattern of results was obtained when stimuli of a fictitious brand were used.\nOriginality/value: This research addresses an important issue in marketing communication and extends the understanding of the perception of overt and covert promotions by examining the underlying mediating and moderating variables, which have rarely been explored before. The results guide marketers in developing effective marketing communication strategies for well-known, less well-known and even new brands. It also directs policymakers to consider whether integrated branded content in recipes should be disclosed to protect consumers from surreptitious promotions, which may help to lower consumers' skepticism toward advertising in the long run.\n",
"role": "assistant"
}
] |
Shifting consumers into gear: car sharing services in urban areas
|
[
"Collaborative consumption",
"Sharing economy",
"Car ownership",
"Car sharing services"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In the last decade, sharing rather than owning goods in a conventional fashion has grown to become a major trend in markets and economies worldwide. It has been suggested that this emerging sharing paradigm is potentially transformative of established economic and consumption patterns and social trends, and driven by powerful technological forces. For instance, Time magazine included collaborative consumption as one of ten ideas that could change the world (Walsh, 2011).\nConceptual background: Collaborative consumption refers to a broad range of practices from transactions that are not market-mediated (e.g. swapping and not-for-profit communal sharing) to access-based services that induce pay-offs (Acquier et al., 2017). For the purpose of this research, collaborative consumption is defined as people coordinating the acquisition and distribution of a resource for a fee or other compensation (Belk, 2014).\nHypothesis development: Knowledge\nStudy design: As noted above, the aim of this study is to investigate the consumer adoption intention of automobile-sharing systems by conducting a large multinational survey. The survey was carried out in collaboration with the TNS Sofres research institute in three European capital cities: London, Paris and Madrid. As in Schaefers et al. (2016), a self-administered online questionnaire was used. Considering quota sampling (Deville, 1991), 2,159 licensed car drivers were recruited through the online panel of TNS Sofres: 742 individuals in Paris, 691 in London and 726 in Madrid. Each subsample was balanced in terms of age, gender, employment status and location (city centre vs suburbs). The sample was representative of the population of licensed car drivers in each city. Only car owners were targeted, as their use of car sharing services may provide substantial economic, social and environmental benefits (Prieto et al., 2017).\nEmpirical analysis: The measurement model was assessed using Partial Least Squares Path Modelling (Chin, 1998). Following prior studies (e.g. Barnes and Mattsson, 2017), this approach was chosen over covariance-based structural equation modelling due to its advantages (e.g. Reinartz et al., 2009).\nDiscussion: It has been suggested that the future belongs to sharing as powerful technological forces allow companies to provide people with goods and services at the exact moment they need them, without the burden and expense of owning them outright (Botsman and Rogers, 2010). So far, however, this new phenomenon (Benoit et al., 2017) has remained a largely understudied area in the literature and there is a strong consensus on the need for further investigations.\n",
"role": "user"
},
{
"content": "Purpose: Recently, the sharing economy has attracted considerable attention. This emerging paradigm is driven by powerful technological forces and has the potential to change the way consumers access very important markets such as the car market. Indeed, access-based consumption may attract more participants as it enables consumers' freedom of lifestyle and more flexible identity projects. The empirical literature has so far paid very little attention to car sharing services; when it has, it has focussed mainly on people who are already using them. The purpose of this paper is to consider the drivers behind the adoption intention of car sharing services.\nDesign/methodology/approach: A large multinational survey is carried out in three European capitals: London, Madrid and Paris. Using quota sampling, 2,159 licensed car drivers are recruited through the online panel of TNS Sofres. The sample is representative of the population of licensed car drivers in each city. The questionnaire is developed using established scales from previous research. An OLS regression analysis is performed to test our hypotheses, with a likelihood of choosing a car sharing option as the dependent variable.\nFindings: The study demonstrates that knowledge, environmentalism, possession-self link and involvement with cars are important determinants of consumer behaviour in the car sharing services market. In addition, the user demographics suggest a target market of younger, predominantly male and urban customers. The empirical findings are consistent across the three capital cities, implying that providers can market their car sharing services in a similar manner.\nPractical implications: As important determinants of consumer behaviour in the car sharing services market are underlined, several managerial implications arise from the study. Car sharing providers should promote awareness to help people not only to expand their experience with the service but also to be informed about the potential environmental benefits. Further, a stronger possession-self link in the automotive context is suggestive of a greater willingness to use car sharing systems. Managers should also take into account that it is much more difficult to engage individuals in car sharing services who are highly involved with car, than those who express very little attachment to the product. As people under 45 years old are far more likely to use these services, this generation effect is progressively moving the frontiers of the car retailing sector to a broader mobility service sector.\nOriginality/value: The foremost contribution of this paper is to demonstrate empirically how consumer intention to use car sharing is driven. To do so, the study addresses the general population of car drivers, interviewing users of the service as well as non-users.\n",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.