title
stringlengths 6
301
| keywords
listlengths 0
23
| messages
listlengths 2
2
|
|---|---|---|
Product positioning based on search, experience and credence attributes using conjoint analysis
|
[
"Product attributes",
"Consumer behaviour",
"Product specification",
"Product positioning",
"Product management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\nIntroduction: Just a few years ago, teeth whitening was considered a cosmetic procedure, something to be done at the dentist's office. Now, casual reading of magazines or watching TV commercials indicates that teeth-whitening products are entering the mainstream acceptance by the consumers. In fact, Stevenson (2002) predicted that teeth whiteners were in place to be analogous in market penetration to the deodorant market. Once considered an option, it is now simply a \"way of life\". The teeth-whitening products had their origins in the dentist's office; they are now accepted as \"at home\" products.As the market for teeth-whitening products grows, so does the variety of teeth-whitening products. For example, Crest's most popular Whitestrips is now joined by Night Effects. According to Fasig (2004), Crest is estimating its sales of whitening products to reach 2 billion by 2009. The increased market size is also having an effect on distribution practices of these products. An estimated 45,000 dentist offices now sell Whitestrips.In a competitive market (as well as in new product introductions), it is vital to understand the customers' benefit preferences. For example, P&G is positioning its Crest Whitestrips as simple to use. Rembrandt positions its Oral B Whitestrips as one that delivers white teeth in only five days of applications. In a growing or mature market these characteristics become even more important in differentiating these competitive products. The knowledge of consumer benefits is critical in product positioning. One question that is salient in positioning a product is the following: which attributes should be stressed in positioning a product? Marketing literature offers some guidance in this regard. Chernev (2004) proposes that products with attributes compatible with individuals' goal orientation tend to be over weighted in their choice of the products. Their findings reveal that consumers choose a brand with attributes that are consistent with their goals. For example, subjects with hedonistic goals choose a whitening product, whereas subjects with a cavity-prevention goal choose a tooth paste shown to prevent cavities. Johar and Sirgy (1991), develop a framework that describes when a particular advertising appeal (value expressive versus utilitarian) will be more effective. This framework is based on congruity between the self concept (self image) and functional approach to attitudes. According to this framework, it is expected that individuals desiring whiter teeth would attach higher importance and higher utility preferences to products offering such benefits.One major goal of this investigation is to determine salient attributes in positioning teeth-whitening products. The major marketing issues explored in this research are described next. The first issue is the selection of attributes. To aid in selection of attributes, information theory researchers have suggested three categories of product's attributes: search attributes, experience attributes and credence attributes. The research includes attributes that represent these three categories. Next the research will use conjoint analysis to estimate the utilities of various attributes in the choice of teeth-whitening products. The use of conjoint analysis is especially relevant when consumers make choices amongst products with differing attributes. Conjoint analysis helps determine the relative importance and utilities of each attribute.Second, an important area in using health and beauty care products is the side effects that may result from using these products. This research will investigate the impact on consumer choice from disclosing side effects. Third, an important contribution of this research is that it will extend earlier research by investigating the role of penny-a-day (pricing in cents) versus full pricing in the context of teeth-whitening products using conjoint analysis. Pricing will be one of the attributes investigated in the study.In conclusion, the earlier paragraphs identified major marketing issues to be investigated in this research. These issues are discussed in the following sequence: search, experience and credence-based attribute framework; the impact of disclosing side effects from the use of products; and penny-a-day versus full pricing.\nLiterature review: Teeth-whitening products form the setting or the context within which several important aspects of marketing will be explored in this research. This setting is important in that the size of the teeth-whitening products is large in number of consumers and total revenue. It is a product that has widespread awareness amongst consumers. This product category lends itself to investigation of these marketing issues. The major marketing issues to be explored in this research are described next.Selection of attributes\nMethod: Conjoint analysis\nResults: The part-worth utility estimates for skills of the applicants are shown in Figure 1. The CVA program provides part-worth estimates for each level of the selected variable. It also provides standard errors for these estimates and an estimate of R square indicating an overall measure of goodness of fit. A total of ten utility estimates resulted in R square values of less than or equal to 0.70. These ten cases were dropped from further analysis. In calculating the utility estimates, the utility scores for the less preferred level of the attribute are set to zero. Thus the utility shown for each variable reflects the additional utility to the respondents if they were to receive the service at the preferred level of the attribute.The relative importance of these attributes is shown in Table II. The importance values are calculated from the range of utility estimates for the attributes. The range of utility of each attribute is divided by the aggregated utility range to arrive at individual importance values.\nDiscussion and implications: The purpose of this paper was to explore several marketing issues that would help marketers on various aspects of marketing positioning. First the question of role of search, experience and credence attributes. Based on utility estimates, the most significant attribute is longevity of effectiveness. The next two important attributes are the time it takes to see the results and side effects. None of these are search attributes. Thus, consumers would have considerable difficulty in assessing the product or making an informed choice without trying the product. An effective strategy in such case is sampling to help consumers in trying the product and using high credibility sources to create or change their attitudes toward using the whitening products.Next consider the issue of price framing; i.e. is it better to promote the product using total of package or price per day for using the product? The utility estimates reveal a relatively small but statistically significant preference towards penny-a-day pricing approach.Regarding the disclosure of side effects, the utilities for reporting negative side effects is -0.42. Thus, a whitening product that may cause any side effects, even amongst a very small percentage of population (less than 1 percent in this case), is likely to reduce the product utility significantly. Marketers would be well advised to make necessary changes to the product so there are no side effects, because increasing the salience of side effects, no matter how negligible is detrimental to marketing efforts. The whitening industry does not seem to stress safety as an issue. In fact one of the few products that explicitly mentions safety is \"TresWhite Professional Teeth Whitening System.\" The product claim asserts that it is, \"safe and effective in-home treatment for people with healthy un-restored teeth.\"Lastly, a few comments related to the market side, i.e. the current promotion of these products is worth noting. One of the heavily-promoted themes is that the benefits of using these products are visible in as few as -- days. The utility estimates we obtained were for 7 days versus 14 days. It appears that this difference of one week is also large and significant. Individuals' utility is improved by seeing results in 7 days or lost by extending the time period to 14 days. A related, but less promoted feature is the enduring results of white teeth. This item has the highest utility. A whitening product that promises (and delivers) the results to have long-term benefits can offset other benefits such as the form in which the product comes, choice of flavors, and even the drawback of any perceived side effects. Marketers would be better served by having an effective and safe teeth-whitening product that promises quick and enduring results. All other attributes are secondary. Choice of flavors was the least important attribute. This finding is consistent with industry practice. In fact, very few products stress multiple flavors.CVA provides individual level utility estimates which are critical in strategic policy implications such as the ability to group individuals into meaningful segments based on these preferences. An obvious segmentation variable is gender. A t-test of mean utilities for these six attributes indicates the market to be fairly homogeneous on most attributes. One attribute that revealed moderate statistical significance was the number of days to see the effects of the product. The utility for males was 0.36, while for females it was 0.60 (p=0.06). Although this attribute is considered to be the second most important attribute (approximate importance=21 percent) its relative importance varies by gender with females attaching a much higher utility.Casual study of the whitening-product market reveals that Crest and Oral B Rembrandt are two products that use the number of days as a salient positioning attribute (Table III). In fact Crest has introduced Crest Whitestrips Supreme and Crest Whitestrips Premium that are consistent with the findings and recommendations of analyses in this paper.In addition to above findings we also looked at some other relevant questions. For example, does the current coloration of the teeth (relatively white or yellowish) have a differential impact on the utilities? For example, those with relatively stained teeth might attach higher utility to the time it takes to attain whiter teeth. The findings revealed no significant difference across the sample. Those with whiter teeth or stained teeth had similar utility estimates across the attributes.Respondents were asked if they were currently using any teeth-whitening product and if so, whether the product was a gel or strips. We wanted to explore the relative utilities for the attribute-form of the product based on their current behavior. The results indicated that users of whitening products as gel were equally predisposed to product form, while the users of product in the form of strips had a significantly higher utility (1.23) for the product in the form of strips. This indicates that the latter segment is highly loyal to the form of the product.Lastly, a few cautions are worth noting. One of the limitations of conjoint analysis is that the levels of the attributes must be discrete. For example, respondents were given the choice that the results of teeth-whitening product may last for 3 months or 6 months. While the levels chosen appear to be reasonable in view of available products in the market, questions do arise such as what if the results of teeth whitening were to last say 4 months? Thus, caution is advised in extrapolating the results.\n",
"role": "user"
},
{
"content": "Purpose: - Using the framework of search attributes, experience attributes, credence attributes and conjoint analysis methodology, the purpose of this paper is to estimate the utilities of various attributes that may be used in positioning teeth-whitening products. In addition, this research also investigates the role of \"penny-a-day\" (pricing in cents) versus full pricing. Third, disclosure of side-effects that may result from using these products (an important health care issue) and its impact on consumer choice are also investigated.\nDesign/methodology/approach: - The study uses conjoint analysis to estimate the utilities of key attributes of teeth-whitening products.\nFindings: - The results reveal that consumers prefer a product that provides benefits lasting for several months after use, a product that does not result in any side-effects and a slight preference for penny-a-day pricing approach.\nResearch limitations/implications: - Caution is advised in extrapolating the results beyond the range investigated in the study.\nPractical implications: - The findings help marketers in designing new products and positioning teeth-whitening products based on desirable attributes.\nOriginality/value: - Use of conjoint analysis in investigating the above issues (penny-a-day pricing, disclosure of side-effects) has received limited attention in teeth-whitening products.\n",
"role": "assistant"
}
] |
Mediating role of process simplification in process integration and upstream supply chain flexibility
|
[
"Top management commitment",
"Mediation",
"Supplier relationship",
"Indian automotive industry",
"Process simplification",
"Upstream supply chain flexibility"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: With the rising global pressure for quick production and delivery of automobiles, automotive organizations are responding by focusing on flexibility as an added dimension of their supply chain system. Flexibility may be defined as the ability to change or react within a limited period of time, effort and cost (Upton, 1994). Specifically in the supply chain context, Slack (1987) explains flexibility as the ability of the manufacturing system to respond to changes in product and volume. Thereby, having flexibility in the supply chain represents a potential source for improving organizational efficiency (Vickery et al., 1999; Hoole, 2005; Weingarten et al., 2016). The automotive organization involves many external supply chain partners and it has been noticed that flexibility initiatives carried out by a chain of plants outperforms the flexibility initiatives of a single plant (Jordan and Graves, 1995). Therefore, it becomes quite critical to improve the flexibility initiatives of the automotive organization and its partners taken together as a single entity comprising the supply chain. The supply chain must have the capability to fill items in the product portfolio of the manufacturing plant from their respective suppliers in a timely and cost-effective manner, also referred to here as upstream supply chain flexibility. Researchers have recommended that automotive organizations that concentrate largely on process-based performance must focus more on upstream practices as compared to downstream practices of the supply chain (Guide and Wassenhove, 2001; Vachon and Klassen, 2008). Researchers have also argued that customers change their buying decision(s) due to late delivery of automobiles caused by problems with upstream supply chain flexibility (Scannell et al., 2000). Hence, it becomes more pertinent to focus on upstream supply chain flexibility with respect to automotive organizations.\n2. Literature review and hypothesis development: This section of the study focuses on formulating relationships among the constructs and identification of the research gap which also forms the basis for the proposed research model. The sub-sections emphasize the importance of upstream supply chain flexibility and the role of process integration and process simplification in enhancing the upstream supply chain flexibility.\n3. Methodology: This section of the study is divided into three sub-sections. The survey instrument and data collection section concentrate on the formulation of the research questionnaire, sampling techniques to be used, respondent's profile, organization's demographics and response rate for the study. The internal consistency section focuses on measuring the Cronbach's a value for the constructs using SPSS version 21. In the analysis section, correlation analysis is performed first followed by the three-step regression process procedure proposed by Baron and Kenny (1986) using SPSS macro named PROCESS (Hayes, 2013).\n4. Results and analysis: This section of the study is divided into two sub-sections that discuss the mediating impact of process simplification on the relationship of process integration constructs and upstream supply chain flexibility.\n5. Theoretical contribution and managerial implications: This section of the study is divided into three sub-sections. The first sub-section discusses the theoretical contribution of the study. The focus of the next two sub-sections is on the managerial implications related to the mediating impact of process simplification on the relationship between process integration (top management commitment and supplier relationship) and upstream supply chain flexibility.\n6. Conclusion and future scope: The study empirically validates that process simplification mediates the relationship among process integration (top management commitment and supplier relationship) and upstream supply chain flexibility in the Indian automotive organization. The analysis suggests that there is a partial mediation impact of process simplification on the relationship of top management commitment and upstream supply chain flexibility. Also, complete mediation impact of process simplification is suggested on the relationship between supplier relationship and upstream supply chain flexibility.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the mediating role of process simplification on the relationship between process integration and upstream supply chain flexibility in Indian automotive organizations.\nDesign/methodology/approach: The three-step mediation analysis was performed using SPSS macro PROCESS to assess the mediating role of process simplification on the relationship between process integration constructs: top management commitment and supplier relationship; and upstream supply chain flexibility.\nFindings: The results indicate a complete mediation effect of process simplification between supplier relationship and upstream supply chain flexibility, while partial mediation effect is noticed between top management commitment and upstream supply chain flexibility.\nPractical implications: For Indian automotive managers, the study suggests that for improving the upstream supply chain flexibility, organizations must have a strategy towards improving the simplification of supply chain processes by upgrading technology and providing training to their suppliers. This understanding will help the automotive managers to simplify their upstream supply chain processes for gaining competitive positioning and maximizing the organizational profit.\nOriginality/value: This study has considered the mediating role of process simplification (a relatively less studied variable) specifically in the context of its impact on upstream supply chain flexibility. Also, the presented study explores this role in the Indian automotive domain which further enhances its value for both practitioners and researchers alike.\n",
"role": "assistant"
}
] |
Master degree under crisis: The salient motives of business students to enrol in a postgraduate programme
|
[
"Crisis",
"Syria",
"Postgraduate programme",
"Business administration",
"Master degree",
"Student motives"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In an attempt to understand the complex process for students to get enrolled in postgraduate programmes, Kotler and Fox's (1995) model of student buyer behaviour can be utilised. The student buyer behaviour model depicts this process in five consecutive stages. These are \"needs and motives\", \"information gathering\", \"evaluating alternatives\", \"decision and enrolment,\" and finally \"post-choice evaluation\". Most of the previous research has focused on the second, third, or fourth stages in the model. Recently, AL-Mutairi and Saeid, (2016) highlighted family, alumni, universities, and media as the main sources of information towards choosing an MBA programme. Other studies proceeded to the selection criteria of areas of study, programmes, or universities, which may be the cost, reputation, convenience, programme duration, or the university's location (Subramaniam et al., 2014). However, the exploration of the students' motives behind their behaviours is still not enough considered. Such exploration is highly important for university marketers to enable them plan their postgraduate programmes according to the real needs of students.\nLiterature review: In spite of the increasing importance of postgraduate education at individual, institutional, and national levels, an intensive review of literature unveiled a lack of research on postgraduate education (Keane, 2016; McCulloch and Thomas, 2013; Morgan, 2014). The dearth of research on the transition to postgraduate education and related academic practices has also been noted (Tobbell and O'Donnell, 2013; Tobbell et al., 2010). Researchers argued that the fact may be due to suppositions of the \"expert status\" towards the postgraduate students (in other words, these students are capable of handling postgraduate education practices, given their successful completion of undergraduate level and advancement to postgraduate level) (ibid). Research, however, intensifies the need to help postgraduate students manage their learning and life (ibid), and that institutions must consider the heterogeneity of postgraduate student groups, especially in terms of their different aspirations (O'Donnell et al., 2009).\nMethodology: This study sought to explore the students' motives to attend a postgraduate business programme based on the Syrian context. Epistemologically, acknowledging the subjective and individualistic reality regarding the motives was essential to understanding this phenomenon (Pocock and McIntosh, 2010). In the same manner, recognising the existence of such multiple realities was considered important for generating an in-depth insight on which to recognise the students' motives from insider's perspective. Exploration of students' motives was generated in this study using semi-structured interviews as advised by Robson (2002) for exploratory research. Interview questions were developed based on the research objective and in light of relevant literature (Al-Fattal, 2010) (Appendix).\nThe results of the interviews: In response to the study question, which aims at exploring the students' motives to attend a postgraduate programme, this study found that motives could mainly be grouped into six categories. These are: self-motives, social motives, professional motives, academic motives, delaying military service, and lack of vision. These motives are not necessarily crystallised inside the students since long periods of their lifetimes. Rather, a considerable number of students reported that they were motivated to attend a postgraduate programme during their last years of bachelor programme, or even after getting the degree. A detailed description of the six categories of motives is listed below. Students' names in the quotations are fake for privacy reasons.\nConclusion and recommendations: The results from the interviews indicate the existence of six different motives for students to pursue a postgraduate study: self-motives, professional motives, social motives, academic motives, lack of vision, and delaying military service. Amongst those motives, only the last one is directly caused by the current war conditions in Syria where students seek postgraduate study to legally postpone their military service. However, this is only applicable for males while females are not obliged to do the military service in Syria. The authors, therefore, suggest adding a motivation criterion to the selection process of postgraduate candidates. Such criterion should help the higher education system maintain students with the highest academic and professional motives.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore why students decide to enrol in a business postgraduate programme at Damascus University in the current Syrian crisis.\nDesign/methodology/approach: Exploration of students' motives was generated in this study using semi-structured interviews. On the basis of saturation sampling, 11 interviews took place in the leading Syrian university providing postgraduate programmes, Damascus University.\nFindings: The results from the interviews indicate the existence of six different motives for students to enrol in a postgraduate study: self-motives, professional motives, social motives, academic motives, lack of vision, and delaying military service, which is directly caused by the current war conditions in Syria.\nPractical implications: Understanding postgraduates' motives is essential at the national level, the institutional level, and also at the individual level to make better future plans related to opening new programmes or altering admission criteria. Recommendations to higher education policy makers are highlighted in the study.\nOriginality/value: The majority of previous studies concentrate on students' motives to pursue postgraduate studies during financial crisis. However, very little is known on why students decide to enrol in a business postgraduate programme in a war context.\n",
"role": "assistant"
}
] |
Reaching beyond the wire: challenges facing wireless for the last mile
|
[
"Case studies",
"Wireless",
"Communication technologies",
"Broadband networks",
"Business policy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Ongoing developments in wireless technology have prompted communities, educational institutions, internet service providers and various other organizations across the world to explore using wireless to complement or even substitute for existing wireline broadband internet services. Practitioners seeking to implement wireless technology have to account for many context specific factors, including how local regulations, geographies and socioeconomic characteristics might affect the likelihood of success. Relying on the varied contexts that arise in the USA, this paper provides a critical perspective on the major opportunities and challenges that diverse organizations are likely to face when attempting to use wireless to extend broadband access.\n2. Opportunities and challenges in rural USA: 2.1 Opportunities\n3. Opportunities and challenges in urban USA: 3.1 Opportunities\n4. Discussion: New developments in wireless technology, coupled with changing regulations that have permitted these technologies to be deployed, have created new opportunities for expanding internet broadband coverage and competitive options. Although wireless service has, in the past, been largely associated with mobile wireless, which has notable differences from wireline (especially cable- and fiber-delivered) broadband service in terms of speeds, reliability, data caps and end-user devices, the role of wireless is evolving to encompass both fixed and mobile usage. In addition, wireless has become a supplement and sometimes a substitute for wireline as a form of backhaul, a way to transmit data between different parts of a wireless network rather than as only a means for consumers to communicate with that network.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to provide a critical perspective on the major opportunities and challenges facing the future of wireless access to broadband internet services in the USA. In particular, it seeks to develop realistic expectations for new entrants and existing wireline broadband providers, hoping to diversify into wireless to extend the reach of broadband to previously underserved markets.\nDesign/methodology/approach: The paper synthesizes a series of empirical case studies of efforts to develop wireless projects across the USA. Cases were chosen to reflect variations in recent initiatives related to spectrum used, housing density, topography, types of services provided and the type of entity providing them. Findings are supplemented by a thorough review of applicable sections of the Code of Federal Regulations, Federal Communications Commission Orders and state and local regulations.\nFindings: Although developments in wireless technology and regulation have permitted novel approaches for using wireless to extend the reach of broadband, myriad challenges must be addressed for successful expansion. These include limitations on throughput, infrastructure and equipment hurdles, regulatory uncertainty and relatively low revenue potential in underserved markets.\nPractical implications: This study can serve as a guide to the nuances that wireless broadband providers must take into account when extending service in different settings. Because social and regulatory contexts are critical, the authors anchor their findings in a study of US cases and regulations. Nonetheless, the opportunities and challenges that are explored can be of relevance for practitioners considering wireless access in other countries where comparable situations may arise.\nOriginality/value: Existing scholarship generally focuses on the opportunities of wireless technologies for extending the reach and quality of broadband without methodically exploring potential hurdles. In this paper, the authors carefully consider both, paying attention to the social, financial, policy and technical challenges that must be addressed to successfully design and implement different wireless initiatives.\n",
"role": "assistant"
}
] |
Product packaging and consumers' emotional response. Does spatial representation influence product evaluation and choice?
|
[
"Emotions",
"Packaging",
"Construal level theory",
"Face reader",
"Spatial distance"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Packaging is one of the most important factors in the buying decision process. It communicates to consumers at the point of sale, thus becoming an essential part of the selling strategy (Prendergast and Pitt, 1996; Rettie and Brewer, 2000; Silayoi and Speece, 2004). In today's increasingly competitive marketplace, package becomes a primary vehicle for communication and branding, particularly in the case of new brand introductions, brand repositioning or extensions, product changes and product categories in which the use of pictures on packages would provide a strategic method of differentiation (Garber et al., 2000). In light of this evidence, an understanding of the impact of package design elements is crucial to enhance point-of-purchase communication.\nLiterature review: The role of packaging in the buying decision process\nMethodology: The emotional state of individuals watching the picture of the product on the package in the two spatially distance conditions (foreground and background) was measured through the affective computing technology. The FaceReader 5 software by Noldus was used to track and measure the presence and intensity of emotional reactions of participants as captured in their facial expressions while watching the stimuli, namely, the product package images. Such facial-recognition device captures images of human face through the use of a webcam and can identify the six basic emotions described by Ekman (1970): happiness, sadness, anger, surprise, scare, disgust and a neutral state. Moreover, the software provides measures of overall emotional valence, which indicates whether the person's emotional status is positive or negative. The face modelling method uses a database of over 10,000 annotated images and calculates the main sources of variation found in the model dimensionality. New faces can then be described as deviations from the mean face, using a vector. The actual classification of the facial expressions is done by training an artificial neural network, which takes the above vector as input. FaceReader can recognize facial expression with an accuracy of 90 per cent. The output of FaceReader is a vector of values for the seven emotions and an overall valence of emotional state. The possible values of each emotion range from 0 to 1, and valence ranges from -1 to +1. The valence measure compares the conformity of the facial expression to \"happy\", the only positive emotion, with that to the four negative emotions (sad, angry, scared and disgusted).\nStudy 1: Experimental stimuli\nStudy 2: Experimental stimuli\nDiscussion: In a context of growing competitive market conditions, the package has become a key element for conveying image, symbolism and/or brand meaning, thus effectively differentiating the product from competitors (Underwood and Ozanne, 1998). As a consequence, opportunities for packaging to attract, communicate and influence consumers' choices are growing relative to more traditional marketing communication vehicles (e.g. mass media advertising). According to Harckham (1989), the package is the \"shopper's window\" to a product because it often projects the initial impression he forms about the product, its quality or its value. Therefore, literature has focused on the communication elements in the package and, among these, on the presence of the product imagery as a vehicle to influence consumers' evaluation in terms of perception of intrinsic characteristics of the product and purchase decisions.\nLimitations and future research: Some limitations of this work merit future inquiry. First, the sample composition focused on a student target may have partially influenced the findings of the study, limiting its generalizability. Further studies are recommended to extend the investigation to a larger sample that is representative of the general population. Second, because the position of the product imagery on the package does not influence purchase intention, it would be appropriate to include in the analysis variables that may mediate the causal relationships. Third, only intentions were measured, not actual behaviours. A new empirical research should be undertaken to investigate if the differences in the product representation on the package translate into different purchasing decisions. Contextually, further research in this field should verify whether the product imagery position is capable of affecting other aspects of the consumer's response to the product (e.g. attention, evaluation and taste expectation). Fourth, it would be interesting to replicate the study focusing on private labels to assess if the consumers' emotional reactions differ compared to industrial brand products. Finally, investigating the hypothesized relationships through an in-store experiment would enrich the results of the study. The attention prompt in the experiments could elevate cognitive elaboration or product involvement, which may in turn affect the purchase intention measured. If in the store, individuals are less thoughtful and show a lower level of involvement towards the product, then the packaging would have more impact on behaviour.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the influence of different spatial representations (foreground and background) of a product image on the package on the consumer emotional state and intention to buy the product, drawing upon the construal level theory (CLT).\nDesign/methodology/approach: Two between-subjects experimental designs were conducted. A face-reading software was used to measure the emotional reactions of participants while watching the stimuli.\nFindings: Empirical results showed that the foreground representation of the product, triggered by a lower perceived psychological distance between the subject and the product, generates more positive emotional responses compared to the background representation. However, the purchase intention did not significantly differ between the foreground condition and the background condition.\nResearch limitations/implications: The study enriches the literature on CLT and provides some guidance to manufacturers and product managers for the understanding of the role of packaging in influencing consumer behaviour.\nOriginality/value: The paper contributes to the research on the presence of the product imagery on the package. The focus is shifted from the presence/absence of the picture to its position (proximal or distal). This aspect has been neglected in previous studies. Moreover, it is the first application of face reading to support CLT basic assumption and enrich extant marketing research on packaging.\n",
"role": "assistant"
}
] |
The influences of relational benefits on repurchase intention in service contexts: the roles of gratitude, trust and commitment
|
[
"Relationship marketing",
"Gratitude",
"Trust",
"Relational benefits",
"Relationship commitment"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Relationship marketing (RM) has increasingly been a major focus in both academia and the service industry (Payne and Frow, 2017). This is mainly because, as industrial competition grows, maintaining relationships with existing customers can generate higher profits than acquiring new ones (Reichheld and Earl Sasser, 1990). As a result, research on and strategies for RM have been mounting steadily and are critical in the service industry (Gummesson, 2017; Payne and Frow, 2017).\n2. Literature review and hypothesis development: 2.1 Relational benefits\n3. Methodology: 3.1 Sample and data collection\n4. Results: A partial-least-square-based structural equation modeling (PLS-SEM) was adopted to evaluate the measurement of research constructs and test the research hypotheses in this study. We used PLS-SEM to examine the model for two reasons. First, PLS-SEM is a structural equation modeling technique that simultaneously assesses the reliability and validity of the measures of constructs and estimates the relationships among them (Lohmoller, 1989). Second, contrary to other SEM techniques such as LISREL, PLS-SEM does not require a larger sample (Chin, 1998). Given that small data points were being tested for this study, PLS-SEM is suitable.\n5. Discussion: This study proposes an RM model to understand why the relational benefits perceived by customers in a service context can motivate repurchase intention. In addition to trust and relationship commitment, we add gratitude, a positive emotion incurred after the reception of benefits, to the model as an additional mediating factor between relational benefits (relationship antecedents) and repurchase intention (relationship outcome). On the basis of the proposed model, we tested the research hypotheses with a sample size of 253.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to develop and examine an integrated mediating model that contains relationship benefits as the relationship antecedents; trust, relationship commitment and gratitude as mediators; and repurchase intention as the relationship consequence in a service business context.\nDesign/methodology/approach: The study recruited 253 respondents to test the proposed model. A partial-least-square-based structural equation modeling was adopted to evaluate the measurement of research constructs and test the research hypotheses that underpinned a proposed conceptual model.\nFindings: The results show that confidence benefits increase trust more than social and special treatment benefits; social benefits have the most influence on relationship commitment; and special treatment benefits most greatly enhance feelings of gratitude. Thus, depending on the improvement objectives, marketers can accordingly choose the proper relationship benefits as a primary approach. Furthermore, both the direct and indirect effects of confidence benefits and social benefits on repurchase intention were significant.\nResearch limitations/implications: The main limitation of this study is that all research variables were measured at a static point rather than as they were developing. Ignorance of the temporal sequence may incur the concern of possible spurious causal relationship inferences.\nPractical implications: The study's findings offer several important implications for service providers, which, in this study, were separated into high (e.g. such as hairdressing, beauty and massage) and low personal contact service groups (e.g. telecommunications and equipment maintenance). First, social benefits are more effective at increasing gratitude in the low personal contact service group, and at boosting trust in the high-contact group. Second, special treatment benefits increase gratitude much more in the high-contact group than in the low personal contact group. Third, gratitude has a much stronger effect on repurchase intention in the high personal contact group than in the low-contact group.\nOriginality/value: The interactions among relationship benefits, mediators (i.e. trust and relationship commitment) and outcomes remain unclear. Although the importance of gratitude as a mediator in relationship marketing (RM) has been shown, little research has examined how relational benefits influence repurchase intention through the combined mediating effects of commitment, trust and gratitude. Therefore, this study advances RM theory by examining the degree of additional explaining power that gratitude contributes to the commitment-trust model. The study also makes a practical contribution by building a research model to understand the nature of mediating roles played by relationship commitment, trust and gratitude in the service industry.\n",
"role": "assistant"
}
] |
RFID-based traceability in the supply chain
|
[
"Supply chain management",
"Radio frequencies",
"Identification",
"Modelling",
"Tracer methods"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: RFID (Radio Frequency IDentification) technology promises revolutions in areas such as supply chain management and ubiquitous computing by its peculiar characteristics of pervasive, low-cost sensing and identification (EPCGlobal, 2005). RFID technology allows a sensor (RFID reader) to read and write, without line of sight, many items at the same time using a unique tag attached to an item. RFID provides an alternative to bar code identification systems and it is being gradually adopted and deployed in a wide area of applications, including supply chain management (Wal-Mart, 2005).Large retailers like Wal-Mart and Target have begun implementing RFID systems in their warehouses and distribution centers, and are demanding their suppliers to tag products with RFID at the pallet and case levels. In the near future, we can expect tags to be placed at the individual item level since the cost of tag is continuously decreasing. Such enormous amount of low-level data acquired from RFID systems and high-level queries pose great challenges to traditional data warehouse (DW) technologies. Several recent studies have addressed these challenges and proposed methods to overcome those issues related to RFID data management (Gonzalez et al., 2006b, c). Moreover, some works have been researched about the traceability requirements and data models to support the end-to-end traceability but disregarded the \"combine and split\" process of product items that is frequently occurring in the manufacturer tier which will end up forming an incomplete traceability. In this paper, we propose a novel RFID dynamic tracing task model (DTTM) in which the \"combine and split\" concept is taken into account with the use of BOM (Bill of Material) and related modules.Several recent studies have shown that RFID data will create a \"flowgraph\" (Gonzalez et al., 2006a, c) in which all the path information of a certain object is stored. However, since objects in the supply chain combine and split a lot to become a product, more than a flowgraph is needed. In this sense, our DTTM will be perfectly suited in the actual supply chain since it has the capability to trace not only the end item but also the components of that item giving the whole span of traceability view along the supply chain. DTTM incorporates the generic BOM and transforms the BOM structure into a series of sets in which the information of BOM structure is inserted. It then uses this set to create certain rules in order to make the connections between the end item and the subparts in terms of traceability. We also present the data model and system architecture for supply chain management in order to be applicable to our DTTM, so that companies can actually adapt their systems to obtain the full traceability.In this paper, we assumed the supply chain is consisted of four big tiers, namely supplier, manufacturer, wholesaler and retailer. We, especially, focus on the manufacturer tier where items or components are being merged and divided through their production process. The primary role of RFID technology in the supply chain is to provide automatic non-line-of-sight scanning. This means items do not require particular scanning works. RFID readers can scan multiple items simultaneously and this capability supports the automation of many SCM tasks that have been labor intensive. Companies will also have accurate information on stock levels which will help reduce inventory costs (Michael and McCathie, 2005). With proposed data model and system architecture, DTTM will act as core functionality in the existing enterprise applications to obtain the full traceability by extracting BOM information from ERP (Enterprise Resource Planning) and performing path connection activity. Without this DTTM, companies will only have huge amount of raw RFID data of items and fail to know which item is used for its parent item. To implement the DTTM in the existing enterprise applications, a simple path reasoning module and the BOM information are required.The rest of the paper is organized as follows. Related works are discussed in Section 2. Section 3 presents the characteristics of RFID data. Section 4 deals with an overview of traceability, followed by the details of the proposed model, DTTM and the system architecture in Sections 5 and 6. We conclude our study in Section 7.\n2 Related works: RFID technology has been extensively studied from several perspectives. Shih et al. (2005) discusses the privacy and security concerns of RFID technology and specifically provides the secure web services framework in the EPCglobal Network environment. Galanxhi-Janaqi (2004) talks about a so-called U-commerce (ubiquitous commerce) in which RFID technology is used as one of the means. RFID is also used for item identification method in automated purchasing.Another research area dealing with RFID technology is about constructing architecture for RFID-enabled system. Chang et al. (2007) presents the WSARE (web services-enabled Architecture for RFID Environment) which is not only based on RFID but also web services. Jankowska et al. (2007) introduces the architecture for agent-based mobile supply chain event management (SCEM). SCEM is an approach to fill the gap between supply chain planning and supply chain execution by the use of RFID technology and mobile computing. Those are somewhat close to our work since it proposes the architectures for RFID-based system but differs in that our architecture is mainly concerned with making the full traceability along the supply chain using RFID.RFID technology poses new challenges for data processing and management. By adopting RFID technology, companies will be confronted with a huge amount of data generated by RFID tags. Research dealing with RFID data management suggests several methods to clean and filter the RFID raw data and present a data model to process data effectively (Gonzalez et al., 2006a, b, c). Along with this approach, there are some works that deal with transforming the raw data into more valuable, meaningful data which will support management decisions. Complex event processing is one of these approaches that take RFID activities as an event and processes it with predefined logic (Wang et al., 2006; Zang and Fan, 2007). However, there were few studies related to the traceability based on RFID technology. DeVries (2008) deals with RFID track and trace in the airline industry and shows some examples of effective baggage tracking but it is without going into the detail at the level of subcomponent data structure. Our approach are designed to handle each individual item at every tier in the supply chain and showed the methodologies to dynamically track and trace in order to give the full view of item's history. Some of the related works above had parts similar to that of ours but none are close to our model.\n3 Characteristics of RFID data: Since the topic of the paper is RFID-based traceability and deals with RFID data, we pointed out the two critical characteristics of them. It is quite clear that these elements need to be fully considered in constructing our DTTM. The two characteristics are listed below:1. Dynamic. Not only the word dynamic means that data changes over time but it imposes the concept of the item being merged with other subparts and being divided into several components through the production process.2. Procedural. Items will leave a trail of data as they pass through several different locations. As an object is equipped with a tag, when it passes by the vicinity of a reader it will send information messages to the interrogator. This process will make the identification of a tag for the reader at every location creating a consistent path of the object.3.1 Path database\n4 Traceability outline: Traceability is the ability to trace the history, application or location of an entity, by means of recorded identifications (ISO, 1995). This is a general definition of traceability given by ISO. As stated in the definition, traceability requires a mean of identification and in this paper we assume RFID technology is fully-fledged to satisfy a complete method for identifying items. In order to trace the location and eventually the history of an entity, RFID readers should be placed in the area of interest along the supply chain including supplier tier, manufacturer tier, wholesaler tier and retailer tier. In the cases where end items are only of interest, the path database shown above will be sufficient to meet the requirements of traceability. However, in general, that is not sufficient to be capable of having an end-to-end traceability, because products are made up of one or more subparts having the history of themselves and if there are no data of a product configuration to connect each item's paths it will end up creating incomplete traceability. This is why we developed DTTM to make a complete traceability throughout the whole supply chain.If we look at the supply chain, we assume there are four tiers (supplier, manufacturer, wholesaler, and retailer). In supplier, wholesaler, and retailer tiers, almost all the products are transported in end items. But in manufacturer tier, items are combined and divided through the production process to create a final product (Figure 1). For this reason, in the former three tiers the existing path database will be used to provide the traceability and in the latter tier, the proposed DTTM will be applied so that it will give a full view of traceability (Figure 2).\n5 Dynamic tracing task model: In this section, we introduce DTTM to make the end-to-end traceability viable. It incorporates the concept of traditional BOM and is implemented with several modules supporting the dynamic tracing task. The bill of material is considered to be an engineering document that specifies the ingredients or subordinate components required physically to make each part number or assembly (Vollmann et al., 2005). And in the supply chain, especially in manufacturing process, objects usually combine and split to become a product. For this reason, companies store the BOM in their databases to know all the items that consist of a product (Figure 3).Although both the MRP (material requirements planning) and MRP II (Manufacturing Resource Planning) also incorporate the BOM in their system our DTTM is completely different from them in terms of its functionality. BOM in the MRP is used with start times and lot sizes in order to generate gross requirements of any required components. In the MRP II, the BOM is one of the core records so as to perform MRP. Since most companies have already implemented either MRP or ERP system meaning that the BOM data can be easily obtained, DTTM can simply exploit this BOM data to execute dynamic tracing task.5.1 BOM data transition\n6 Data model and system architecture: 6.1 Data model\n7 Conclusion: RFID technology gives several advantages for enterprises to effectively handle their daily operations. One of its benefits is the product traceability. If a product can be tracked and traced, product recall, product pedigree analysis, and consumer visibility can be improved and enhanced. However, few studies have introduced various methodologies to provide the full traceability of products. In this paper, we developed DTTM in order to make a complete traceability, by:* transforming BOM into path information sets;* creating path connection rules based on path information sets; and* connecting an item's path database to its proper parent item's path database.The cost of applying DTTM in the enterprise systems can be relatively low because it makes use of the existing system at the maximum and requires simple software materialized with the algorithm based on the proposed path connection rules. The data model and system framework are also presented to support the proposed DTTM. With our models and framework, the product traceability along the supply chain can now be formed completely. And with this full traceability, the expected benefits such as distribution channel transparency and fair trade can be obtained. But most of all, there will be several advantages in operations management that help solve the problems of matching supply with demand which is extremely difficult for managers as well as economists. Integrating our model to actual enterprise solutions along with the validation and verification is left for our future work.\n",
"role": "user"
},
{
"content": "Purpose: - The paper aims to propose a novel dynamic tracing task model to enhance the traceability range along the supply chain beyond simple distribution channels. It further extends the study by implementing the system architecture with the proposed data model to support the dynamic tracing task.\nDesign/methodology/approach: - Typical processes of supply chain in manufacturing industries using bill of material data to extract and define information requirements are followed. The data elements are systematically selected and explained in the proposed model step by step.\nFindings: - This paper found that existing Radio Frequency IDentification (RFID) data management scheme has to be modified so as to provide end-to-end traceability.\nResearch limitations/implications: - Validation of the proposed model and system architecture should be done through actual implementation in industrial settings.\nPractical implications: - The paper gives an insight to many system managers and executors in how full traceability along the supply chain can actually be realized.\nOriginality/value: - Several studies have discussed the traceability issues but failed to construct a system and methodologies to obtain the full traceability range. However, this paper contributes by presenting the specific methodological steps which can provide powerful solutions to build a system for traceability, exploiting RFID technology.\n",
"role": "assistant"
}
] |
Marketing-specific intellectual capital: conceptualization, scale development and empirical illustration
|
[
"Intellectual capital",
"Knowledge resources",
"Marketing",
"Scale development",
"Spain"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: According to the knowledge-based view (Grant, 1996; Spender, 1996), knowledge is a firm's most strategically important resource, since it is considered the cornerstone of competitive advantage and value creation. For that reason, managers need to focus on producing, acquiring, retaining and utilizing knowledge (Spender, 1996). An organization's combined knowledge resources constitute its intellectual capital (IC) (e.g. Nahapiet and Ghoshal, 1998; Sullivan, 1999; Youndt et al., 2004).However, even though some authors equate IC to knowledge, others consider that IC encompasses all kinds of intangible resources (e.g. Brooking, 1996; Edvinsson and Malone, 1997; Sveiby, 1997; Roos et al., 1998; Marr, 2006). Giving the prominent role of such resources in the generation of superior and unique returns for firms (due to their substantive inimitability), the early IC literature put great emphasis on taking stock of IC components and its constituents. Usually, a three-component classification has been suggested in which intangible and/or knowledge resources are sorted out according to where they reside: people (i.e. human capital (HC)), the organization (i.e. structural or organizational capital) or relationships (i.e. relational or social capital). Thus, IC components constitute designed or constructed conceptual variables with the aim to logically structure the intangible assets of the firm.After the effort made in inventorying and classifying IC, numerous studies have been carried out to analyze its influence on company performance (for a review, see Inkinen, 2015). Many of them set out from the IC categories previously identified and tried to prove the influence of each IC component on performance using \"constructs\" or \"proxy variables\" made up of several indicators that try to capture their content. Despite the designed or constructed nature of such components, common factor models (see Bagozzi, 2011) have mostly been applied, as if they were referring to actually existing unobservable variables that give rise to the indicators included in each subscale (i.e. human, structural/organizational and relational/social capital). However, when a conceptual variable is viewed as a combination of different elements (i.e. when such elements \"define\" the variable but do not cause it), a composite measurement model applies (Henseler, 2017).Beyond this general methodological challenge, the wide spectrum of elements that could be included within each IC component poses additional measurement problems. First, as the majority of existing studies analyze the influence of IC on performance with a single construct per IC component, a selection is made by authors about the specific elements to be included in each IC category, which affects the comparability of the results obtained. Second, when very heterogeneous aspects are included within the same construct, each of them tends to be represented by very generic and synthetic indicators. Thus, analyses on the IC-performance linkage based on such indicators provide very little information beyond highlighting the relevance of hiring bright and skilled people or that of investing in information systems and documenting knowledge. Moreover, very often the scales used (e.g. Bontis, 1997; Hsu and Fang, 2009) mix practices, resources and outcomes within the same construct, which implies that dependency relationships between such elements may also exist.Therefore, to improve the relevance and consistency in IC measurement, in the current study a proposition is made to concentrate clearly on knowledge resources (i.e. to exclude other intangible assets, practices and outcomes) and to take a more specific approach into the qualities of knowledge. As pointed out by Kianto et al. (2018), one of such qualities is contextuality: knowledge is always a contextual phenomenon, with strong local and institutional components, something that has been largely overlooked by the IC literature. Although since 1998 onwards IC-performance studies exist that have been carried out in different contexts (especially in terms of industry, ownership, company size and type of country), these studies do not seek to highlight the specific types of knowledge or intangible items that may be of particular interest in these contexts, but they rather aim to show the general relevance of IC by using generic scales as discussed previously.To overcome this research gap and considering that organizational units or functions constitute one of the less studied contexts in previous IC research, this study examines IC in the context of the marketing and sales function as a distinct but broadly relevant organizational domain. According to Porter (1985), while all of a business's functional areas contribute to the delivery of goods and services, marketing plays a key role in adding and creating value for customers. Additionally, the numerous and rapid changes in the marketing field have consequences for the knowledge resources needed for successful performance of this function. Marketing is therefore relevant in the present context in light of its key role in attracting and retaining customers and shaping innovation and value creation, all of which are essential for company survival (see for example Kotler and Armstrong, 2018).By distinguishing between human, organizational/structural and social/relational capital (RC) (Youndt et al., 2004) within the marketing context, a measurement scale will be proposed that provides a more fine-grained instrument for researchers assessing the influence of domain-specific knowledge resources on marketing capabilities and performance. This will in turn provide more tailored recommendations for marketing professionals in relation to managing IC. Marketing managers can also use the scale as a self-assessment tool to diagnose the knowledge-related strengths and weaknesses of the marketing function. More broadly, the development of a domain-specific IC scale serves to demonstrate the utility of more tailored but generalizable approaches to IC measurement.Next, a literature review will be presented on the conceptualization, categorization and specification of IC and its consequences for empirical research, including an overview of IC-performance studies in specific contexts. Research design will then be explained, whereupon knowledge resources that shape IC within the marketing function will be identified and classified. Subsequently, a scale to measure marketing-specific IC will be developed and validated before being empirically illustrated in a survey of 346 Spanish firms. Implications for research and practice will be discussed in the end.\n2. Theoretical background: 2.1 Conceptualization, categorization and specification of IC\n3. Research design: This section explains the rationale to develop an IC scale for the marketing function and the methods applied to validate and test it. The proposed scale could be applied to any type of company regardless of industry, size and geographical context, provided it has a marketing team (i.e. a group of people who work on marketing-related issues). Of course, certain company characteristics may influence the degree of development of some marketing-specific IC elements and categories, but any firm with a marketing team could use the scale for self-assessment purposes, and researchers could also use it to analyze the influence of marketing-related knowledge resources on organizational capabilities and performance with the appropriate controls. In such studies, firm characteristics could also be treated as moderating factors that may affect the investigated relationships. 3.1 Scale development\n4. Conceptualization, categorization and specification of marketing-specific IC: This section presents the scale developed to assess IC in the marketing function. To that end, the concept of marketing-specific IC is first presented. Following the knowledge view of IC, marketing-specific IC is defined as all of the available valuable marketing-specific knowledge resources that an organization manages in developing its marketing capabilities and achieving its marketing-related goals.This concept complements and extends the market orientation (MO) concept traditionally addressed in the marketing literature. From a market information processing perspective (e.g. Jaworski and Kohli, 1993; Hult et al., 2005), MO refers to the extent to which a company engages in the generation, dissemination and response to market intelligence pertaining to current and future customer needs, competitor strategies and actions, channel requirements and abilities and the broader business environment (Morgan et al., 2009). It involves customer orientation (i.e. understanding customers' needs and wants), competitor orientation (i.e. understanding rivals' strengths and weaknesses and how they are satisfying customers' needs and wants) and inter-functional coordination (e.g. Narver and Slater, 1990; Hult et al., 2005). Thus, marketing-specific IC could be conceived as the result of a firm's MO: because firms are market-oriented, they are able to generate marketing-specific knowledge resources. The identification and classification of such resources facilitates their subsequent management and monitoring, which increases the chances that MO will lead to superior performance.In the following, marketing-specific IC categories and subcategories are presented, together with their indicators. Figure 1 provides an overview of the measurement architecture. 4.1 Marketing-specific human capital\n5. Scale validation: As content and face validity were already verified before the survey research, only convergent validity remained to be tested at this phase. Table 3 shows the correlation between each composite and the corresponding summary indicator (i.e. items marked with an asterisk in Table 2), and Table 4 presents descriptive analyses and VIF values for each indicator. The decision to consider mode A or mode B composites for future analyses depended on the VIF values obtained.As shown in Table 3, all but one of the correlations are higher than 0.707; the exception is the correlation for human capital, market knowledge (HCMK), with a value of 0.692. As this value is extremely close to the threshold, it can be concluded that convergent validity is good enough in all composites. Moreover, as VIF values larger than 3 (see Table 4) were returned for human capital, customer knowledge (HCCK); human capital, product/service or technical knowledge (HCTK); structural capital, marketing-specific IT capital (SCIT) and all RC subcomponents, a mode A composite was proposed for these subcategories (see Table 3). Notice that differences in the value of N (i.e. the number of respondents) in Table 4 are due to the presence of some missing values (i.e. non-responses) in some of the items.\n6. Marketing-related IC in Spanish firms: Once the marketing-specific IC architecture had been validated, a diagnostic of the participating Spanish firms was carried out. Table 5 reports the descriptive statistics and correlations for the whole sample (in order to obtain a clearer and more synthetic picture, the average score of the items making up each IC subcategory was used). As can be observed, the most developed subcategories of marketing-specific IC are those related to HC, followed by RC and then SC. Regarding HC, product/service or technical knowledge exhibits the highest degree of development (5.75), followed by market knowledge (5.56), marketing skills (5.55) and customer knowledge (5.48). In the case of marketing-specific RC, customer capital is the most prominent subcategory (5.51), followed at some distance by other types of RC, with values that are closely bunched (around 5.1 in all cases). Finally, marketing-specific SC is some distance from the other blocks, with values of 4.77 for marketing-specific IT capital and 4.44 for marketing-specific organizational memory.Having gained this overall picture and because previous studies have shown that IC differs across firms of different types (e.g. Kianto et al., 2010; Buenechea-Elberdin, 2017), T-tests were conducted among the previously defined groups of firms to examine whether this also applies to marketing-specific IC (Table 6). The significant differences (p > 0.10) between groups are reported below.The first comparison examined differences between B2B and B2C companies. As marketing differs in B2B and B2C contexts (Kotler et al., 2006), it can be assumed that marketing-specific IC will also differ. The results show that, overall, marketing-specific IC is more developed in B2C than in B2B firms, with significant differences in several IC subcategories. There were differences in HC, where product/service or technical knowledge, as well as market knowledge, were more developed in B2C than in B2B firms (comparative scores were 5.93 vs 5.65 and 5.68 vs 5.48, respectively). Further, all SC subcomponents returned larger scores in B2C companies (4.97 vs 4.65 in the case of marketing-specific IT capital and 4.63 vs 4.32 in the case of other types of organizational memory). Regarding RC, values obtained for internal relational subcomponents were also significantly larger in B2C firms (5.32 vs 5.08 in the case of internal RC at the department level and 5.32 vs 5.01 in the case of internal RC at the inter-department level).The second comparison looked at the differences between manufacturing and service companies. Although, according to Kianto et al. (2010), IC differs between manufacturing and service firms, few differences were found within the marketing context. In relation to HC (marketing-related skills) and SC (marketing-specific organizational memory), service companies returned significantly larger values than manufacturing firms (comparative scores were 5.68 vs 5.44 and 4.59 vs 4.30, respectively).The third and final comparison explored differences between high-tech and low-tech companies. As others have shown (e.g. Saenz et al., 2017), IC configurations differ according to company technology intensity, and some significant differences also emerged in the marketing context. In terms of HC, marketing employees in low-tech companies exhibited higher product/service technical knowledge (5.82 vs 5.59), although their educational background and experience was lower than that of employees in high-tech firms (5.66 vs 5.43). Unsurprisingly, the level of marketing-specific IT is significantly higher in high-tech companies (4.84 vs 4.59).\n7. Discussion: In attempting to develop a model of IC with an improved managerial application, the present study developed and validated a knowledge resources measurement scale, which is specific for the domain of marketing. The marketing function is a useful context in which to develop such a scale in light of its distinctive but universally important organizational role.Previously, a literature review was carried out aimed at tracing back the conceptualization, categorization and specification of IC and the influence of such issues on empirical research analyzing the IC-performance linkage. An examination of such research revealed several challenges related to IC measurement. These include measurement model misspecification (i.e. using common factor models instead of composites); lack of comparability between studies due to differences in the breadth of content selected by different authors for similar IC categories; usage of very generic and synthetic indicators to grasp each of the multiple elements making up each IC component (which gives rise to very unspecific recommendations) and potential dependency relationships between elements within the same construct (especially when practices, resources and outcomes are mixed up within the same construct).Therefore, to improve relevance and consistency of IC measurement, the proposition was made to focus on the knowledge view of IC (i.e. to exclude other intangible resources, practices and outcomes) and to deepen into the qualities of knowledge, in particular, into knowledge contextuality. As already argued by Kianto et al. (2018), knowledge is always idiosyncratic and organization-specific, right down to the micro level of teams, functions and divisions. The present study adopted this micro-level view to contextualize IC within the marketing function. In the following, theoretical and practical contributions of the research are discussed in detail. 7.1 Theoretical contributions\n",
"role": "user"
},
{
"content": "Purpose: The aims of this paper are to identify and classify the knowledge resources that shape intellectual capital (IC) within the marketing function, to develop and validate a related scale and to demonstrate the scale's applicability in an empirical context.\nDesign/methodology/approach: A literature-based approach was adopted to identify and classify knowledge assets in the field of marketing. The new scale's content was then tested in a number of companies with different profiles. A subsequent survey of a representative sample of 346 Spanish firms sought to validate the scale and to assess those companies' marketing-related IC.\nFindings: The literature search provided the basis for a marketing-related IC architecture comprising three main categories, nine subcategories and eighty items whose validity was tested and confirmed. The survey revealed that marketing-specific human capital (HC) is the most developed knowledge resource in Spanish firms, followed by marketing-specific relational capital (RC), while marketing-specific structural capital (SC) is the least developed. Significant differences were also found among companies with different profiles (B2C vs B2B, high-tech vs low-tech and manufacturing vs services).\nOriginality/value: This study makes a valuable contribution to the IC literature as one of the first to deploy the general IC framework in a specific functional area (here: marketing and sales) for more meaningful and in-depth assessment of firm-specific knowledge resources.\n",
"role": "assistant"
}
] |
An approach to mastering the marketing mix
|
[
"Return on investment",
"Marketing mix",
"Mathematical modelling"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: With eyes fixed firmly on growth, CEOs are now acutely aware that their marketing investments are perhaps the last significant elements appearing on financial statements that lack clear links to revenues and profits. CFOs are in lockstep with CEOs, calling for results from the last few quarters' investments and limiting future spending if they don't like what they see. In fact, it is common for most C-suite executives to view marketing as a sinkhole full of investments with undocumented returns.\nBarriers to achieving marketing ROI: There is no shortage of opinions about how to determine marketing ROI. First, there is the question of which ROI numbers to use. For example, one recent poll found that marketers have 27 ways to define leads. There are few company-wide standards, let alone industry standards. It is not likely that any metrics that marketing may possess will be in line with the CEO's agenda. At the same time, the proliferation of customer \"touch points\" increases the number of data elements that must be monitored. In turn, the more data that is collected, the greater are management's expectations of being able to derive value from it. There is also a tendency to track data only by individual product lines or across a function, but not across several functions in the company.There is also the perennial mismatch between long-term marketing goals and shareholders' short-term payback targets.One other impediment is worth mentioning: it's what we call the \"low-hanging fruit problem.\" It is very typical for the marketing programs whose returns are more easily quantified to enter into a vicious cycle of ever-increasing funding, regardless of their business impact. Recently, though, some companies have found ways to isolate and quantify the factors that influence customer behavior. They are deploying new marketing science techniques to yield fact-based analyses that make it easier for managers to decide where to invest.These techniques fit into a hierarchical framework where the top levels focus on how well marketing investments stack up against other business investments. The next level down allows marketers to gauge one marketing investment against another within well-defined categories. For example, they might compare the ROI of direct mail versus telemarketing, or a corporate brand-building campaign versus several regional product advertisements. The levels enable crisp decisions about specific program trade-offs: a 60-second ad spot on local radio versus a 30-second one, for instance (see Figure 1).\nThree critical ROI techniques: In this article we describe three related techniques that correspond to the hierarchical framework. We will address the advantages and limitations of each technique and give three examples of where they are used in practice to deliver significant benefits. Drawn from econometric analysis, the techniques are now being used increasingly in business. Used separately - each for its own application - they can provide significant gains. Used together, they offer marketers a powerful advantage. The three techniques are described below.Structural equation modeling\nThe power of structural equation modeling: Company A, a leading online grocery provider, had focused mainly on corporate branding that communicated an \"ease of use\" message. Although useful when the category was new, that message had begun to seem undifferentiated as the category had grown. Yet Company A's advertising spend has tripled from the late 1990s, with almost three-quarters of that outlay dedicated to TV placements.The company's marketing, customer service, and product management operations had all grown up separately; the factors that influenced customers' selection of Company A were not clear to any group. Aware that the \"ease of use\" message was losing effectiveness but unable to pinpoint why, the grocer's marketing team launched a major research program to help define an overall portfolio of spending and to reveal the best trade-offs among different business spending elements (marketing versus service or product experience, for instance), as well as the trade-offs among different marketing investments (for example, word-of-mouth vs. TV, direct mail, and e-mail) (see Figure 2).Working with brand consultants from Lippincott Mercer, Company A's marketers first used structural equation modeling to map all the influences on the loyalty of their customers and on the likelihood that prospects would sign up with the company. The marketers then kicked off an extensive research effort, polling 3,500 customers and prospects and interviewing others in depth.Analyzing the findings, the company was surprised to learn just how important word-of-mouth was to how customers and prospects viewed Company A's value. In fact, TV ads - Company A's traditional vehicle for delivering its message - came in a very distant second. The return on a dollar spent on word-of-mouth activities was likely to be at least two-and-a-half times more effective than if that dollar were spent on TV advertising. And it was troubling to find that the company had very little positive reputation with non-customers.The realization that such influencing factors were not part of marketing's traditional responsibilities allowed Company A management to see the value of a \"whole-company\" approach to customer loyalty - a first for the company. In addition, the research revealed that although TV advertising was still seen as \"cool and exciting,\" it was not necessarily perceived as relevant, since the market had outgrown the \"ease of use\" messaging.The results of the research have convinced the company to take word-of-mouth seriously (see Figure 3). Company A has now shifted significant marketing resources to develop specific \"buzz-creating\" programs to leverage its strong positive word-of-mouth reputation and to reduce negative word-of-mouth among prospects. For the first time, it is aiming public relations efforts at consumers, because PR ranked higher than expected as an influence factor. The company has also recently modified its TV messaging to focus more on communicating specific benefits rather than on conveying the \"ease of use\" of online grocery services. The new TV messages have played out well in the marketplace. They are more relevant to prospects and customers and have increased the numbers of prospects who view the company favorably by nearly one-third.\nThe value of deep historical analysis: Company B, a leading maker of computer accessories, was spending about $300 million a year on advertising. It had three major advertising goals, each with significant challenges, as explained below.Corporate-wide brand advertising\nThe value of in-market experiments: Company C, a leading regional provider of home-maintenance products, has made good use of in-market experiments. Although the company was spending more than $200 million each year on marketing efforts in its consumer group, it was not using any sophisticated testing approaches to determine and improve ROI. In fact, Company C lacked both the resources for testing and a disciplined testing methodology that was efficient and scalable.As a result, good ideas were going untested, and marketing strategies were put into play based on individual opinions or anecdotal evidence rather than hard facts. As regional markets became even more competitive and marketing budgets were whittled down, Company C's marketing leaders welcomed any approaches that could improve ROI. The marketing team turned to the following two specific sets of experiments.Determining what makes a great sales call\nConclusion: Marketers now have an unprecedented opportunity to demonstrate the value of their initiatives. The analytical ROI framework helps them make sense of complex and seemingly chaotic marketing investment patterns and allows them to quickly reach compelling conclusions about future marketing commitments. It provides the impetus necessary to break away from \"always done it this way\" inertia and the fact base to push back against outside agencies - particularly advertising firms - that can be reluctant to accept quantitative measures of effectiveness.With new ROI techniques at hand, marketing leaders can confidently take a portfolio approach to their investments. This approach is relevant to spending beyond their traditional span of control, in functions such as customer service and product development. The new framework encourages, rather than confounds, collaboration across business functions.Perhaps most satisfying of all, marketers now have the wherewithal to silence critics who have long griped about marketing's \"squishy\" decisions.If anything, the pressure to rein in \"non-growth\" costs will only increase. There will always be barriers to gauging returns on marketing investments, but there are fewer excuses for not doing so. Leading companies are gaining a competitive edge by applying the new marketing ROI approaches. They are gaining experience with them, fine-tuning them, and gathering more valuable data with them. Competitors who do not take a keen interest in what the marketing leaders are doing may not be effective competitors a few years from now.\n",
"role": "user"
},
{
"content": "Purpose: - The goal of this paper is to use three condensed case studies to present an analytical ROI framework that helps marketers make sense of complex and seemingly chaotic marketing investment patterns and allows them to quickly reach conclusions about future marketing commitments.\nDesign/methodology/approach: - The article focuses on three marketing science techniques with roots in econometric analysis. The three techniques are: structural equation modeling, historical analyses, and in-market experiments.\nFindings: - The case studies give evidence of hard returns not only in terms of marketing funds conserved but also in terms of revenue gains, more positive brand equity, and greater marketing efficiency. The article suggests that changes in marketing practices to include the new ROI techniques will yield measurable benefits.\nPractical implications: - The article will be of value to marketing leaders, brand managers, and senior executives who seek quantifiable returns from their marketing functions. The three quantitative techniques described make sense of previously inscrutable marketing choices such as spending more on word-of-mouth campaigns than on national advertising or selecting one direct-mail approach over another.\nOriginality/value: - To date, marketing ROI work has been constrained by confusion over ROI definitions, by decisions about what data to collect, and by an emphasis on pricing, coupons, and promotions rather than broader marketing questions about, say, the efficacy of direct mail or the impact of regional advertising. The article reports on a fresh approach termed marketing science - techniques whose fact-based analyses make it easier for managers to decide where to invest.\n",
"role": "assistant"
}
] |
Effects of organizational learning on performance: the moderating roles of trust in leaders and organizational justice
|
[
"Organizational justice",
"Knowledge sharing",
"Organizational performance",
"Organizational learning",
"Trust in manager"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The concept of organizational learning (OL) has received significant attention over the past several decades. Given its the multi-disciplinary background (Easterby-Smith, 1997), the concept has been widely applied and broadly defined in research on processes of knowledge transfer (Huber, 1991; Nonaka, 1994), exploitation and exploration in organizational adaptation (March, 1991), learning capability (Chiva et al., 2007), communities of practice (Brown and Duguid, 1991), power and political processes (Lawrence et al., 2005), as well as interaction in multi-level learning (Crossan et al., 1999). While scholarly perspectives may differ, most definitions of OL emphasize the interaction between individual-, group- and organizational-level learning (Milia and Birdi, 2010) and pay particular attention to the flow of knowledge sharing across levels. That is, the way in which new knowledge flows from the individual to organizational level and the simultaneous flow of existing knowledge from the organization to individual level (Crossan et al., 1999; March, 1991). Multiple learning activities are produced by the social and psychological experiences of an organization's members and eventually lead to organizational change (Crossan et al., 2011).\nTheoretical framework: This research adapts several theories proposed in OL literature in three ways. First, this study shares the perspective of resource-based theory, which is premised on the claim that organizations rely on embedded knowledge resources to achieve their competitive advantage (Barney, 1991). Many OL studies support the notion that learning occurs at multiple levels within an organization, namely, at the individual-, group- and organizational-level (Crossan et al., 1999; Eraut and Hirsh, 2007; Jyothibabu et al., 2010) and regard their distinct learning outcomes as competitive resources for survival (Jerez-Gomez et al., 2005; Sabherwal and Becerra-Fernandez, 2003). Several scholars have argued that learning outcomes can be stored as specific types of resources at each level; in the form of tacit or explicit knowledge (Nonaka, 1994) at the individual-level, a shared mental model (Kim, 1993) at group-level and as a learning system and procedure (Nevis et al., 1995) embedded in the organizational-level. These knowledge resources - the so-called \"learning capability\" at each level - become a driving force, motivating the organization to learn when it has a common goal or needs to make a change. Moreover, knowledge resources can serve to drive or facilitate knowledge sharing activities - particularly in the utilization of existing knowledge or the exploration of new knowledge, according to organizational strategy (Garcia et al., 2003).\nLiterature review and hypotheses: Organizational learning: learning stocks and learning flows\nMethod: Participant and procedure\nResults: Table II presents each construct's mean, standard deviation, skewness and kurtoses coefficient and correlations coefficient. Most learning constructs are statistically correlated with organizational justice, trust in management and perceived performance. The mean for organizational performance (4.00 out of 6) is relatively higher than others. Moreover, most informants perceive that knowledge accumulation (3.72) occurs more often than knowledge sharing (feed-forward 3.57 and feedback 3.60) at their organizations, and that trust in management (3.71) is more prevalent than trust in the organization (3.37).\nConclusion and implication: This study used Crossan et al.'s SLAM model (Bontis et al., 2002; Crossan et al., 1999) as an OL process to examine the relationships between learning stocks, their transfer (feed-forward and feedback learning flows) and perceived organizational performance. This study also hypothesized that greater levels of organizational justice or trust in management would result in a better association between learning stocks and organizational performance through feed-forward and feedback knowledge transfers. This study started to answer the question of what motivates people and groups in organizations to share or hoard their knowledge. In considering this question, this study was designed to identify whether trust in organizational systems and management plays a moderating role in the mediated relationship of learning flows between learning stocks and organizational performance from a critical OL perspective. The findings of this research make the following contributions.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to identify how organizational learning processes influence perceived organizational performance and examine the moderating roles of organizational justice and trust in managers therein.\nDesign/methodology/approach: This study develops a theoretical model to exhibit how knowledge acquisition and transfer activities influence perceived organizational performance. Data were collected from 515 respondents and analyzed by PROCESS macro for SPSS.\nFindings: This study found that feedback learning flows are strongly mediated between learning stocks and organizational performance. It also found that organizational justice moderates the effect of learning stocks on organizational performance through feed-forward learning flows, while trust in manager moderates the effect of learning stocks on organizational performance through feedback learning flows.\nResearch limitations/implications: This study has a limitation in which it uses self-report data to measure all constructs. The objective measure may be necessary for future study.\nPractical implications: The implications of this study are twofold. First, it finds that the higher organizational justice, the better the transfer of knowledge from the bottom up. When firms need to explore new knowledge, fairness in procedure and the distribution system is critical. Second, the higher the trust in management, the better the transfer of knowledge from the top down. The role of managers is instrumental in persuading employees and disseminating knowledge.\nOriginality/value: Combining functionalist and critical perspectives and developing a theoretical model, this study contributes to the understanding of how trust and justice facilitate learning activities within organizations.\n",
"role": "assistant"
}
] |
Value creation through big data application process management: the case of the oil and gas industry
|
[
"Big data",
"Trends",
"Challenges",
"Oil and gas",
"Value creation",
"Knowledge creation",
"Business process",
"Knowledge implementation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Rapid advances in information technology and process automation have enabled companies to generate massive amounts of data, widely termed \"big data\". In practice, big data is a key factor to searching and understanding the dynamics of the market and is a source of competitive advantage (Pauleen and Wang, 2017; Dezi et al., 2018). The \"implementation of big data\" suggests gaining knowledge, improving learning and making sense of data (Abbasi et al., 2016). According to recent empirical studies (Rialti et al., 2018; Sumbal et al., 2017; Khan and Vorley, 2017; Ferraris et al., 2018; Santoro et al., 2018), organisations are paying attention to big data. Further, studies (Rialti et al., 2018; Sumbal et al., 2017; Khan and Vorley, 2017; Ferraris et al., 2018; Santoro et al., 2018) have emphasised that organisations can gain maximum value from big data by experimenting with it to improve performance and replace human decision-making with automated algorithms (Sumbal et al., 2017). Nonetheless, big data research is still at the embryonic stage (Gupta and George, 2016; Intezari and Gressel, 2017; Rothberg and Erickson, 2017), and the big data phenomenon has spurred academic debate in recent years.\n2. Theoretical background: The term \"big data\" has taken the world by storm. It has gained momentum and become a hot topic in diverse industries such as social media, corporate management, health care, business and governance. Big data differs from traditional data's big Vs (LaBrie et al., 2017). It has historically been known by 3 Vs: velocity, variety and volume (LaBrie et al., 2017). Recently, a fourth V, veracity has been added (Schroeck et al., 2012). Volume refers to the scale of the data, ranging from gigabytes to petabytes. It requires new solutions to enable hardware and software to handle so much data. Velocity refers to the speed at which these large volumes of data are being generated. Variety indicates the various types of data such as documents, videos, images and numbers, in structured and unstructured formats. Finally, veracity refers to the uncertain nature of data in terms of its trustworthiness and reliability.\n3. Methodology: Based on the unexplored nature of the phenomenon under investigation and the unavailability of secondary data, this study adopted a qualitative research design and inductive research approach. We used theoretical and purposive sampling techniques (Miles and Huberman, 1994) to obtain detailed explanations.\n4. Results and analysis: This section discusses the results obtained from the collected data. These results have been categorised according to different themes developed during the coding process.\n5. Issues and challenges with big data implementation: This section discusses the main issues and challenges revealed during the interviews on implementing big data initiatives.\n6. Discussion and conclusion: 6.1 Discussion and comparison with existing studies\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is twofold: to investigate the role of big data in firms' co-knowledge and value creation and to understand the underlying drivers behind value creation through big data in the oil and gas industry by underscoring the role of firms' capabilities, trends and challenges.\nDesign/methodology/approach: Following an inductive approach, semi-structured interviews were conducted with senior managers and analysts working in oil and gas companies across eight countries. The data collected from these key informants were then analysed using the qualitative data analysis software ATLAS.ti.\nFindings: Value creation through big data is an important factor for enhancing performance. It has a positive impact on both tangible (organisational performance) and intangible (societal) aspects depending on the context. Oil and gas companies understand the importance of big data to creating value in their operations. However, implementing and using big data has been problematic. In this study, a framework was developed to show that factors such as the shortage of data experts, poor data quality, the risk of cyber-attacks and unsupportive organisational cultures impede its implementation and utilisation.\nResearch limitations/implications: The findings from this study have implications for managers and executives implementing big data and creating value across various data-intensive industries. The research findings, are contextual, however, and should be applied cautiously.\nOriginality/value: This study contributes to the value creation literature in the big data context. The findings identify the key areas to be considered for the effective implementation and utilisation of big data in the oil and gas sector. This study addresses a broad but under-explored issue (i.e. knowledge creation from big data and its implementation) and strengthens the academic debate within this research stream.\n",
"role": "assistant"
}
] |
Doing more with less: An analysis of work life balance among layoff survivors
|
[
"Job satisfaction",
"Quality of life",
"Stress",
"Downsizing",
"Family roles"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Research on the role of work-family issues in understanding career experiences has experienced major growth in the past few decades. A major review of the relationship between work and non-work appeared in 1980 and provided guidance for developing theoretically grounded research on this topic (Near et al., 1980). A few years later, an influential article appeared outlining the major sources of conflict between work and family roles (Greenhaus and Beutell, 1985). After publication of this article, research on work-family issues steadily increased. Today a plethora of research has led to increased understanding about the interface between work and family roles (see Eby et al., 2005). Despite this large body of research, recent reviews of the methodologies used in work-family research suggest that work-family research tends to use homogenous samples and often fails to examine special segments of the population that may be particularly prone to work-family conflicts (Casper et al., 2007). The current study examines work-family issues among one of these special segments of the population, employees who survive layoffs.\nTheory and research on layoffs: The frequency with which today's employees are confronted with downsizing is staggering (Koretz, 2001; Worrall et al., 2004). Recent instances of large scale downsizings include firms such as Hewlett Packard, where 14,500 jobs, or 10 percent of the workforce is estimated to be cut, and Sony, where 10,000 individuals will lose their jobs (Burrow and Gagnier, 2005; Levy, 2005). Such events have made understanding the experiences of those who survive such downsizing extremely important. Accordingly, substantial literature examines layoff survivors who remain with the organization (Brockner et al., 2004; Kernan and Hanges, 2002).Research has examined the \"survivor syndrome\", a term coined by David M. Noer (Fong and Kleiner, 2004) to describe the layoff survivors' lost confidence in management, and the anger, insecurity, guilt, frustration, perceived unfairness, and depression that are associated with remaining in an organization post-layoff (Baruch and Hind, 2000; Sahdev, 2004). In trying to understand the negative consequences of downsizing for survivors, Fong and Kleiner (2004, p. 10) note that \"one of the effects of downsizing is the exacerbation of work overload\" for those who remain in the organization. Evidence tends to support the notion that employees who remain with an organization after a layoff find themselves working harder (McGinn et al., 2001; Worrall and Cooper, 2001). A study by the Families and Work Institute (Galinksy et al., 2001) found that 42 percent of workers whose employers had cut staff in the last year reported high levels of overwork.While there is evidence that downsizing survivors face increased workloads (Kozlowski et al., 1993), little research has examined whether high workloads of layoff survivors are related to work-life balance and work and personal attitudes (Galinksy et al., 2001). This is an important oversight because if survivors experience increased work-nonwork role conflict, such conflict can impact performance outcomes such as turnover and absenteeism (Kossek and Ozeki, 1998).Role theory and layoffs\nMethod: Sample and procedures\nResults: Table I presents the means, standard deviations, intercorrelations, and reliabilities of the measures used in the study. Mean substitution was used to deal with missing data. Although this technique underestimates the covariances, it is recommended when less than 10 percent of the data are missing (Donner, 1982), and was adopted here since there was less than 3 percent missing data. Although correlations are provided in Table I, the hypotheses were tested examining the standardized regression weights from structural equation modeling rather than correlation coefficients. This decision was made because the regression weights provide an estimate of the bivariate relationship after controlling for other model variables, and thus, provide a \"cleaner\" picture of the true relationship.Examination of Table I reveals that layoff survivors perceived an increase in their workload (5.33 on a seven point scale). This is consistent with our expectation that layoff survivors would be faced with a greater workload after the layoff due to absorbing additional tasks into their job that were formerly performed by employees who were laid off.Since the measures were obtained from self-reports, we conducted the Harman's one-factor test (Harman, 1967; Podsakoff and Organ, 1986) to detect the presence of common method variance. Podsakoff et al. (2003) note that common method variance is indicated if a single factor emerges from the factor analysis or if one general factor accounts for a majority of the variance. The results of the unrotated factor analysis indicated that we had five factors, the same number as included in our model. All the five factors had eigenvalues greater than 1. The first factor accounted for 33.87 percent of the variance, and the remaining four factors accounted for 40.55 percent of the variance. Since we obtained five factors and the combined variance of the four factors was greater than that of the first factor, it is likely that relationships exist between the variables in the model which are due to more than a common method factor.Structural equation modeling using AMOS 6.0 with maximum likelihood estimation was used for model testing. Prior to testing the structural model, we ran a confirmatory factor analysis to assess the fit of the measurement model. Although results showed that the chi-square was significant (kh2 (125)=520.21, p<0.01), because this statistic is sensitive to high power resulting from large samples as in this study, other fit indices were used to assess the fit of the measurement model including the goodness of fit index (GFI; Joreskog and Sorbom, 1994), comparative fit index (CFI; Bentler, 1990), the Tucker-Lewis index (TLI; Tucker and Lewis, 1973), and the normed fit index (NFI; Bentler and Bonnett, 1980). Examination of these indices revealed that they were all above the recommended level of 0.90 (GFI=0.90; NFI=0.91; TLI=0.91; CFI=0.93). The root mean square error of approximation (RMSEA; Steiger and Lind, 1980), which should be below 0.08, was 0.079. Moreover, an alternative measurement model was tested in which all items loaded on a single factor (kh2 (153)=5647.56, p<0.01). The chi-square difference test indicated that the fit of the hypothesized measurement model was significantly better than the fit of the alternative measurement model (Dkh2 (28)=5,127.4, p<0.01), supporting our hypothesized measurement model.After the measurement model was tested, the structural model depicted in Figure 1 was tested. Although chi-square was significant (kh2 (128)=532.39, p<0.01), all the other fit indices suggested a good fit (GFI=0.90; NFI=0.91; TLI=0.91; CFI=0.93; RMSEA=0.079). Howeh Examination of the parameter estimates, presented in Figure 2, revealed that all of the relationships were in the predicted direction and all of the paths were significant except for the direct path between role overload and job satisfaction. The squared multiple correlations indicated that the model explained 22 percent of the variance in role overload, 37 percent of the variance in work-life balance, 11 percent of the variance in job satisfaction, and 33 percent of the variance in life satisfaction.Alternative model testing\nDiscussion: This study contributes to our knowledge by examining the work-life experiences of layoff survivors. Our findings regarding specific model links confirm most of our hypotheses. First, we found that workers that remain employed after a layoff perceive an increase in their work demands, which is associated with greater role overload. Second, we found that role overload was linked with important attitudinal outcomes such as work-life balance, job satisfaction, and life satisfaction, and that the effects of role overload on satisfaction outcomes were mediated by work-life balance. Finally, and unexpectedly, once work-life balance was entered in the model, the relationship between role overload and life satisfaction became positive. In the section which follows, each of these findings is discussed along with their practical implications for organizations and directions for future research.Our findings are consistent with previous research which has suggested that survivors are faced with additional work responsibilities in a post layoff situation (Mishra et al., 1998; Moen and Sweet, 2004). In the current study, survivors did perceive an increase in their workload post layoff and this increased workload was significantly related to role overload. This suggests that organizations should carefully consider survivor experiences after a layoff because increased workloads and the stress associated with them may be problematic for these employees. Some recent literature found that an overemphasis in considering the treatment of \"leavers\" during layoffs in contrast to \"stayers\" may be detrimental to the organization because it can make leaving the organization more attractive than staying (Sahdev, 2004). Greater focus on the redistribution of job responsibilities may be critical in ensuring that the increased responsibilities are still manageable by survivors. Work redesign during downsizing may be useful (Chadwick et al., 2004). However, Sahdev (2004, p. 193) cautions that \"organization redesign must result in process simplification and not process complication in order to avoid survivors feeling overwhelmed by the aftermath of downsizing\".Second, our findings also suggest that the role overload experienced by layoff survivors is related to other important attitudinal outcomes. Specifically, role overload was related to work-life balance, consistent with past research (Frone et al., 1997), and work-life balance mediated the relationship between role overload and job satisfaction. However, the role overload-job satisfaction relationship was fully rather than partially mediated by work-life balance. These findings support the notion that survivor role overload and impaired work-life balance that may occur after a layoff are associated with important attitudes such as job satisfaction which, in turn, have been linked to important outcomes such as organizational citizenship behavior (Organ and Ryan, 1995) and absenteeism (Hardy et al., 2003). The fact that work-life balance is a key mediating mechanism through which these effects occur, suggests steps that could be taken by organizations to ameliorate the deleterious effects of role overload among layoff survivors. For instance, organizations might assist layoff survivors in managing work-life issues that arise from increased workloads by allowing flexible work arrangements such as telecommuting or flextime. Such an approach may be practical from several perspectives. First, because such flexible work arrangements are related to enhanced work attitudes (Baltes et al., 1999), they may offset the negative effects of role overload on job satisfaction. Second, they may enhance job satisfaction through increasing work-life balance. Third, flexible schedules provide greater control to employees, and research has found that greater control during downsizing is related to higher job satisfaction (Parker et al., 1997). Finally, because companies that have recently downsized are likely to have limited financial resources to provide work-life benefits to remaining employees, flexible schedules represent a mechanism by which organizations can support work-life balance with little or no expense.Work-life balance also partially mediated the role overload-life satisfaction relationship as hypothesized. However, unexpectedly, the relationship between role overload and life satisfaction changed from a negative bivariate correlation to a positive regression coefficient once work-life balance was included as a mediator in the model. Thus, after accounting for the shared variance between work-life balance and role overload in predicting life satisfaction, the unique variance associated with role overload was actually positively related to life satisfaction. This is not surprising given the fairly high multicollinearity between role overload and work-life balance. Still, the positive relationship between role overload and life satisfaction (after controlling for work-life balance) is counter-intuitive and warrants consideration. Several possible explanations for this effect exist. First, it is possible that the unique variance associated with role overload may tap into an aspect of increased work demands that makes survivors feel needed at work, given the variety of tasks they can perform and the fact that they may be the only remaining member of the organization that can perform some of those tasks. Given the organization in which this study was conducted had experienced multiple waves of layoffs, if greater demands (associated with role overload) promote some sense of job security among survivors, this could also enhance life satisfaction (Lim, 1996).Second, it is also possible that although the construct of role overload aims to assess the negative overwhelming experience of multiple roles, it also taps into other aspects of multiple roles, which can also be positive. For instance, in addition to perceiving added responsibilities as increasing work demands, survivors may also perceive these added responsibilities to enrich their jobs and enhance their skills. Job characteristics theory suggests that when job responsibilities are enriched to include more task variety and use more skills, employees experience this work as more meaningful and are more satisfied (Hackman and Oldham, 1976). Research supports the basic premise of this theory (Fried and Ferris, 1987). Thus, after controlling for the negative impact of role overload through the shared variance with work-life balance, the remaining variance of role overload may have tapped into the enriched aspect of the job which would be positively related to life satisfaction.Our course, these suggestions are speculative and should be confirmed with future research. Future research might examine whether increased job responsibilities survivors take on after a layoff result in true job enrichment or merely job enlargement. This is an important consideration, given research suggesting that job enrichment is related to motivation, satisfaction, and performance (Campion and McLelland, 1993). Kernan and Hanges (2002, p. 926) note that \"enriched jobs often lead to increases in role-related variables such as role ambiguity and role overload, at least in the short term.\". As such, individuals may have different reactions to being a layoff survivor depending on whether they see the reorganization in work as enlargement or enrichment. Research suggests that management and non-management employees may differ in how they perceive additional job duties, with management employees viewing them as enriching, and non-management employees viewing them as merely enlargement (Tombaugh and White, 1990). Campion et al. (2005) also warn of \"job engorgement\" during downsizings, a situation of excessive job enrichment, in which individuals may feel that they do not have the knowledge, skills, or abilities to perform the job. Such excessive enrichment may have deleterious effects. Companies should therefore try to provide moderate levels of job enrichment in conjunction with resources to help employees \"manage\" role overload, so that added responsibilities are perceived favorably.Limitations and directions for future research\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this research is to examine how increased work overload of layoff survivors relates to their work-life balance and job and life satisfaction.\nDesign/methodology/approach: - Survey methodology was used to collect data from 510 layoff survivors in a high tech company. Regression analyses and structural equation modeling were used to test the hypotheses.\nFindings: - The study found that layoff survivors experience higher levels of workload which impact overall role overload that negatively affects work-life balance. Findings suggest that high workloads experienced by layoff survivors contribute to reduced job and life satisfaction through reduced work-life balance as a mediating mechanism.\nResearch limitations/implications: - The data used in this paper is cross-sectional and conducted within a single organization. Also, most of the data is obtained from self report survey data and subject to common method bias. As such, longitudinal studies are recommended for future research.\nOriginality/value: - This study makes a contribution by joining two distinct research streams - the job loss literature with research on work-family issues. Findings suggest that high workloads experienced by layoff survivors contribute to reduced job and life satisfaction with work-life balance acting as a mediator. Future research should determine whether these findings generalize to diverse layoff survivors in distinct industries, and assess whether these phenomena change over time.\n",
"role": "assistant"
}
] |
Guest editorial: a new dawn for qualitative service research
|
[
"Interviews",
"Qualitative research",
"Ethnography",
"Case study research"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Service research has grown in size and importance over the last 30 years and developed from a subfield in marketing to a global, interdisciplinary research field. Since its launch in 1987, the Journal of Services Marketing (JSM) has provided a fresh perspective on service theory and practice. Reflecting on the history of the journal, Martin (2012, p. 3) noted that \"the journal has evolved as the field of services marketing has evolved - from many conceptual, 'how to' and idea articles to those more empirically-based and theory-driven.\" This suggests that the research methodology used in service research to address marketing problems evolved because the research field matured and marketing problems changed. However, what research methodology is best suited to address present problems in service marketing?\nWhat is qualitative research?: Bansal and Corley (2011) suggested that qualitative research is scholarship that primarily relies on qualitative data and inductive theorizing. Qualitative data refer to non-reduceable text, including words and visuals, which can be digitized, synthesized and even counted once the data have been interpreted to discern patterns and insights (Bansal and Corley, 2011). Qualitative research surfaces new insights that can lead to developing new theoretical models and even new directions of research. When adopting qualitative research methods, researchers draw on observations from the data to introduce abstract knowledge. Inductive theorizing based on qualitative data is particularly appropriate in new or emerging empirical contexts for which existing research is relatively scarce. However, there are several different views of what qualitative research is and how it should be performed. Researchers (including interested readers and reviewers) might even confuse different types of research that seem similar because they use qualitative data, and these researchers may be disappointed if the resulting research does not match their understanding of qualitative research (Eisenhardt and Graebner, 2007). Alam (2005) noted that qualitative research in marketing is among the least understood and most criticized research methods today, often criticized more for how researchers have used them and less for flaws in their nature.\nAlternative qualitative research methods: A plethora of qualitative research methods are used in service research. The most common ones are interview studies and case studies; however, both ethnographies and search are becoming more common. Highlighting that qualitative research is not one type of study but that qualitative studies actually differ in their origin, assumptions and activities is key for performing and evaluating qualitative research. Subsequently, we will briefly discuss the most common qualitative research methods (Table I).\nWhy is qualitative research scarce?: In this special issue, Valtakoski (2020) confirmed that the share of qualitative research in the Journal of Services Marketing is approximately 7.5 per cent and that this share only has slightly increased over time. This finding supports previous research on the shares of qualitative research in the European Journal of Marketing, the Journal of Marketing, and JSM, which have consistently been between 5 and 10 per cent of published research (Hanson and Grimmer, 2007). It is possible to argue that marketing as a field heavily relies on a single paradigm (logical positivism) and has developed certain inherent methods biases (Deshpande, 1983), which may result in developing new theoretical contributions using methodologies more appropriate for deduction compared to induction. These, however, are possible limitations that we cannot influence in the short term. Instead, we would like to highlight aspects of qualitative research that can be improved to increase success in the publication process.\nInterest, rigor, relevance and richness: In this section, we discuss four characteristics of qualitative research: interest, rigor, relevance and richness. We argue that these characteristics are important for performing and publishing excellent research in service marketing. However, what is meant by interest, rigor, relevance and richness can vary amongst scholars with the terms meaning different things to different people.\nContributions to the special issue: This special issue consists of ten studies that address qualitative research methods in service research and that intend to make a methodological contribution to the field. The manuscripts can be grouped into four areas:\nDiscussion: This special issue attempts to initiate a discussion if there is a new dawn for qualitative service research. We do not necessarily say there should be more or less qualitative service research, but argue that the qualitative service research that is performed is evaluated based on its own merits. It is not the research methodology that decides whether a manuscript is cited or not, but using the wrong methodology to address a particular research question can kill the best research idea.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to highlight the role of qualitative research in service research. This study discusses what qualitative research is, what role it has in service research and what interest, rigor, relevance and richness mean for qualitative service research.\nDesign/methodology/approach: This study examines the most common qualitative research methods and discusses interest, rigor, relevance and richness as key characteristics of qualitative research. The manuscripts in the special issue are introduced and categorized based on their contributions to service research.\nFindings: The findings suggest that the amount of research using qualitative research methods has remained stable over the last 30 years. An increased focus on transparency and traceability is important for improving the perceived rigor of qualitative service research.\nOriginality/value: This special issue is the first issue that is explicitly devoted to the qualitative research methodology in service research. In particular, the issue seeks to contribute to a better use and application of qualitative research methodology.\n",
"role": "assistant"
}
] |
In the margins of training and learning
|
[
"Workplace training",
"Workplace learning",
"Social class",
"Work process",
"Learning society",
"Marginalization"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Continuing work-related training and the skill development of the labour force are seen as a prerequisite for successful business and a prosperous economy in advanced societies. Through its many policy documents and programmes, the European Union (EU) has urged the member states to implement more efficient and attractive lifelong learning policies and practices to widen the access to learning opportunities for their adult populations (European Commission, 2000, 2001, 2006, 2007). The differences in the adult education policies and systems as well as their participation rates are substantial between the EU countries (Hefler and Markowitsch, 2010; Rubenson and Desjardins, 2009). Participation rates vary between less than 10 per cent and approximately 70 per cent. Statistics show that the highest participation rates in adult education and training are found in the Nordic countries, along with Luxembourg and The Netherlands (Cedefop, 2015; OECD, 2015). In all of these countries, more than 50 per cent of the adults aged 25 to 64 years participate in adult education yearly. Adults participate primarily in work-related learning activities, fully or partly financed by their employers. Only a small minority of adults participate in work-related learning activities without support provided by an employer. Therefore, most of the training that adults participate in is employment related; moreover, training can also be seen as a type of a fringe benefit for the personnel of a company (Silvennoinen and Lindberg, 2016).\nThe \"vicious circle\" of the training margin: Much inspired by Labor and Monopoly Capital: The Degradation of Work in the Twentieth Century, by Braverman (1974), the nature and direction of change in work processes have been under intensive debate and research. According to Braverman, technological development and automation, along with organisational changes, impoverish work and restrict employee autonomy (deskilling thesis). Employers have a strong motive to compartmentalise and simplify work processes so that production is not dependent on the scarce skills of the employees. To achieve an uninterrupted production process, it is important that workers are easily exchangeable and replaceable. Deskilling of tasks is not bounded to manufacturing industries only. The examples from McJobs (after McDonald's) and call centres embody much of what Braverman wrote about: work will become impoverished, requiring fewer and fewer skills, autonomy and discretion. Although this impoverishment has undoubtedly been observed, studies do not prove deskilling as the only work process trend (Gallie, 1991; Green, 2006). Research produces increasingly more evidence of work polarisation: some jobs become more meaningful and independent, while others turn into tightly monitored and controlled routines (Goos and Manning, 2003, 2007; Goos et al., 2009; Felstead et al., 2002, 2004; Oesch and Rodriguez Menes, 2011).\nResearch questions, data and methods: In the empirical section of this article, we examine a type of reverse side to the present day \"professional hype\". Then, we examine the employees whose learning and training opportunities fail to match the positive visions often associated with knowledge work and other information society jobs. The following questions will be answered using large-scale Finnish survey data:\nResults: The training margin\nConclusion and discussion: In the above, although we investigated those excluded from the training and learning opportunities at work, we tried to frame the findings as part of a larger phenomenon. For example, the results shown in Tables III-V clearly indicate that employment training and learning margins are not exclusively about training. Missing opportunities for learning and personal development indicate many other undesirable aspects of employment: hard, un-stimulating and uncertain work, minor task discretion and autonomy at work, poor access to the management information flow and the relationship between management and staff is not trusting and open. Primarily, all of these features depend on employment structures, not on the lack of willingness of an individual to increase his or her autonomy or learning opportunities at work. Meaningful work and good learning opportunities feed the willingness to learn and the motivation to train oneself. Correspondingly, un-stimulating work keeps the motivation for training low. Low-skilled, low-paid workers primarily live in a culture in which education and learning are not valued as highly as in the middle classes (Bourdieu, 1984; Jarvis, 2012).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the distribution of learning and training opportunities in the Finnish workforce. It will concentrate on the sectors of the workforce that lack these opportunities.\nDesign/methodology/approach: The Working Conditions Barometer (WCB) data from 2008, 2009 and 2010 are used (N = 3,326) in this investigation. The focus of the study is to examine those employees whose learning and training opportunities fail to match the positive visions of the current \"super professional hype\". The problems associated with the workforce training margin are considered as a part of the so-called work process debate.\nFindings: Findings suggest that age, gender, type of industry and socioeconomic status have an independent influence on an individual's presence in the training margin: women, older employees, employees in manufacturing and the private service sector and manual workers have a greater risk than others do. The relative differences between the social classes are still prevalent: incapacity and marginalization are primarily working class problems.\nPractical implications: The workers should collectively demand certain benefits more forcefully, even when working under fixed-term contracts. Employers should be pressured to organise development opportunities and training for the fixed-term employees, including the type of training that not only benefits an employee in that one company but also the employees' working life in general. Employers should be persuaded to organise the type of training that improves the chances of a fixed-term or otherwise powerless employee to find a new, perhaps better and more secure job.\nOriginality/value: Because the percentage of response to the WCB is exceptionally high, 80 per cent on average, the information obtained from the random sample can be generalised to represent all of the working wage earners in Finland and in a European context as well.\n",
"role": "assistant"
}
] |
Retaliation against reporters of unequal treatment : Failing employee protection in The Netherlands
|
[
"The Netherlands",
"Employment legislation",
"Discrimination",
"Workplace",
"Retaliation",
"Social protection"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The terms unequal treatment and discrimination refer to practices in which people are treated unequally on the basis of inappropriate grounds, such as race, gender or age. This paper reports on an empirical investigation of retaliation against people who experienced unequal treatment at work and decided to address this wrongdoing. We start with an illustrative example to set the scene.A woman of Surinamese origin is almost sixty years of age when she finds a financial accounting job in a foundation in the field of international aid and equal opportunity. Her expectations are high, but she is in for a sobering experience. After some time, she discovers that she is being paid less than her colleagues, that she is denied a tenured position for a longer time than is usual and, that she is denied a well-deserved promotion. The first time she addresses these issues, she is simply ignored. When she then claims her right to an assessment by an internal committee, the director uses his powers to prevent this, and she starts getting negative remarks about the quality of her work. She then files a formal complaint to the Dutch Equal Treatment Commission and gets confirmation of what she suspected. The commission investigates her complaint and concludes that it is well founded. She has indeed been treated unequally on the grounds of race and age. Does this help her? Not really. Soon afterwards, the foundation reorganizes and she is the only employee to lose her job, just a few months before she can regularly retire. As a result, her pension is now considerably lower than it would have been.The above example illustrates employer retaliation in response to an employee who reported unequal treatment at work, but leaves many questions unanswered. How often does this type of retaliation occur? Can we speak of incidents or does retaliation against unequal treatment complaints pose a structural problem? And, which factors contribute to this type of retaliation?In the Netherlands, the principle of equality forbids discrimination of persons and groups on the grounds of gender, race, age, nationality, religion, belief, political opinion, sexual orientation, marital status, handicap or chronic illness and type of labor contract. The principle of equality has a long history and is laid down in the first article of the Dutch constitution and firmly embedded in several national laws, such as the General Equal Treatment Act and the Equal Treatment in Employment Act. Furthermore, the right to equal treatment at work is established in several international and European treaties, including the European Convention on Human Rights, the International Convention on Civil and Political Rights and the United Nations Conventions on Elimination of all forms of Racial Discrimination and Discrimination Against Women.In 1994, the Dutch Equal Treatment Commission, a semi-judiciary body, was established, with the tasks of promoting and monitoring compliance with this equal treatment legislation, and of investigating unequal treatment disputes and providing judgments (these are judgments which are not legally binding). In 2012, these tasks of the Equal Treatment Commission have been transferred onto the newly erected Netherlands Institute for Human Rights.All these formal arrangements, however, do not guarantee that discrimination is effectively banned from the Netherlands or that employees can simply report discrimination to make it stop. As scholarly research points out, an important hindrance to achieving equality in practice is retaliation (Bergman et al., 2002; Brake, 2005; Cortina, 2008; Cortina and Magley, 2003; Cortina et al., 2001; Lee et al., 2004; Mesmer-Magnus and Viswesvaran, 2005; Miceli and Near, 2002; Miceli et al., 1999; Near and Miceli, 1986, 1996; Near et al., 2004; Miles et al., 2010; Parmerlee et al., 1982; Sincoff et al., 2006; Skiveness and Trygstad, 2010). Retaliation, and in particular fear of retaliation silences challenges to discrimination, and thereby hinders the realization of equal treatment in the workplace (Bergman et al., 2002; Cortina and Magley, 2003; Miles et al., 2010). For these reasons, retaliation against reporters of unequal treatment is forbidden in many countries. This is also the case in the Netherlands, where retaliation against reporters of unequal treatment is prohibited under article 8a of the General Equal Treatment Act. This article, however, is, again, a formal rule. It does by no means guarantee that retaliation does not occur in practice.This study focusses on the forbidden practice of retaliation against employees who address unequal treatment at work in the Netherlands. This practice of retaliation is an under-studied phenomenon, which so far has been given little attention outside the USA (Brake, 2005; Skiveness and Trygstad, 2010). The research question steering our investigation reads:RQ1. To what extent do employees who complain about unequal treatment experience retaliation and which factors explain the variation in the nature and extent of retaliation?To answer this question, we proceed as follows. First, in the theory section, we discuss the existing literature on retaliation against employees who address wrongdoing, to arrive at three sets of explanations for variety in the nature and the extent of retaliation. Then, we present the research method used in testing these explanations, including the method of data collection and the measurements. This method section is followed by the results section, and finally by a conclusion, in which we also discuss some practical implications of our findings.\nTheory and hypotheses: As argued above, retaliation against employees who report unequal treatment can have important consequences. It is therefore somewhat surprising that the field of employment discrimination research has not addressed this issue of retaliation extensively (Goldman et al., 2006; Miles et al., 2010). Instead, the employment discrimination literature mostly focusses on explaining antecedents, consequences and outcomes of acts of discrimination (Goldman et al., 2006; Miles et al., 2010) and on explaining outcomes of employment discrimination claiming (e.g. Goldman, 2001; Hirsh, 2008). This does not mean, however, that theory on the topic of retaliation is fully lacking. In the literature on whistleblowing, retaliation has received much more attention in the past and, as Bergman et al. (2002) argue, this literature on retaliation against whistleblowing is conceptually similar because it is also \"intended to end an illegal or unethical situation\" (p. 232). So, what does this existing theory tell us about retaliation and about the factors that explain this phenomenon?First of all, it is found that retaliation refers to a wide variety of punitive actions taken against employees who address unlawful practices (Rehg et al., 2008). The most obvious instance of retaliation is that in which a superior in an organization uses his formal powers to punish an employee by affecting the employee's work conditions and/or the employee's promotion opportunities. Cortina and Magley (2003) call this type of retaliation work retaliation. Examples are unjustified negative performance appraisal, denial of promotion or other advancement opportunities, suspension, demotion and transfer to a different geographic location (Cortina and Magley, 2003; Ormerod and Vaile-Wright, 2003).In addition, Cortina and Magley (2003) distinguish social retaliation as a second important form of punitive action. Social retaliation refers to a variety of actions which do not require any formal power of decision over the employee's work and work conditions, but which instead are based on social powers, i.e. powers to change the social context in which the employee works. These social powers, which rest primarily with actors with whom the employee is in contact on the work floor, can also be used punitively. Examples of social retaliation are gossiping about the employee, addressing the employee in a negative manner, shunning and excluding the employee from social activities and mobbing (Cortina and Magley, 2003; Namie, 2003; Vandekerckhove and Commers, 2003).Given the many forms of work retaliation and social retaliation, punitive actions against employees who challenge unequal treatment can vary widely in both nature and extent. Three types of explanation can be distinguished for this variety in retaliation. The first type of explanation, which has received most attention in the existing literature, links retaliation to the nature of the accusation implied in a complaint or a report about organizational wrongdoing. The second type of explanation concerns the organizational context in which the wrongdoing is addressed and the third type of explanation focusses on the process which is followed by the employee in question, when he or she addresses organizational wrongdoing. These three types of explanation will now be discussed to arrive at seven separate hypotheses regarding the explanation of retaliation against employees who report unequal treatment.Nature of the accusation\nMethod and data: Data collection for the empirical part of the study consisted of telephone interviews with employees who had taken their case outside the organization, namely by filing a formal report about unequal treatment at work to the aforementioned Dutch Equal Treatment Commission.Based on experiences in earlier research, this sample was selected carefully. First, following the findings of Jaspers and Kleinwee (1985), the target population was limited to those employees who had submitted a complaint concerning unequal treatment against their then employers. This implied that two particular types of complaints were excluded, namely, claims about unequal treatment during application procedures, and claims from employees who at the time of filing their complaint had already left their employer. Second, the study was limited only to those employees whose complaints had been taken up by the Equal Treatment Commission. The reason for this was that only if the commission had taken up the case, it was certain that the employer had been informed about the complaint and that the conditions had been present in which retaliation could have occurred (see also Miles et al., 2010). Third, the target population was limited to those employees who had submitted their complaint between one and four years before the research was conducted. The latter limit was chosen because the study was meant to shed light on the recent state of affairs. The reason not to study more recent cases was that retaliation can take some time to develop and to be recognized as such.In total 126 persons were interviewed about their experiences, which amounted to a response rate just under 55 percent, which is considered quite high in the Netherlands. Most non-response was caused by the practical problem of contacting the persons involved (31 percent). This was often due to out-dated contact information, but also the result of failure to get telephonic contact within six attempts. In total, 12 percent of those contacted refused to cooperate altogether or did not complete the whole interview. In less than 2 percent of the cases the interview could not be conducted as a result of insurmountable language problems.The interviews\nResults: The results of the statistical tests of the three sets of hypotheses are presented in Tables III through V. In addition, we also present the mean scores for both social and work retaliation, for the different categories of each independent variable.As shown in Table III, the nature of the implied accusation was found to matter. First of all, H1 was supported by the finding that social retaliation was significantly higher when complainants accuse actors closer to the work floor - such as colleagues and/or direct supervisors - than when complainants accuse their formal employers or \"plans and regulations.\" As expected, the relationship between the type of actor accused and the extent of work retaliation was not statistically significant. As presented in the same table, we also found support for H2, as the extent of social and work retaliation were both significantly related to the equal treatment ground involved. On the basis of Tamhane post hoc test it could be concluded that especially complainants who had addressed racial discrimination, had experienced significantly more social retaliation than complainants who had addressed discrimination on the grounds of age or type of labor contract. With respect to H3, no significant difference in social or work retaliation was found between cases in which the claim had been acknowledged to have merit and cases in which the claim had been refuted.The tests of H4 and H5, concerning the organizational context, provided a mixed picture (Table IV). A significant correlation was found in support of H4, that both social and work retaliation are less severe when an employee has more colleagues who may be discriminated against on the same equality ground. H5, about the positive effect of more extensive institutional protection in larger organizations, was not supported by the data. Here we found no significant correlation between the size of the organization and the extent of social and work retaliation.Finally, Table V shows, that both process H6 and H7 were not supported by the data. No evidence was found for the assumption that involving a specialized adviser or respecting the inside-out norm helps to reduce the risk of social or work retaliation.\nConclusion: This investigation has revealed a high incidence of retaliation against employees who address unequal treatment in the Netherlands. The survey showed that almost two-thirds of the employees who had submitted a complaint to the Dutch Equal Treatment Commission reported retaliation. Therefore, the first conclusion must be that retaliation is a widespread and structural problem in the Netherlands and that, despite the formal statutory ban on retaliation against unequal treatment complaints, true freedom from this type of retaliation does not exist.The second finding is that the nature and extent of retaliation are clearly linked to the organizational circumstances in which the unequal treatment initially developed. These initial circumstances not only concern the nature of the accusation, but also the availability of support from co-workers with similar characteristics. Consistent with prior research on retaliation in the broader context of whistleblowing, we find that work retaliation and social retaliation are significantly more severe when a report of unequal treatment comes from a person who is relatively alone on the work floor and when the inequality addressed concerns an issue of higher social condemnation, especially racism. Furthermore, social retaliation is significantly higher when co-workers and/or supervisors close to the work floor are responsible for unequal treatment.Third, no empirical support could be found for the positive effects of organizational context and proper process. The fact that no relationship was found between the extent of retaliation and organizational size is remarkable in view of the efforts taken in larger organizations. Despite the fact that many of them invest in policies and protocols for dealing with internal complaints, in complaints committees, in specialized personnel officers and in confidential advisers, retaliation is just as common in larger organizations as it is in smaller organizations, in which this type of protection is hardly formalized. Furthermore, it is remarkable that prudent behavior in addressing unequal treatment does not help to avoid retaliation. Complainants who took the trouble to seek specialized advice were hit just as hard by retaliation as complainants who did not, and complainants who took the prudent route of first addressing the issue internally were in the end not better off than those who immediately involved the Equal Treatment Commission.Limitations and further research\n",
"role": "user"
},
{
"content": "Purpose: - Equal treatment in the workplace is considered one of the most fundamental rights of employees. This right also implies that employees must be able to address any form of unequal treatment freely and effectively, without fear of retaliation. The purpose of this paper is to investigate the prevalence of retaliation against complaints of unequal treatment in The Netherlands and its underlying factors.\nDesign/methodology/approach: - This article is based on a telephone survey among employees who filed a formal complaint about unequal treatment at work to the Dutch Equal Treatment Commission.\nFindings: - The main finding is that retaliation against equal treatment reports is commonplace in The Netherlands and in many cases takes on serious forms. Furthermore it is found that the nature and extent of retaliation are primarily explained by the circumstances in which unequal treatment develops and that the extent of retaliation is neither explained by the manner in which unequal treatment is addressed nor by the level of institutionalized protection that is available.\nOriginality/value: - The need for a better understanding of retaliation is high, because retaliation and fear of retaliation have broad consequences for the psychological and physical well-being of individual employees and for employment relationships. Yet, studies on retaliation in the context of employment discrimination are rare, especially outside the USA. This article contributes to a better understanding of retaliation, which may be used by policy makers, employees and employers to address it more effectively.\n",
"role": "assistant"
}
] |
Are healthcare middle management jobs extreme jobs?
|
[
"Middle management",
"Extreme jobs",
"High work investment",
"Hospitals"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The ward leaders - sisters and charge nurses - make or break the hospital's reputation. We have over 50 wards, on two sites. Each has a team of 20-30 staff operating 24/7. It's where the costs arise, and where patients and visitors make judgements about standards, depending on staff approach and discipline. That's where the key business of patient care is (Chief Executive, Greenhill Hospital).\nResearch aims and settings: This paper is based on results from a wider study of the roles of middle managers in six acute hospitals, profiled (using pseudonyms) in Table I. Following a review of management research priorities (Christian and Anderson, 2007), this wider project aimed to identify how managerial roles were changing, focusing on the pressures and demands facing middle managers, and on their motivations and rewards.\nThe intensification of middle management work: It is not surprising that, in the face of so many pressures and demands, hospital management work had intensified. Middle managers in general appear to have been subject to work intensification for some time. From their study of eight public and private sector organizations in six Western European countries,\nResearch methods and construct development: Information was gathered between 2009 and 2012, from hospital websites, and from 93 setup interviews, 40 focus groups, a \"proof of concept\" pilot study with six participants, a management survey with 611 responses, and feedback briefing sessions with a further 221 respondents (Table II). Figure 1 summarises this approach, illustrating the sequence in which these methods were deployed and the linkages between them. Interview questioning and focus group discussions explored how managerial roles (pure plays and hybrids) were changing, their motivations and rewards, the pressures and demands facing middle managers, and how management contributions could be strengthened.\nSupporting evidence: Following the pilot, to investigate this extreme job profile more systematically, with a larger sample, the design of the \"job characteristics\" section of the management survey was partly based on the \"proof of concept\" items. Other sections of the survey incorporated items relating to personal work experiences, work-life balance, and stress. The survey was administered during the final quarter of 2011, at five hospitals (the sixth was involved in merger plans and did not welcome the distraction of a survey). The survey was administered online using hospital email distribution lists. A total of 611 replies were received (a 24 per cent response rate). Of that total, 18 per cent indicated that they had management roles, and 51 per cent had hybrid roles. However, 31 per cent did not answer this question, making comparisons between these groups difficult, so this has not been attempted.\nConclusions and implications: This research has tentatively established a healthcare variant of the \"extreme job\" profile, incorporating the 12 characteristics summarised in the following list:\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore the incidence of \"extreme jobs\" among middle managers in acute hospitals, and to identify individual and organizational implications.\nDesign/methodology/approach: The paper is based on interviews and focus groups with managers at six hospitals, a \"proof of concept\" pilot with an operations management team, and a survey administered at five hospitals.\nFindings: Six of the original dimensions of extreme jobs, identified in commercial settings, apply to hospital management: long hours, unpredictable work patterns, tight deadlines with fast pace, broad responsibility, \"24/7 availability\", mentoring and coaching. Six healthcare-specific dimensions were identified: making life or death decisions, conflicting priorities, being required to do more with fewer resources, responding to regulatory bodies, the need to involve many people before introducing improvements, fighting a negative climate. Around 75 per cent of hospital middle managers have extreme jobs.\nResearch limitations/implications: This extreme healthcare management job model was derived inductively from a qualitative study involving a small number of respondents. While the evidence suggests that extreme jobs are common, further research is required to assess the antecedents, incidence, and implications of these working practices.\nPractical implications: A varied, intense, fast-paced role with responsibility and long hours can be rewarding, for some. However, multi-tasking across complex roles can lead to fatigue, burnout, and mistakes, patient care may be compromised, and family life may be adversely affected.\nOriginality/value: As far as the authors can ascertain, there are no other studies exploring acute sector management roles through an extreme jobs lens.\n",
"role": "assistant"
}
] |
When rehabilitation and duplication collide
|
[
"Tidal waves",
"Fishing",
"Rehabilitation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nBackground of the village: Kameshwaram is a fisherfolk village that comes under keelvelur taluk of Nagapatnam District. There are 210 families in the village and fishing is the prime occupation. Before tsunami this village had 40 fiber boats, 20 catamarans with outboard motors and 20 catamarans without engines. The fiber boats found in this village before tsunami were small in size compared with the latest design. In fiber boats four people will go fishing and the catch will be divided into six parts. The owner will take three shares for his assets and the workers will take one share each. The catamarans are made of wooden logs and three people will go fishing. The catch will be divided into five parts and the owner will take three shares and the workers will take one share each which is the common practice in many fisherfolk villages.Prior to the tsunami the village of Kameshwaram was just 50 metres from the sea. The villagers used to keep their fishing assets on the seashore. There are fresh water canals on the northern and southern sides of this village. The southern canal is very close to this village and for three months only (during rainy season) the canal will be open and water will enter the sea. During other months, due to lack of water flow, the mouth of the canal will be closed by sand due to sea waves.Before the tsunami, the fishing assets were owned by a few families and many worked as fish workers in the boats and catamarans for their livelihoods. The fisherfolk will start their fishing activity in the early morning and the crowd will gather in the shore before starting for the sea. There was a strong relationship between their traditional wisdom, resource, livelihoods and harmony and mutual support in pushing the boats/catamarans into the sea.\nTsunami - boon or bane?: Tsunami had the worst impact in Kameshwaram village as 36 persons died and among them 11 were children as this village was very close to the sea. In the aftermath of tsunami, more than 50 Non Governmental Organizations (NGO) and 25 International Organizations (INGOs) flocked to this village with relief and rehabilitation programmes. There was a competition among NGOs and INGOs in relief distribution with main focus on spending funds rather than considering the need of the community. As a part of livelihood rehabilitation many INGOs and NGOs distributed fiber boats and at present this village now has 100 fiber boats. The present situation is that every two families own a boat on their own. The boats are distributed with joint ownership between two families.\nPresent situation: The author made a visit on 03.02.2006 to study the situation on take up of the rehabilitation efforts. Mr Gandhi a youth told that:...before the tsunami we all will come to the shore to go for fishing in the morning and we help each other in pushing the boats/catamarans into the sea, whereas at present since the number of boats has increased and owned by families with relatives, they prefer fishing at different times.Many times there will be insufficient numbers of persons on the shore to push the boats into the sea as each boat needs minimum six to eight persons for launching into the sea. He expressed that revival of 150 metres of southern canal with 30 feet breadth and 10 feet depth inland and 40 meters of length into sea to control sand deposits in the mouth of the canal will solve their problem of needing to push their boats out to sea, as they will bring their boats inland through the canal. When asked about the cost of the project he told it may be a hundred million (see Figure 1).Mr Ramesh from this village told that \"after tsunami the level of sea has raised and during full moon and new moon days the sea will be rough and hence the boats kept on the shore will collide with one another and are damaged due to high tides\". In the earlier days only 40 boats were anchored in the 500 meter stretch of shore whereas now 100 boats are anchored in the same stretch. Due to overcrowding of boats within the stretch, damage is happening during times of rough waves. He said that revival of canal will provide safety and cost of canal revival will be a hundred millions Rs.Also the NGOs distributed boats for women SHGs. In this village Ms Indira Gandhi a member of women SHG told that \"we got the boat in our SHG but we are not involved in fishing activity and the local panchayat had control over the asset and they gave Rs.60,000/- to our SHG and took the ownership of the boat\".\nWhat has been learned from the rehabilitation?: * There are surplus boats and there is no labour force to mutually share the fishing activity.* The ownership of boats is within families and fishing timings are changed to their convenience.* After tsunami people are not interested to take up hard labour in pushing the boats, whereas prefer canal renovation (scientifically and technically a mega project) for bringing the boats inland.* Over crowding of boats within the stretch is a reason for damages during rough seas.* Lack of base line village information resulted in over subscription of resources (fishing assets).* Lack of NGOs knowledge on field realities and idea on traditional wisdom and coping mechanisms has resulted in duplication of relief and rehabilitation efforts.* There was also pressure on International NGOs to spend money within a time frame has resulted in in-effective resource distribution.* Distribution of boats to women SHGs is not effective as the control was in the hands of local village Panchayat (the rich and the dominant families represent this unorganized group).\nConclusion: Over-distribution of boats has resulted in lack of labour and also pampering with relief and rehabilitation interventions made the community demand, and get all their requirements met, through greed rather than necessity. Since mutual support is lost, people are experiencing livelihood constraints. The expectations of the people of this village are routed more in canal revival which needs more resource and more technical inputs than are currently available.We feel there is a development conceptual change from \"teach the poor man to fish rather than provide him with fish\" to \"provide boat/net, do the fishing, sell them and give the money to the poor man\" is the situation after the tsunami in many fisherfolk villages.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to describe the way the functioning of a fishing village changed as a result of the tsunami.\nDesign/methodology/approach: - The paper outlines the experiences of the village before and after the tsunami.\nFindings: - The paper lists all that has been learned from rehabilitation following the tsunami.\nOriginality/value: - It is felt that a conceptual change from \"teach the poor man to fish rather than provide him with fish\" to \"provide boat/net, do the fishing, sell them and give the money to the poor man\" is the situation after the tsunami in many fisherfolk villages.\n",
"role": "assistant"
}
] |
Consumer cultural identity: local and global cultural identities and measurement implications
|
[
"Cultural identity",
"National identity",
"Ethnocentrism",
"Nationalism",
"Global citizenship through global brands",
"Global identity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Although the processes of globalization date back centuries, many globalization scholars date the modern era of globalization to the collapse of the Berlin Wall in 1989, the end of the Cold War and invention of the internet (Friedman, 1999). Interestingly, early assessments pointed to a win of globalization over nation-state sovereignty and isolation, such that Fukuyama (1992) described this time as \"the end of history\" and the achievement of the final liberal democratic form of government. Subsequent technological and digital developments, the enactment of the European Union and growing trade agreements worldwide further empowered globalization as a world force (Hardt and Negri, 2000). In 2018, however, the promises and successes of globalization are in question with many nations engaged in trade wars, increased isolationism and proclamations of sovereignty (e.g. MAGA campaign in the USA, Brexit in the UK, Catalonia's fight for independence in Spain). These global political and economic tensions have given rise to the increased salience of globalization and nationalism discourses related to consumer identity and consumption practices.\n1. Conceptualizing consumer cultural identity: Cultural identity is defined as \"a broad range of beliefs and behaviors that one shares with members of one's community\" (Jensen, 2003, p. 190, 2011; Berry, 2001). As globalization has progressed, community has been contemplated as the global (worldwide, universal) community, one's local (national, traditional) community and as an intermingling of the global community and one's local community. Arnett (2002), consistent with earlier work on psychological acculturation (Berry and Kim, 1986; Graves, 1967), argued that globalization would have psychological influences on identity formation and development, i.e., how consumers think about themselves in relation to the social environment. Arnett's key premise is that most individuals develop some form of \"bicultural identity, in which part of their identity is rooted in their local culture while another part stems from an awareness of their relation to the global culture\" (p. 777). This body of work posits that the integration of local and global identity markers will be challenging and may cause identity confusion, yet concludes that for most individuals, some integration of global and local identity markers is inevitable.\n2. Conceptualizing local and global consumer cultural identities and measurement approaches: Extant research has laid the conceptual groundwork for identifying consumers who are (vs are not) engaged with local and/or global discourses. Figure 1 provides a framework for considering engagement/unengagement and further differentiates those with local, global and glocal identities and consumption practices. In the following sections, we begin by considering those engaged in the local and global discourses with attention to reviewing the conceptualization and measurement of local, global and glocal cultural identities and consumption practices, and follow with a discussion of the \"unengaged\" segment, those who are not attentive to these global/local discourses.\n3. Perspectives on consumer cultural identity for future research: Marketing scholars have addressed consumer cultural identity from many perspectives, including for example, contextualized within a global consumer culture, as segmentation variables and as predictors of marketplace outcomes. Consistent with past work on cultural identity both within and outside the field of marketing, our perspective is that consumers negotiate the local and global spaces, that these spaces are independent, yet can be co-mingled (Berry and Kim, 1986; Berry et al., 1986; Cannon and Yaprak, 2002; Steenkamp and de Jong, 2010). Herein, we have examined consumer cultural identity, focusing on local and global discourses and dialogical markers to explicate a consumer cultural identity framework marked by two dimensions: consumer engagement/unengagement with the globalization-localization discourses, and consumer engagement via cultural identity beliefs and/or consumption-based beliefs. As shown in Figure 1, for those who are engaged, we distinguish among those who have a stronger local (weaker global) identity or a stronger global (weaker local) identity and those who have strong local and global identities. Upon examination of the measurements to assess consumer cultural identity, we also distinguish between measures of a more general identity focus (e.g. national identity, global identity) vs those measures that are associated with consumption practices (e.g. consumer ethnocentrism, global citizenship through global brands) (see Tables I-III). In the next section, we provide an overview of measurement approaches to consumer cultural identity from the perspective of their ability to identify consumers with more global, local and glocal affiliations vs those unengaged with the global-local discourse. We then follow with a discussion of avenues for exploring research on consumer cultural identity.\n4. Conclusion: At a time when globalism and nationalism debates are on the rise, we and our colleagues interested in consumer cultural identity beliefs and consumption practices need to be attuned to the evolving socio-political-economic happenings around the world. These societal-level local and global discourses are likely to trickle down to impact consumers' identity beliefs, with the potential of creating additional stresses and tensions to these negotiated identities. Undoubtedly, conceptualization of consumer cultural identity and measurement approaches will continue to evolve, grounded in the recognition of both global and local influences on identity and consumers' varying levels of engagement with globalization-localization discourses. We encourage interested researchers to explore broadly as provocative questions emerge related to consumer cultural identity beliefs and consumption practices.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to offer a framework for considering the interplay between local (national) and global (world-based) identities and consumption practices with attention to various conceptualizations and measurements of consumer cultural identity.\nDesign/methodology/approach: This is a conceptual paper reviewing major works on consumer cultural identities and offering a framework for future considerations of the interplay between global and local identities.\nFindings: The framework identifies two dimensions which underlie consumer cultural identity conceptualizations and measurements: first, consumer engagement with globalization-localization discourses, and second, more general identity beliefs vs consumption-based identity beliefs. Clustering and categorical measure approaches (vs a compensatory approach) are preferred for identifying and exploring global/local/glocal and unengaged consumer cultural identity segments. Research foci should guide use of global and/or local general identity vs consumption-based identity beliefs as predictors of marketplace outcomes or as segmentation variables.\nResearch limitations/implications: The conceptualization of consumer cultural identity is based on Berry et al.'s (1986) early work on acculturation and Arnett's (2002) bicultural identity theorizing, and thus the authors acknowledge four consumer segments, those with: stronger global (weaker local) identity, stronger local (weaker global) identity, strong global and local identities and those unengaged with global-local discourses. The authors review measurement approaches to examine consumer cultural identity and determine that categorical and clustering (vs compensatory) approaches are consistent with the conceptualization of consumer cultural identity segments.\nPractical implications: International marketers can gain insights into major conceptualizations and measurements of consumer cultural identity, and understand the advantages and limitations of different measurement approaches. The authors highlight two important dimensions underlying cultural identity that demand managers' attention and consideration for strategic decisions. Social implications - this paper brings attention to various conceptualizations and measures of consumer cultural identity, highlighting the need to further examine differences between various cultural identity segments, specifically the unengaged consumers and glocally engaged consumers.\nOriginality/value: The paper provides a broadened lens to understanding conceptualizations and measurements of consumer cultural identity, identifying two dimensions underlying consumer cultural identity: consumer engagement with globalization-localization discourses, and more general identity beliefs vs consumption-based identity beliefs.\n",
"role": "assistant"
}
] |
Uncertainty avoidance as a moderator for influences on foreign resident dining out behaviors
|
[
"Attitude",
"Ethnic restaurants",
"Behavioural intention",
"Subjective knowledge",
"Uncertainty avoidance",
"Cultural trait"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Along with globalization, multiculturalism has become a common trend in the world. South Korea is no exception, with international business growing, multinational human resources and a wave of immigration leading to a more multicultural society. This is illustrated by information from the Seoul Census Bureau (2016); foreign residents are 2.7 per cent of Seoul city population, an increase of 23.2 per cent since 1970. This noticeable increase of foreign residents in Seoul emphasizes the importance of understanding their consumption behavior, especially considering the cultural aspects of service operations directly relating to service success (Patterson et al., 2006; Reimann et al., 2008). Indeed, foreign residents compose a large portion of restaurant patrons in Korea, even more than Korean residents in terms of dining out frequency (Ministry of Health and Welfare, 2013), so it is inevitable that foreign residents consume the foods offered in the host country. Despite the importance of this emerging market in South Korea, few studies have examined their dining out behaviors, especially from the perspective of cultural traits. Food choice is a complex phenomenon, and many factors affect the choice of foods, so a detailed theoretical examination of individual subjects would help explain the dining out behavior of foreign residents to better help restaurants meet their needs and desires.\nLiterature review: Relationship between subjective knowledge and behavioral intention\nMethodology: Sample and data collection\nResults: Demographics of respondents\nDiscussions and conclusions: This study explicated the relationships among subjective knowledge, attitude toward Korean foods and BI by integrating a cultural trait (UA) of foreign residents in Korea. Specifically, we adopted Hofstede's (1984) definition of UA because the trait is significant as a cultural dimension in international settings; it is strongly related to tolerance for risk taking in eating ethnic foods and dining out (Wong, 2004). Our findings were consistent with basic attitudinal research suggesting that attitude is a predictor for behavior (Armitage and Conner, 2001) but, in our research, for consuming ethnic food. In addition, subjective knowledge of Korean foods was important in explaining the behavior of foreign residents dining out at Korean restaurants. Subjective knowledge was critical in forming foreign resident attitudes toward Korean foods and their subsequent BI, with positive attitudes toward Korean foods enhancing the likelihood of foreign residents visiting Korean restaurants.\nTheoretical implications: The study results add to the literature in several ways. First, despite the proliferation of research into the sequence of steps from attitude to BI among foreign patrons in Korean restaurants, little research has considered subjective knowledge as a proximal factor leading to patron attitude and BIs. This study sheds light on the importance of subjective knowledge and attitude in predicting BIs for dining out. Moreover, this study shows that increased subjective knowledge of Korean foods leads to a positive attitude toward Korean foods, which ultimately affects the intention of foreign residents to visit Korean restaurants. In other words, increased knowledge of Korean foods seemed to reduce the concerns of foreign residents about novel foods and helped them to build more positive attitudes toward Korean food, which ultimately leads to higher intentions to visit Korean restaurants. Therefore, given that subjective knowledge is a critical part of understanding foreign patrons, the concept deserves more attention from researchers and practitioners.\nManagerial implications: The study findings suggest several important implications for practitioners as well. The findings indicate that enhancing subjective knowledge about Korean foods should increase the probability of foreign residents visiting Korean restaurants, so restaurant marketers should consider subjective knowledge as they work to encourage foreign residents to try Korean foods. The concept could help in tailoring marketing not only to new customers but also to existing customers, increasing the frequency with which they visit Korean restaurants. Promotion campaigns that offer Korean foods at temporarily lower prices may encourage customers to try the foods so that people begin to feel knowledgeable about Korean foods. In addition, Korean restaurant operators who want to target foreign residents should also consider other ways to enhance subjective knowledge of Korean foods among foreign residents. Opening a cooking class or distributing leaflets with information about ingredients, history and cooking methods would be effective in helping foreigners learn about Korean foods.\nLimitation and future research: We recommend future studies to supplement the limitations of this study. The first limitation pertains to the narrow geographic scope of the study and the use of non-probability sampling. Although Seoul was chosen because it is the most popular place of residence among foreigners in Korea, the choice may have contributed to biased results. Foreigners who live in Seoul, Korea's capital as well as a major metropolitan area, are more likely to be exposed to Korean culture. Thus, generalizations based on the findings of this study should be approached cautiously and interpretations should remain within the specific framework of the sample. A cross-validated model with larger samples in other regions would be helpful.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the moderating effect of uncertainty avoidance (UA) on the relationships among subjective knowledge, attitude toward Korean foods and dining out behavioral intentions (BI) of foreign residents in Korea.\nDesign/methodology/approach: A total of 247 foreign residents in Korea were participated through a street intercept survey at several locations in metropolitan areas of South Korea. Subsequently, the samples were divided into two groups (a low UA group and a high UA group) for multiple group analysis to examine the moderating role of UA.\nFindings: The results of structural equation modeling showed that subjective knowledge and attitude toward Korean foods significantly influenced intention to visit Korean restaurants. Furthermore, multiple group analysis results showed that UA had a significant moderating effect as a cultural dimension on the relationships between subjective knowledge and BI, as well as between attitude and BI.\nResearch limitations/implications: This research has made the first attempt to account for UA in examining the relationships among subjective knowledge, attitude and BIs, especially for ambiguous situations where foreign residents who are new to the mainstream Korean food culture face challenges in visiting Korean restaurants.\nPractical implications: The findings indicate that enhancing subjective knowledge about Korean foods should increase the probability of foreign residents visiting Korean restaurants, so restaurant marketers should consider subjective knowledge as they work to encourage foreign residents to try Korean foods. Furthermore, planning strategies for marketing to foreign residents should consider level of UA among foreigners.\nOriginality/value: This study first illustrates the value of considering the cultural trait of UA in examining dining out behavior at ethnic restaurants. The UA trait sheds light on how subjective knowledge helps predict attitude and dining out BI at ethnic restaurants.\n",
"role": "assistant"
}
] |
Leading beyond tragedy: the balance of personal identity and adaptability
|
[
"Leadership development",
"Personality",
"Organizational change"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nTragedy: September 8, 1989: The Friday night patrons at the National Theater in Oslo, Norway were surprised by an unexpected announcement asking for all employees of Wilhelmsen Lines (WL) to immediately meet in another room at the theater[1]. There were many Wilhelmsen people present that evening, all proud employees of the more than 100-year-old Norwegian shipping firm, one of the largest players in the global ocean transport business. Earlier that day, Wilhelmsen's managing director (CEO) and 49 other employees, including the entire top management team, had flown from Oslo to Hamburg for a ship-naming ceremony. The traveling employees had won the privilege to take the prestigious trip by having their names drawn in a company lottery, with chances in the lottery weighted by years of seniority; most on the trip had long associations with Wilhelmsen.However, as the company's employees at the National Theater that September evening were told, the plane carrying the Wilhelmsen group had disappeared. In the following hours they learned that the plane had gone down; there were no survivors. All weekend the company's office building was open, and pastors, counselors, and psychologists were available as the Oslo employees and their spouses numbly tried to grasp the enormous loss. The top management of Wilhelm Wilhelmsen Ltd (WWL), the parent company of WL, worked round the clock, quickly naming an executive from a related subsidiary company as interim managing director of WL. The first company meeting under the new leadership took place on that Sunday, and the WL doors were open for business Monday morning. However, things were far from normal. Entire groups of employees had disappeared, and the company was in shock.An employee reflected back on the sense of shock, working in an office suddenly devoid of many close friends and coworkers, saying:It was a very difficult time when everybody around me died ... Everybody was gone, you know. All the old people that I knew and that I trusted ... were gone ... I sort of felt like I was just walking around in a daze ... the whole building was filled with flowers, the phones were ringing ... and then they didn't find everybody, you know. So every day you heard, \"They found him. They found him. But they haven't found her.\" And there are still five people they haven't found.Another manager remembered sadly:I lost half my department, and some entire sections were lost. I really had problems getting into the computer, because everybody did not know the password. And we had to bring in people who had left the company. We worked day and night. It was a very, very, tough time. A lot of us had to go to many of the funerals, because being here a long time, I knew everybody.Thinking back on those difficult days, another long-term employee wondered how they actually got the work done and moved forward, saying:It was a very tough time. We had very long days, and after a while it was funerals in the day, and basically you had to work almost 'round the clock in order to cope with the workload.The company was moving forward, but its long-term survival was clearly at stake until a permanent leadership team could be put in place. WL needed to find the right leader to face this extraordinary challenge.\nThe new leader: After the initial crisis, in which the entire senior and middle management of the WL company was lost in the plane crash, an extensive search for a new leader for WL commenced. Recruiting ended when Ingar Skaug, then head of Scandinavian Air Systems (SAS) in Norway, agreed to take the position at WL. Skaug, a native Norwegian, came to the post at WL after 18 years in the airline industry, first within the Lufthansa organization and then at SAS, where he worked as a manager in several major European posts and rose to become head of US operations for SAS. Following his successful tenure in the USA, he returned to SAS Europe in 1982, and was appointed vice-president in 1986 to head up SAS' Norwegian operations. Throughout his time at SAS, Skaug worked closely with Jan Carlzon, a leader well-known for his focus on both employees and customer service (see Carlzon, 1987).While service oriented and skilled in running large organizations, Skaug lacked direct experience in the shipping industry - and certainly had not faced circumstances such as those at WL. Few leaders have. He joined the company with many reservations and concerns, but also with a strong sense of excitement about taking on such a formidable challenge. On his arrival at WL, although the interim management team expressed confidence and satisfaction with the state of affairs at the company, Skaug, the incoming CEO, saw many issues that he planned to address. He carefully studied the organization's characteristics and capabilities, and talked with many people inside the organization and throughout the shipping industry. Four of the key challenges Skaug identified were: grief, culture, management style and strategy. The accident had left a profound sense of organizational and personal loss, striking a deep mark on the company and the whole country. The company had a strong family-based heritage. However, there were now two very different groups - employees present prior to the accident and the newcomers; neither was really open to change. Also, Skaug saw the company as having a very hierarchically driven \"command and control\" management style. As such, employees were not accustomed to taking responsibility and making decisions. Skaug discovered that little about the current strategy of the company had been written down; when senior leadership perished, so did the strategic knowledge.\nSkaug's choices: As the new leader, Skaug faced three key choices in his change efforts at WL. To present and frame this leadership dilemma, we examine how Skaug, as leader of WL, answered three central questions; these serve as key decision-making polarities which must be addressed by all organizations facing dramatic change efforts and guide our analysis of Skaug's leadership:1. Speed - fast or slow? As change agent, the pace of change is a core issue and Skaug wondered how quickly to move forward. Was this a situation where the first 100 days were critical to long-term success? Did he need to establish himself and his ability rapidly? Or did he need to study the terrain, come to understand the culture, and spend time listening? Should he have purposely honored people's passage through the chaos and uncertainty, or quickly prompted a new focus?2. Values and vision - dictate or dialogue? Vision and values play center stage in many discussions of leadership and organizational change. Both Covey (1991) and DePree (1989) point to the importance of principle in leadership. Likewise, vision is a common dimension in the success of many (see Collins and Porras, 1997; Bennis and Nanus, 1985). At the same time, leadership, in some circumstances, must be quite directive. In this crisis situation, how much initiative should Skaug have taken in articulating organizational values and vision? In a time of grief and upheaval, would the company have found solace in the quiet of the past, or responded positively to a firm articulation of a vision and values to guide the future?3. Power sharing - drive change or empower? There are two very different uses of power in organizational change: hard and soft. Some use a hard approach; such leaders usually utilize downsizing, restructuring and economic incentives (e.g. \"Chainsaw Al\" Dunlap's tenure at Scott paper). Conversely, other leaders focus on organizational capability, investing in the culture and human resource strength, all the while fostering organizational learning (Useem, 1998). How much did he need to drive change versus empowering others?Skaug's approach: a paradoxical embrace\nIdentity and adaptability in action: Next, we explain how Ingar Skaug demonstrated his leveraging of these leadership metacompetencies in his CEO tenure at WL.Skaug's personal identity\nSummary and conclusions: There are many well-known leaders today who exhibit complex leadership behavior that varies widely as the situation and environment shifts. We acknowledge that Ingar Skaug, as our primary example in this paper, is only one example and is someone who is, of course, not perfect. In interviewing dozens of people associated with Skaug and in meeting with him personally numerous times, it became apparent that, while he is well respected, he has a few rough edges. For instance, some tell of his notable temper (even though he was said to get over his outbursts quickly). And while he actively attends to issues of balance in his personal life, it is difficult to do a perfect job at it when he has often spent well over 100 days a year traveling the globe at Wilhelmsen.However, the paradoxes (and leadership warts) do not belong to Skaug alone. For example, Jack Welch's profit/values matrix philosophy is well publicized, but we may forget his initial \"Neutron Jack\" nickname, earned for spearheading massive downsizing at GE early in his CEO tenure. The neutron phase of Welch's leadership may have helped make GE an unhappy place for a full ten years. Peter Browning, profiled in classic business school case studies, wisely employed \"bend but don't break\" organizational change strategies at both Continental Whitecap and Sonoco Products to enact successful leadership transitions in these heritage-rich companies - but he also presided over the painful Chapter 11 bankruptcy proceedings at National Gypsum and left Sonoco when growth projections were not achieved. Anita Roddick built her Body Shop organization around environmental and social causes, but operational concerns also had to be addressed when the company's profitability turned for the worse, and she left the company she founded over 25 years ago (Simms, 2000).As indicated in our description of Skaug's leadership, these successful - and complex - leaders seem to bear the common thread of clear sense of self, or identity, grounded in a clearly communicated values-orientation. This identity awareness shows through in those defining \"moments of truth\" (Carlzon, 1987), critical times when leaders drive their business strategy in their own unique ways. A second thread is their ability to adapt and change their leadership styles in concert with changing external forces. Conversely, leaders like Don Burr at People Express, while passionate, talented, and values-driven, were less able to shift when other styles and strategies were called for. It is, therefore, this combination, of identity plus adaptability, \"I/A balanced\" leadership, which we see as being critically important for contemporary leaders.Major leadership lessons\n",
"role": "user"
},
{
"content": "Purpose: - As a way to understand contemporary leadership, aims to examine an organization living through tragedy and study how its leader responded to extraordinary challenges over 12 years as chief executive officer. Proposes exploring leadership behavior through the lens of paradox, where leaders face conflicting choices in their decision-making processes.\nDesign/methodology/approach: - Primary case study research was conducted through on-site interviews, participant observation, and archival research. Building from the careers and leadership literature, illustrates how one leader sorted through key paradoxical choices and studies how his sense of personal identity and adaptability served as bedrock metacompetencies in navigating complex change.\nFindings: - Two metacompetencies - personal identity and adaptability - help explain how effective leaders respond to situations requiring seemingly contradictory sets of behaviors. Here, leaders navigate traumatic situations and conflict by drawing on a deep awareness of self and high levels of adaptability, knowing how to respond appropriately to crisis, and invoke behaviorally complex responses to challenges. Frames this as being evidence of \"identity/adaptability balanced leadership\" to offer leadership development lessons and suggest directions for future leadership research.\nOriginality/value: - This research examines leadership through the metacompetencies of identity and adaptability, emphasizing the critical mutuality across both dimensions. Draws direct linkages between the careers literature and the study of leadership and presents a meaningful way to understand the complexity of contemporary leadership, especially in times of great stress. These reflections offer broadly applicable lessons that will be useful to individuals seeking to grow as leaders and to development professionals who seek to foster growth in others.\n",
"role": "assistant"
}
] |
Leader's communication style, LMX and organizational commitment: A study of employee perceptions in Peru
|
[
"Peru",
"National culture",
"Leader-member exchange (LMX)",
"Affective organizational commitment",
"Organizational communication",
"Leader’s communication style"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Although the benefits of communication for organizations have been studied for decades, the line of research on the components of the leader's communication style (LCS) and their effects remains, comparatively, underdeveloped. Leaders are often not aware of their communication and do not give importance to it, so they do not realize that the way they communicate is perceived by their subordinates as their way of leading. In line with this, communication often seems so obvious and so embedded in human interactions that are hard to perceive. Fairhurst (2011, p. 43) quotes a story by David Foster Wallace in order to explain the nature of communication: \"Two young fish are swimming along, and they pass an older fish swimming the other way. The older fish nods at them and says, Morning, boys. How's the water?\" As the two young fishes swim on, one eventually turns to the other and says, \"What the hell is water?\" The attitude of the young fish could be predicated of many leaders who consider communication occurs naturally and are not aware of the impact the way they communicate has on their subordinates and on management outcomes (Fairhurst, 2011).\nBackground: The communicative behaviors of leaders\nResearch hypotheses: Effect of the leader's communication on AOC\nThe impact of the LCS on AOC through LMX: Drawing from the literature, one might expect that the LCS is a significant factor in the construction of LMX, and that the quality of the exchange contributes in turn to affective commitment to the organization. Some studies have found evidence on the relationship between an LCS and LMX (Fairhurst and Chandler, 1989; Fix and Sias, 2006; Geertshuis et al., 2015; Michael et al., 2005; Mueller and Lee, 2002), as well as between LMX and organizational commitment (Dulebohn et al., 2012; Eisenberger et al., 2010; Gerstner and Day, 1997; Wayne et al., 2002). The mediating role of LMX in the relationship between LCS and affective commitment toward the organization, however, remains largely unexplored. Our study seeks to help fill this gap.\nMethod: Sample\nResults: Table II presents the mean, standard deviation and correlations. AOC is significantly correlated with four dimensions of the LCS and correlates with LMX. The mediator variable LMX correlates with all the dimensions of the LCS.\nDiscussion: From a conceptual perspective, this research contributes to a better understanding of communication style mechanisms that support leaders in building leadership relationships with their collaborators and enhance AOC. This is the first research made in a Peruvian organizational context on these issues. Over the past decade, Peru has been one of the region's fastest-growing economies, with an average growth rate 5.9 percent in a context of low inflation (averaging 2.9 percent) (World Bank, 2018). Much of the growth of the Peruvian economy is explained by an exogenous factor: the price of metals and not by endogenous factors. Peruvian leaders are challenged to take advantage of the window of opportunity generated by the price of metals to boost the economy and build endogenous growth factors. One of the factors that can contribute to growth is through leadership itself. Therefore, this research seeks to contribute to the strengthening of leadership through the tool of the leader's own communication style.\nLimitations and future lines of research: The use of the CSI (de Vries et al., 2011) has been an essential tool as we have tried to advance theory in the context of the aforementioned thesis, in analyzing the organizational effect of different dimensions of leaders' communication. This instrument was originally drawn up for The Netherlands, and this paper employs it in a substantially different cultural context. The perceptions of communicative behaviors are demonstratively culture-specific, and some of our results suggest that future research may address the development of adapted models and measures that work well in different cultural environments. Future studies can explore the question whether the relationships we have identified would hold in societies with higher levels of individualism and lower power distance than Peru.\nManagerial implications: Communication styles can enhance affective commitment to the organization, because they project qualities of leadership that are highly valued today, as shown by the theories of charismatic and transformational leadership. A leader can use these communicative behaviors with a subordinate to forge a relationship of openness, trust and empathy; the subordinate receives support, sufficient and timely information, takes part in frequent dialogs and has the right to give an opinion to be respectively considered; all these are behaviors that, according to LMX theory, reinforce leadership (Graen and Uhl-Bien, 1995). The subordinates perceive that their supervisor represents values the subordinates find satisfactory, and generates a positive workplace climate, helping to bolster their AOC.\nConclusions: The reported research has been conducted in the context of Peru, a country characterized by high power distance and collectivist. This study contributes to clarify that each feature of the LCS has a different impact on the perception of the subordinate. Thus, the leaders should be trained to understand this model and be able to make the necessary adjustments to obtain the desired leadership results. Four of the six dimensions of the leaders' communication style - expressiveness, preciseness, questioningness and verbal aggressiveness - are related to the quality of the LMX, and indirectly related to the AOC. The leader's ability to communicate with a style characterized by expressiveness, precision and questioning makes it easy to build high-quality LMX relationships for Peruvian employees. On the contrary, a communication style characterized by high levels of verbal aggressiveness will negatively affect subordinates, limiting the possibility of building high-quality LMX relationships.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyse the impact of a leader's communication style (LCS) on the quality of interpersonal exchanges between leaders and followers (LMX), and how this translates into the employee's affective organizational commitment (AOC), in the context of Peru.\nDesign/methodology/approach: An integrated model of six dimensions is used to measure LCS. Using multiple hierarchical regressions and the Preacher and Hayes mediation model, the study focuses on determining the direct and indirect effect of each of the dimensions on LMX and organizational commitment.\nFindings: The dimension preciseness shows a significant direct association to AOC. Four dimensions are significantly related with LMX: expressiveness, preciseness and questioningness with a positive sign, while verbal aggressiveness records an important negative one. The same four dimensions show an indirect effect on AOC through LMX. Emotionality and impression manipulativeness do not record significant results.\nResearch limitations/implications: The research was carried out with a sample of 253 white-collar Peruvian professionals with high-level studies and managerial experience, which are not necessarily representative of the labor population. This research provides comprehensive evidence on how leaders' communicative behavior may contribute to desirable outcomes such as employee commitment in a Latin American cultural context, although the findings may apply to other cultures.\nPractical implications: This study contributes to clarify that each dimension of the LCS impacts differently on subordinate perceptions; leaders should understand this model and be able to make the necessary adjustments to their communication in order to obtain the desired results of leadership. The leader's ability to communicate with a style characterized by expressiveness, precision, and questioning makes it easy to build high-quality LMX relationships for Peruvian employees. On the contrary, a communication style characterized by high levels of verbal aggressiveness may negatively affect subordinates, limiting the possibility of building high-quality LMX relationships. This, in turn, affects AOC of employees.\nSocial implications: This study is a contribution to clarify that each feature of the LCS has a different impact on the perception of the subordinate, for which the leaders should be trained to understand this model and be able to make the necessary adjustments to obtain the desired results of leadership. The leader's ability to communicate with a style characterized by expressiveness, precision and questioning makes it easy to build high-quality LMX relationships for Peruvian employees. On the contrary, a communication style characterized by high levels of verbal aggressiveness will negatively impact subordinates, limiting the possibility of building high-quality LMX relationships.\nOriginality/value: The value lies in revisiting the construct \"leader's communication style\" to turn it into an instrument for the exercise of leadership. It is a contribution in favor of leaders becoming aware that their own communication style constitutes an instrument of effective leadership and a lever to optimize the commitment of their collaborators toward the organization.\n",
"role": "assistant"
}
] |
Escaping the corner of death? An eye-tracking study of reading direction influence on attention and memory
|
[
"Online marketing",
"Attention",
"International consumer",
"Location-driven placement",
"Reading direction"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Internet users are constantly exposed to marketing stimuli such as banners, sponsored search, logos, pop-up ads and rectangle ads in multiple locations of a webpage. However, Agarwal et al. (2011) highlighted the matter that the specific location on a webpage for placing the marketing stimuli determines different levels of revenue generation by the stimuli. Along this vein, research about the most effective and memorable location of such elements on the webpage is fragmented and inconclusive. Hill (2010) warns advertisers to avoid the \"corner of death\", a term coined to indicate the lower right-hand corner of any layout, whether it is a web site page, a TV screen, a print ad, direct mail piece or billboard. It is important to note that Hill's (2010) advice is based on Outing and Ruel's (2006) exploratory study with American participants. Considering the diverse internet audience, the question of whether this scanning pattern also applies to any internet reader remains unanswered. Despite being one of the typical places where marketers and advertisers place their logo, according to Outing and Ruel (2006), the corner of death is typically the second to last place an audience looks on a page. On the other hand, contrary to the advice of avoiding the corner of death or lower right-hand corner, Goodrich (2010) found that online ads placed on the lower visual field of the screen received more attention than those placed on the upper visual field. Therefore, decisions about marketing stimuli placement in online settings become critical if we consider the average ad viewing time per reader usually is less than 2 s (Hill, 2010).\nConceptual background: Should we place a logo to the left or to the right of the content on a webpage? Previous research has offered insights about the directional/lateral organization of marketing stimuli. The directional or lateral display influences consumers' cognitive (e.g. information processing and value judgment) as well as affective (e.g. attitude) evaluations of stimuli (Janiszewski, 1990, 1993; Valenzuela and Raghubir, 2009; Chae and Hoegg, 2013; Romero and Biswas, 2016). Among them, Atalay et al. (2012) examined the effect of product layout and found how items are laterally displayed has a prominent influence on consumers' visual attention in a shopping environment.\nHypotheses: Relevant to the study of two-dimensional space, there are two main mechanisms of visual attention: spatial attention and feature-based attention (Liu et al., 2007; Carrasco, 2011). Spatial attention is the attention process onto a specific spot from a visual field. It allows individuals to selectively process all visual information of a visual area. It is overt and is indicated by eye movement to the focus of attention. On the other side, feature-based attention can be deployed regardless of the focus of attention. Feature-based attention is deployed to differentiate and highlight particular features in a visual field. It operates both within and outside the spatial focus of attention. For example, individuals follow two mechanisms of attention when browsing products on an e-tailing website:\nMethod: In this section, we discuss how we carried out the experimental procedures. This process includes recruiting participants who are specifically bidirectional and unidirectional readers, creating an experimental webpage as the marketing stimuli and instructing the participants to conduct the experiment in lab settings.\nResults: Eye-tracking data\nDiscussion: The corner of death has been identified as the least desired place for a logo because of the lack of attention (Hill, 2010). Such notion was based on experiments conducted with North American consumers. Previously, how the different locations on a webpage are processed or remembered across countries and cultural groups has not been fully understood. Among the earliest to shed light on this research topic, our findings contribute to the understanding of the differences between unidirectional and bidirectional readers in their responses to directional organization of marketing stimuli. The results from the verbal measures revealed that consumers who are bidirectional readers pay more visual attention on the corner of death than the unidirectional (LTR) readers. Further, from the eye-movement analysis, it was found that unidirectional (LTR) readers and bidirectional readers differ in attention as measured by total fixations and duration time. In particular, we found that bidirectional readers spent a higher percentage of time fixed on the bottom right corner (the corner of death). They also have a larger amount of eye fixations on the right side of the webpage and, in particular, the top right corner. These results are consistent with previous findings reported by Roman et al. (2013) and Dobel et al. (2007). Interestingly, we found non-significant effect in logo memory. This finding converges with Agarwal et al. (2011), who did not find a correlation between attention and memory.\nManagerial implications: From a managerial standpoint, the findings can help global marketing managers understand how to improve the effectiveness of advertisements and webpage design. The findings render clear managerial expertise in regard to logo placement. The placement of a logo on the corner of death is not as detrimental in an advertising platform targeting bidirectional (LTR/RTL) readers. This is especially important when the time of fixation is abundant. For ads or web pages intended for certain international populations that are bidirectional readers, the corner of death might be strategically used for logo placement to achieve good results in visual attention.\n",
"role": "user"
},
{
"content": "Purpose: The authors aim to examine the effect of location-driven logo placement on attention and memory on the web addressing differences between individuals that read unidirectionally (left-to-right [LTR]) versus bidirectionally (both right-to-left and LTR).\nDesign/methodology/approach: Using an eye-tracking approach combined with traditional verbal measures, the authors compared attention and memory measures from a sample composed of bidirectional (Arab/English) readers and unidirectional readers.\nFindings: The findings reveal that unidirectional and bidirectional readers differ in attention patterns. Compared to bidirectional readers, unidirectional readers pay less attention to the logo on the bottom right corner of the webpage based on verbal measures. The eye-tracking data of the two groups further identify differences based on total hits and duration time. Unidirectional LTR readers demonstrate higher fluency in feature-based attention whereas bidirectional readers show higher fluency in spatial attention.\nOriginality/value: The authors expand on scarce research on reading direction bias effect on location-driven stimuli placement in online settings. They contribute to the understanding of the differences between unidirectional and bidirectional readers in their cognitive responses (attention and memory) to organization of marketing stimuli.\n",
"role": "assistant"
}
] |
Knowledge construction by users: A content analysis framework and a knowledge construction process model for virtual product user communities
|
[
"Communities",
"Knowledge processes",
"Virtual communities",
"Problem solving",
"Knowledge construction",
"User knowledge construction"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: It is widely accepted that knowledge is a key source of competitive advantage for organizations (Nonaka, 1991; Nelson, 1991; Nonaka and Takeuchi, 1995; Spender and Grant, 1996; Brown and Duguid, 1998). Yet the types of knowledge that organizations need are various: they are not limited to patents or strategic knowledge. They also require knowledge of efficient processes, and fixes to simple product issues. Knowledge-related processes whereby users of products themselves contribute are both theoretically and practically important. Users can provide valuable innovative ideas for product development and contribute to marketing strategies (Wurster and Evans, 1997; Mahr et al., 2014; Cui and Wu, 2016). However, there has also been considerable interest in knowledge construction by users within virtual communities where product users simply share their knowledge and solve technical problems collaboratively (Anderson, 2005). From such interactions, producers may be able to gain knowledge of product usage and applications, discover design defects and improve product design (Anderson, 2005; Mahr et al., 2014). A \"virtual product user community\", as such groups will be referred to here, can be defined as \"a producer-sponsored customer aggregation existing on the Internet to share usage experience and collaboratively to find technical solutions to problems within specific brand products\". Examples would be support forums run by large IT companies such as Dell, HP or Lenovo.\n2. Literature review: 2.1 Virtual communities of product users\n3. Research design: 3.1 Methodology\n4. Findings: 4.1 A content analysis framework for knowledge construction\n5. Discussion: 5.1 A new knowledge construction model\n6. Conclusion: The analytic framework developed in this study captures key elements of knowledge behaviours in contexts where critical thinking is not a key aspect, and can be used to study not only virtual product user communities, but also other similar contexts. Unlike knowledge construction in formal learning contexts, these activities may be widely practised by people in their daily lives and work, yet are less explored. Thus, the framework can complement existing analytical frameworks and tools exploring high-level cognitive development and critical thinking in CSCL. It encapsulates the key knowledge construction constituents in this type of problem solving and clarifies the relationships between their main categories.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to develop a content analysis framework and from that derive a process model of knowledge construction in the context of virtual product user communities, organization sponsored online forums where product users collaboratively construct knowledge to solve their technical problems.\nDesign/methodology/approach: The study is based on a deductive and qualitative content analysis of discussion threads about solving technical problems selected from a series of virtual product user communities. Data are complemented with thematic analysis of interviews with forum members.\nFindings: The research develops a content analysis framework for knowledge construction. It is based on a combination of existing codes derived from frameworks developed for computer-supported collaborative learning and new categories identified from the data. Analysis using this framework allows the authors to propose a knowledge construction process model showing how these elements are organized around a typical \"trial and error\" knowledge construction strategy.\nPractical implications: The research makes suggestions about organizations' management of knowledge activities in virtual product user communities, including moderators' roles in facilitation.\nOriginality/value: The paper outlines a new framework for analysing knowledge activities where there is a low level of critical thinking and a model of knowledge construction by trial and error. The new framework and model can be applied in other similar contexts.\n",
"role": "assistant"
}
] |
Organic food online shopping in Poland
|
[
"Online shopping",
"E-commerce",
"Poland",
"Consumer behaviour",
"Organic food",
"Distribution channels"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: There is an abundance of literature on organic food consumers (e.g. Aschemann-Witzel and Zielke, 2017; Petrescu et al., 2017; Anisimova, 2016; Bryla, 2016a; Lee, 2016; Lee and Hwang, 2016; Hemmerling et al., 2016; Henryks et al., 2014; Hsu and Chen, 2014; Kesse-Guyot et al., 2013) and on online grocery shopping (e.g. Huyghe et al., 2017; Mortimer et al., 2016; Grzybowska-Brzezinska and Rudzewicz, 2016; Grzywinska-Rapca and Grzybowska-Brzezinska, 2016; Cyrek, 2013). However, there is a shortage of studies combining these two approaches. That is why this study examines the characteristics, attitudes and opinions of organic food online shoppers (e-consumers). This relatively small segment of the population is compared with a representative sample of 1,000 Polish consumers. Organic e-consumers constitute 6.3 per cent of the total sample. This paper aims to answer the following research questions:\nMaterial and methods: A survey was addressed to Polish consumers. The sample consisted of 1,000 inhabitants of Poland aged 15-65. The sample was a representative of the general population, regarding: age, sex, education and the size of the city of origin. The survey was carried out with the use of CAWI (Computer Assisted Web Interview) method by a specialised marketing research agency (ARC Rynek i Opinia) in its online panel (epanel.pl). The questionnaire contained 19 thematic questions and 8 questions on respondent characteristics. Whenever a catalogue of options was proposed, the respondents had the possibility to supplement it with their own answer (semi-open questions) so as to ensure that the respondents' opinions are reflected to the highest degree in the research results. The questionnaire was designed in Polish due to the research setting. It is available from the author upon request.\nResults: Women are slightly more often represented in the group of organic food e-consumers (54.0 per cent compared to 50.1 per cent in the total sample). Considerable differences were observed between organic food online shoppers and the total sample regarding the distribution by age and income (average monthly disposable income per household member): organic e-consumers tend to be younger and more affluent. Unsurprisingly, organic e-consumers have a higher share of organic products in total food consumption, and are characterised by a higher willingness to pay (WTP) a higher price for organic food compared to conventional products (Table I).\nDiscussion: According to the results, organic e-consumers tend to be younger and more affluent than the total population. In the author's opinion, the characteristics of organic e-consumers are related to - organic consumption - consuming organic food only, organic and non-organic, or non-organic (conventional) only; online shopping - buying online only, online and offline, or offline only. Therefore, both streams of literature need to be examined. An examination of a data set of 44,000 American households indicated that education has a strong effect on the likelihood of buying organic products, and that the impact of marital status, income, and access to organic are consistent across models (Dimitri and Dettmann, 2012). In China, certified organic food purchase is associated with demographic variables, such as income, education, age, gender, presence of young children, household size, living in developed cities and overseas experience (McCarthy et al., 2016). The Nutrinet-Sante cohort study among 54,311 consumers revealed that regular organic product buyers, who constituted 14 per cent of the sample, were more highly educated and physically active than other clusters as well as had a lower probability of being overweight or obese (Kesse-Guyot et al., 2013). In Serbia, those who prefer organic food the most are predominantly female, somewhat more mature, with the highest proportion of employed and tend to be married and have children. They strongly value the importance of diet for health, a number of them have experience with a disease of a close person, which is believed that could be prevented by proper diet, but they do the least sports (Grubor and Djokic, 2016). Polish consumers with convenience and price orientation in their food choices are less inclined to buy organic products while consumers open to novelties and willing to try new foods more often declare to buy organic products (Zakowska-Biemans, 2011). A survey in a representative sample of 1,000 Polish consumers demonstrated that there are relatively more organic food buyers among women, older respondents, better-educated people, and rural inhabitants (Bryla, 2016a). Purchasers of food online tend to be young, well educated and in a good financial situation (Grzybowska-Brzezinska and Rudzewicz, 2016).\nConclusion: To the best of the author's knowledge, this is the first study of organic food online shopping in Poland. Organic e-consumers are driven by quality signs to a larger extent than the rest of the sample both in their organic and conventional food purchases. Among organic e-consumers, organic food authenticity is evaluated primarily with the use of the following criteria: product quality, quality signs, taste, and label. In the opinion of organic online shoppers, the crucial barriers to the development of organic market include primarily high prices, low availability, insufficient consumer knowledge and scepticism towards the systems of certification and labelling. The most important organic food selection motives among the online shoppers are: eco-friendliness, healthfulness, quality assurance, taste, and safety. Almost a half of organic e-consumers also buy organic food in specialised physical shops and a third of them in markets and bazaars. In a logistic regression model, age, income, WTP a premium price for organic food, importance attached to product appearance and to quality signs turned out to be statistically significant determinants of online shopping behaviour for organic food in Poland. Higher income, importance attached to quality signs and higher WTP increase the likelihood of being an organic e-consumer, while higher age and importance attached to product appearance decrease it.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyse selected characteristics, attitudes and opinions of organic food e-consumers (online shoppers) in Poland.\nDesign/methodology/approach: A survey was conducted among 1,000 inhabitants of Poland aged 15-65. The sample resembled the general population, regarding: age, sex, education and the size of the city of origin. In total, 63 respondents reported buying organic food online. This segment was compared with the rest of the sample with the use of two-tailed t-tests. A logistic regression model was applied to identify the determinants of organic food online shopping in Poland.\nFindings: In a logistic regression model, age, income, willingness to pay (WTP) a premium price for organic food, importance attached to product appearance and to quality signs turned out to be statistically significant determinants of online shopping behaviour for organic food in Poland. Higher income, importance attached to quality signs and higher WTP increase the likelihood of being an organic e-consumer, while higher age and importance attached to product appearance decrease it.\nResearch limitations/implications: First, some determinants of organic online shopping have been identified. Second, this study has provided some frameworks to analyse organic food selection motives, barriers to the development of the market, food authenticity assessment criteria, distribution channels, and product characteristics.\nPractical implications: The findings may be useful for the industry due to the identification of the role of quality signs in online food purchases and of the importance of selected organic food consumption motives.\nOriginality/value: To the best of the author's knowledge, this is the first study of organic food online shopping in Poland.\n",
"role": "assistant"
}
] |
Enhancing outpatient appointment scheduling system performance when patient no-show percent and lateness rates are high
|
[
"Decision making",
"Patient satisfaction",
"Analysis",
"Clinical effectiveness",
"Capacity management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Large outpatient health-care clinics schedule thousands of patient appointments each year. Productivity improvement in the clinics is becoming a central concern (Arafeh et al., 2014; Carpenter et al., 2011). For example, long wait times by patients seeking consultation is a source of endless complaints (Amar et al., 2015; Huang, 2008). Today's customers seek quality of service for the price they pay (Wang and Fung, 2015). The effectiveness of the scheduling process has a critical impact on the clinical resource usage and patient satisfaction (Speights, 1997; Rohleder and Klassen, 2000; Yu-Li, 2013). The effectiveness of the scheduling process has a critical impact on patient no-shows, which refer to those patients who miss their appointments with no forewarning (Akin et al., 2013; Green and Savin, 2008; Grunebaum et al., 1996).\nMethodology: This section includes the suggested methodology and the patient flow with no-show and lateness representation.\nModel building, experimental design, and analysis: The model building consists of several vital activities including data collection, statistical fitting, and simulation modeling and experimental design.\nResults and discussion: The simulation is run for all designated BASIS and POB combinations. The performance measures are evaluated for all of these runs and averaged over replicates. The following sections explore the results obtained. 3D plots are adapted to graphically show the relation between the input variables and the performance measures.\nConclusions: This work was devoted to studying how outpatient appointment scheduling performs when changing the values of two main design parameters, namely, the BASIS and the POB. Simulation is used to evaluate the different performance measures. An insight into the simulation results shows the following interesting results:\n",
"role": "user"
},
{
"content": "Purpose: High lateness and no-show percentages pose great challenges on the patient scheduling process. Usually this is addressed by optimizing the time between patients in the scheduling process and the percent of extra patients scheduled to account for absent patients. However, since the patient no-show and lateness is highly stochastic we might end up with many patients showing up on time which leads to crowded clinics and high waiting times. The clinic might end up as well with low utilization of the doctor time. The purpose of this paper is to study the effect of scheduled overload percentages and the patient interval on the waiting time, overtime, and the utilization.\nDesign/methodology/approach: Actual data collection and statistical modeling are used to model the distribution for common dentist procedures. Simulation and validation are used to model the treatment process. Then algorithm development is used to model and generate the patient arrival process. The simulation is run for various values of basic interval scheduled time between arrivals for the patients. Further, 3D graphical illustration for the objectives is prepared for the analysis.\nFindings: This work initially reports on the statistical distribution for the common procedures in dentist clinics. This can be used for developing a scheduling system and for validating the scheduling algorithms developed. This work also suggest a model for generating patient arrivals in simulation. It was found that the overtime increases excessively when coupling both high basic interval and high overloading percentage. It was also found that: to obtain low overtime we must reduce the basic interval. Waiting time increases when reducing the basic scheduled appointment interval and increase the scheduled overload percentage. Also doctors' utilization is increased when the basic interval is reduced.\nResearch limitations/implications: This work was done at a local clinic and this might limit the value of the modeled procedure times.\nPractical implications: This work presents a statistical model for the various procedures and a detailed technique to model the operations of the clinics and the patient arrival time which might assist researches and developers in developing their own model. This work presents a procedure for troubleshooting scheduling problems in outpatient clinics. For example, a clinic suffering from high patient waiting time is directly instructed to slightly increase their basic scheduled interval between patients or slightly reduce the overloading percentage.\nSocial implications: This work is targeting an extremely important constituent of the health-care system which is the outpatient clinics. It is also targeting multiple objectives namely waiting times, utilization overtime, which in turn is related to the economics and doctor utilization.\nOriginality/value: This work presents a detailed modeling procedure for the outpatient clinics under high lateness and no-show and addresses the modeling procedure for the patient arrivals. This 3D graphical charting for the objectives includes a study of the multiple objectives that are of high concern to outpatient clinic scheduling interested parties in one paper.\n",
"role": "assistant"
}
] |
Reflections on defining global brands, fragmentation and segmentation, and the emergence of richer brandscapes
|
[
"Global brands",
"Starbucks",
"Hybridization",
"Digital networked technologies",
"International market segmentation",
"Luckin Coffee"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: I thank my colleagues Russell Belk, Aysegul Ozsomer and Saeed Samiee for their thoughtful commentaries on my paper. There seems to be broad agreement that globalization is under pressure currently but that it will survive, probably in a somewhat different form. That is to be expected. Globalization phenomena and signifiers in 2019 are vastly different from those before the Berlin Wall fell three decades ago, let alone in the era of the mighty Roman Empire.\nWhat are global brands?: I share Samiee's concern that global brands are not defined with a degree of precision that one would hope, and different people propose different definitions. It reminds me of the construct of product quality, one of the \"hot\" topics of the 1980s. Countless definitions exist (Steenkamp, 1989), but everybody recognizes it when they see it. I think this also applies to global brands. For example, everybody would agree that Rolls Royce is a global brand despite tiny consumer penetration and lack of meaningful sales in most countries.\nFragmentation and segmentation: Ozsomer thoughtfully points to the impact of digital networked technologies (DNTs) on global consumer culture. I believe that DNT - with their international reach, the world being just one click away - will fragment markets within a country and unify microsegments across countries into global (micro?) segments. The implications are profound. While a microsegment of 1 percent in a local market may not be economically viable in any but the largest economies, a global segment of 1 percent certainly is, especially if it is coupled with a digital marketing strategy (Gielens and Steenkamp, 2019). This renders accurate identification of global target segments an important issue (Steenkamp and Ter Hofstede, 2002), as pointed out by Samiee. Ter Hofstede et al. (1999) wrote an article on segmenting international markets based on attributes, benefits and values, using a relatively simple data collection procedure. This paper won the O'Dell award, suggesting long-term impact, but actually, there has been little follow-up or subsequent method development. If anything, interest in global segmentation methods and applications seems to be waning, not growing. This is troublesome. I think it is fair to say that our field agrees with the following statements:\nEmergence of a richer brandscape: I agree with Belk that today's dynamic processes of globalization and localization, as well as the emergence of new brands from emerging markets noted by Ozsomer (see also Kumar and Steenkamp, 2013) create a richer brandscape. The era of slash and burn of many local brands in favor of a few global brands appears to be drawing to a close, although this is most evident in \"cultural\" categories such as foods and beverages. In consumer durables and high tech, world markets are dominated by global brands like Apple, Samsung, Volkswagen and Toyota.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to present a response to papers written by Professors Samiee, Belk and Ozsomer as commentary pieces on my original paper, \"The uncertain future of globalization: implications for global consumer culture and global brands.\"\nDesign/methodology/approach: The paper is in the form of an essay.\nFindings: The commentaries highlight many areas of agreement with the overall thrust of the original paper, and also pinpoint novel and important additional avenues for reflection and research.\nOriginality/value: The paper builds on the commentaries, identifying additional ways forward for the field.\n",
"role": "assistant"
}
] |
Uncovering the intellectual development of the Journal of Organizational Change Management: A knowledge-stock and bibliometric study, 1995-2011
|
[
"Journal of Organizational Change Management",
"Knowledge‐stock analysis",
"Citation analysis",
"Co‐citation analysis",
"Intellectual development",
"Journals",
"Knowledge management",
"Serials"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Considering that organizational change management (OCM) is one of the highest stakes for organizations (Luscher and Lewis, 2008) and that \"there is a large and growing literature on the causes, consequences, and strategies of organizational change\" (Herscovitch and Meyer, 2002, p. 474), a short break is needed in order to be able to see where we are now and where we come from.The 25th anniversary of the JOCM is the occasion to grasp the intellectual progress of this journal in order to provide constructive elements for its further development. Somehow, a part of the OCM research intellectual development can also be analyzed. The articles published in the JOCM represent an interesting fraction of OCM research efforts as this peer-review journal is one of the leaders in the discipline with a rejection ratio of 92-94 percent of submitted papers[1] Moreover, a significant part of these articles deal with OCM (see Figure 1). Eventually, when looking at the journals specifically addressing the question of change management, the JOCM appears to be the most appropriate source of data. Table I shows indeed that the two other ranked journals deal with change through the lens of different social sciences like \"industrial organization\" (Industrial and Corporate Change) or \"innovation and entrepreneurship\" (Technological Forecasting and Social Change). In 2011, the JOCM is the only scientific journal which deals with organizational change management while being included both in the Association of Business Schools[2] ranking and in the Journal Citation Reports[3].This research proposes a single-journal analysis as a first step towards a full bibliometric study of OCM, which would require a multiple-journal analysis. While for instance the domain of business ethics has a clear leading journal which \"plays an important role in setting the research agenda for the entire field\" (Calabretta et al., 2011, p. 499), OCM seems to be addressed in too many different journals. As this brings major methodological impediments hindering a multiple-journal bibliometric study about OCM (presented in the coming literature review), the choice is made here to focus on a single journal.Conceptual papers about the topic of OCM have mainly turned out to be literature reviews (Armenakis and Bedeian, 1999; Armenakis and Harris, 2009; Buchanan et al., 2005; By, 2005; Oreg et al., 2011; Van de Ven and Poole, 1995; Van de Ven and Sun, 2011; Weick and Quinn, 1999) or meta-models (Young, 2009). To some point, these studies might have been influenced by the subjective views of their authors (Ramos-Rodriguez and Ruiz-Navarro, 2004) and have not yet been complemented with a bibliometric study. Indeed, bibliometric studies \"have the advantages of quantifiability and objectivity\" (Nerur et al., 2008, p. 320). They enable the identification of the most impacting documents (through a citation analysis) and the pinpointing of the linkages between them (through a co-citation analysis).Therefore answering to the call for a bibliometric research, the present paper proposes both a citation analysis and a co-citation analysis of the articles published in a major journal mainly dealing with OCM: the JOCM. The period covered by this research is the last 17 years (1995-2011).Ramos-Rodriguez and Ruiz-Navarro (2004, p. 981) recall that bibliometrics refer to \"the mathematical and statistical analysis of patterns that appear in the publication and use of documents (Diodato, 1994)\". The field of bibliometrics has become well established and has known recent developments regarding its practice (Thelwall, 2008). Nevertheless, the citation analysis remains one of the basic techniques (Moed, 2009) and relies on the hypothesis that frequently cited documents are particularly influential as they represent the basis for the development of researchers' thought (Culnan, 1987; Moed, 2005). Co-citation analysis is also a widely used technique (Usdiken and Pasadeos, 1995) and focus on the documents which have been cited together (co-cited) in the same paper. The pairs of documents are then interpreted as having a similar content (McCain, 1990; White and McCain, 1998). Boyack and Klavans (2010, p. 2390) note that \"Co-citation analysis was adopted as the de facto standard in the 1970s, and has enjoyed that position of preference ever since\".The present article brings significant contribution not only because it is the first to apply knowledge-stock, citation and co-citation analysis to the JOCM, but also because it completes already existing qualitative studies about OCM research.The paper is divided into four parts. The first is a review of bibliometric studies; the second details the methodology; the third lists and discusses the results; the fourth part presents the conclusion of the study.\nLiterature review: Several bibliometric studies can be reported in the various fields of management sciences (see Table II).While most bibliometric works use both citation and co-citation analysis, some of them propose a deep citation analysis (Baumgartner, 2010; Tahai and Meyer, 1999). Co-citation methods vary from absolute co-citation counts (Gregoire et al., 2006) to Pearson's correlation coefficients (Calabretta et al., 2011; Georgi et al., 2010) and principal component analysis techniques (Charvet et al., 2008; Fernandez-Alles and Ramos-Rodriguez, 2009; Ponzi, 2002). Although the majority of citation and co-citation analysis focus on the most cited documents, some of them target the most cited authors (Baumgartner, 2010; Nerur et al., 2008) or keywords (Artto et al., 2009).To our best knowledge, no bibliometric study has been done on OCM research. It might be due to the main methodological impediments inherent to the analysis of its intellectual structure. Contrary to the data sources of Ramos-Rodriguez and Ruiz-Navarro (2004) or Calabretta et al. (2011), a single journal cannot stand for the whole discipline: a multiple-journal bibliometric study looks unavoidable. Contrary to the multiple-journal analysis of Backhaus et al. (2011) on BtoB marketing, there is no clearly identifiable group of peer-reviewed journals publishing around OCM. The retained pool of journals would be debatable.Indeed, OCM can be addressed in many famous peer-reviewed journals like The Academy of Management Journal, The Academy of Management Review, Administrative Science Quarterly, Management Science, Industrial and Corporate Change, Journal of Change Management, Journal of Organizational Change Management, Journal of Operations Management, Strategic Management Journal, etc. Other methodological challenges are likely to arise while using bibliometrics on OCM because a single database may not provide all necessary data in the same format for all the retained journals and a tedious work would be required to filter the articles about OCM.The present research is therefore a first step prior to an OCM bibliometric study. To our knowledge, no bibliometric study has yet been done on the JOCM. This gap is intended to be filled by the present paper. The choice has been made to provide a knowledge-stock analysis, followed by a citation analysis and a co-citation analysis of the most cited-documents in the JOCM articles, over the last seventeen years.\nMethodology: The data\nResults and discussion: Knowledge-stock analysis: JOCM characteristics\nConclusion: Summary\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this study is to identify the documents which have had the greatest impact on the Journal of Organizational Change Management (JOCM) articles and to analyze the evolution of the intellectual structure of the journal.\nDesign/methodology/approach: - A knowledge-stock analysis is performed to assess major trends of the JOCM. A bibliometric study is then conducted thanks to citation and co-citation analysis about the documents which are the most cited by the articles published in the JOCM (between 1995 and 2011).\nFindings: - Through the results of their analysis, the authors: describe the growing stock of knowledge of the JOCM over time; identify the documents having the strongest influence on the JOCM articles; and pinpoint the evolution of the intellectual structure of the journal.\nResearch limitations/implications: - Although the sample of retained articles seems representative of the JOCM publication efforts, the data set presents some limitations. There are also some limits inherent to the research design and to the bibliometric methods. The intention of the present research is to give a quantitative overview of the intellectual evolution of the journal.\nPractical implications: - Grasping the intellectual development of the JOCM enables researchers and practitioners to better understand how issues are being approached by authors who publish in this journal. It also stimulates the scholarly debate.\nOriginality/value: - This knowledge-stock and bibliometric study is the first to be concerned with the JOCM.\n",
"role": "assistant"
}
] |
Retail bargaining in China
|
[
"Bargaining",
"Retailing",
"Negotiation",
"Bargaining power",
"Bargaining goal",
"Field study",
"Negotiating"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Consumers in most post-industrial countries are seldom given the opportunity to negotiate prices (Dwyer, 1984). Most purchases are dominated by giant retailers that set fixed prices for their products. Prices are negotiable only for expensive items (e.g. automobiles and apartments), or in limited settings such as garage sales or flea markets (Herrmann, 2004; Maynes, 1976). When retailers must transact with a large number of customers on a daily basis, a flexible pricing strategy is virtually impossible to implement. Negotiable pricing strategies are therefore more common among smaller retailers who often serve as both owner and salesperson and who make a relatively small number of daily transactions. In many Asian countries such as China and Singapore, these conditions arise frequently (Kaynak, 1988; Lee, 2000). For instance, although giant retailers like Wal-Mart have gradually established their presence in China since China\nConceptual framework: In retail price negotiation, the interests of retailer and consumer are usually negatively correlated. Retailers set higher prices to reap greater profits whereas consumers look for a good bargain. Negotiations are often characterised by an unequal distribution of power, and the concessions made by the negotiation parties are a function of this power. Thibaut and Kelley (1959) defined the power of A over B as the ability of A to affect the quality of outcomes attained by B. According to the power dependence theory (Emerson, 1962, p. 32):\nResearch methodology: Data collection\nAnalysis and results: Antecedents of consumer bargaining goal\nDiscussion and conclusion: Major findings and contributions\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine how consumers set bargaining goal when they negotiate price of relatively less expensive shopping products with salesperson. The impact of bargaining goal on bargaining outcomes is also discussed.\nDesign/methodology/approach: To increase realism and external validity, actual transactions in the shopping malls of two cities of China were observed unobtrusively, followed by a matched survey with the consumer to collect data on variables that could not be observed.\nFindings: Product knowledge, presence of purchase plan and shopping companion, posited to reflect consumer\nOriginality/value: A different research methodology is used to study retail price negotiation. Instead of only using student subjects to understand how consumers negotiate price in a hypothetical setting, the current study collected data from actual transactions. This enables us to study how contextual variables such as shopping companion and purchase plan affect negotiation goal which in turn influences satisfaction and future patronage. A comparison of the reference standards adopted for evaluating bargaining outcome is also performed.\n",
"role": "assistant"
}
] |
Formal accounting planning in SMEs: The influence of family ownership and entrepreneurial orientation
|
[
"Family business",
"Entrepreneurial orientation",
"Management accounting",
"Formal planning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Management accounting practices have gained increasing attention in small business and enterprise (SME) research, and publications on the subject have highlighted the importance of management accounting for SMEs. For example, in their recent extensive review of the literature, Lavia Lopez and Hiebl (2015) concluded that management accounting often has a positive effect on SME performance. Other studies have found that various specific management accounting practices, such as the use of formal planning (Peel and Bridge, 1998), budgeting (King et al., 2010), and control of financial measures and non-financial measures (Stede et al., 2006) have a positive influence on SME performance. However, management accounting is a broad concept (Chenhall and Langfield-Smith, 1998), and in the present study we will focus on the \"formal planning dimension\" of management accounting. Although formal planning has been found to be important for SME performance (McKiernan and Morris, 1994; Peel and Bridge, 1998; Wijewardena et al., 2004), many SMEs are inferior in their formal planning processes (Stonehouse and Pemberton, 2002). This makes it relevant to examine the planning practices of SMEs. Previous studies on planning in SMEs have often been restricted to focusing on mainly strategic planning (McKiernan and Morris, 1994; O'Regan and Ghobadian, 2002; Stonehouse and Pemberton, 2002). However, in management accounting research, planning involves short-term planning as well as long-term planning, and we will address both dimensions of planning. In the present study, short-term planning and long-term planning are considered from a management accounting approach, i.e. short-term planning mainly concerns the use of various budgets with a time-frame of up to a year whereas long-term planning concerns longer periods of time. Although planning in SMEs has been examined in several previous studies (e.g. Honig and Karlsson, 2004; Honig and Samuelsson, 2012; Lyles et al., 1993), few studies have examined formal planning in SMEs from a management accounting standpoint (Lavia Lopez and Hiebl, 2015).\nTheory and development of hypotheses: Formal planning and family ownership\nMethod: Data collection and sample\nResults: The average firm in our sample had 41 employees and was 31 years old. Family firms constituted 66 percent of the sample. Table I shows the correlations among the variables in the study. The results from the correlation analysis did not indicate any serious multicollinearity. The highest variance of inflation factors (VIFs) for all our variables was 1.1. Also, it is not likely that the assessment of whether or not a firm is a family firm is likely to correlate to answers concerning formal planning. Harman's single-factor test on the data collected from the survey (i.e. EO, short-term planning, and long-term planning) resulted in 31 percent explanation for a single-factor solution, indicating that same-source bias is not a major issue.\nDiscussion: In this paper we have examined factors, other than size or growth, that are likely to influence the use of formal long-term planning and formal short-term planning in SMEs. Previous studies on EO (Zahra et al., 2014) and family business (Senftlechner and Hiebl, 2015) have called for more research on these issues, and our aim has been to answer those calls. Our finding of a negative relationship between family firms and the use of formal planning supports the established belief about this relationship, but our study is one of the first to empirically validate this notion. The relationship between family ownership and short-term planning should, however, be interpreted with some caution, as the significance is at the 0.1 level. The positive relationship between EO and formal planning does, however, contradict some previous studies and assumptions, and is an important contribution - and perhaps provides a new insight - into the EO literature.\nConclusion: In this paper we have validated the idea that family firms use formal long-term planning, and plausibly (p\n",
"role": "user"
},
{
"content": "Purpose: Several studies have highlighted the importance of management accounting practices such as formal short-term planning and formal long-term planning for SME performance. However, few studies have considered what actually explains differences in the use of formal planning (from a management accounting approach) in SMEs. Family ownership and EO are two plausible explanations for such differences. The purpose of this paper is therefore to examine how family ownership and EO are correlated to the use of formal short-term planning and formal long-term planning in SMEs.\nDesign/methodology/approach: In this study, the authors examined how family ownership and entrepreneurial orientation (EO) affect the use of formal planning by analyzing a sample of 156 Swedish manufacturing SMEs, using multivariate regression analysis.\nFindings: As could be expected, the authors were able to validate the notion that family firms use less formal planning than non-family firms. However, in contrast to some previous studies, the authors found that there is a strongly positive relationship between EO and the use of formal short-term planning and long-term planning.\nOriginality/value: Whereas many previous studies on family business have assumed that family firms use less formal planning than non-family firms, the present study is one of few to actually confirm this notion. Also, this study has provided strong evidence that EO is positively correlated to the use of formal planning, in the short term and in the longer term.\n",
"role": "assistant"
}
] |
School leadership, social justice and immigration: Examining, exploring and extending two frameworks
|
[
"Immigration",
"Educational leadership",
"Social justice"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Immigration is changing world and local demographics - and it is unclear whether or not school leaders understand the implications that these glocal changes have for their schools and communities (Brooks and Normore, 2010). Certainly, many school leaders are aware of seemingly related classroom dynamics via rich lines of inquiry and practical strategies related to diversity (Dantley and Tillman, 2006), multiculturalism (Banks and Banks, 2009), culturally relevant pedagogy (Ladson-Billings, 2009), social justice (Ayers et al., 2000), multilingualism (Walker and Dimmock, 2002), cultural pluralism (Lopez, 2001), and border-crossing (Alston, 2004). Yet while many scholars have investigated how these concepts influence the practice of leadership in schools, few have explored immigration or connected these related concepts with the growing body of literature related to leadership for social justice (Brooks and Miles, 2010; Capper, 1993; Horsford, 2011; Jean-Marie et al., 2009; Normore and Brooks, 2014; Theoharis, 2009).\nLeadership for social justice: equity and excellence in education: Over the past three decades, educational leadership scholars and practitioners have eschewed traditional leader-centered and difference-blind perspectives on their work, and have instead advanced the notion that principals (and other leaders in schools) can be civil rights leaders for communities and schools (Jean-Marie et al., 2015). This entails a rejection of functionalist approaches to leadership and recognition that leaders' work influences (and is influenced by) dynamics such as racism, sexism, ableism, ageism, classism, religiosity, and various other forms of discrimination and oppression (Brooks and Normore, 2010). It is therefore a leaders' responsibility to be leading with these issues in mind, so as to better advocate on behalf of students, educators and community members whose educational experience is negatively affected by these socio-political forces (Normore and Brooks, 2014). Given that each of these dynamics looks different in every community and school, there is no \"one size fits all\" way to think about or lead for social justice - it is an approach to practice that demands an ongoing cycle of deconstructing inequity, reconstructing more just processes and outcomes, and leading in a way that establishes, sustains and evaluates the intended and unintended consequences of this work.\nLeadership for social justice: the educational domain: Leaders for social justice practice the more traditional aspects of the work with an orientation toward positively influencing equity, in addition to being efficient and working toward high-level processes and outcomes.\nSynergistic leadership for curriculum, assessment and pedagogy: Curriculum leadership, assessment leadership and pedagogical leadership are interrelated aspects of school leadership. They should be considered as complementary aspects of a whole, and it is critical that leaders for social justice strive to realize equitable processes and outcomes in each of these areas. Moreover, as they exert reciprocal influence on one another, changes in any one of these areas necessarily changes aspects of the other two (Capper et al., 2006). Many schools will emphasize one of these areas over another, or neglect to consider the ways in which they are related, but being mindful of their relationships is key to leading in the Educational Domain.\nLeadership for social justice: the equity domain: Leaders for social justice must also develop other aspects of their practice, relative to the ways that equity is conceived and operationalized in schools.\nSynergistic leadership for critical consciousness, knowledge and skills: As with the Educational Domain, the Equity Domain will only be effective if leaders understand, explore and practice critical consciousness, knowledge and skills in concert. It is important to understand that developing and improving in one or two of these three interrelated areas is unlikely to allow for holistic leadership development.\nBringing equity and education together through leadership: As the Equity and Education Domains have dynamic components of their own, it is likewise the case that they are connected with each other in reciprocal influence. Education and equity inform each other in a complex manner that is also idiosyncratic to each school, community and educational system. For example, a leader may realize a need to do more to address issues related to poverty in the school and community. Their strategies may include reforming the curriculum to include materials that help students understand the issue, helping to develop assessment strategies that are appropriate while also developing their skills and knowledge and critical consciousness around this aspect of education.\nLeadership for social justice: the importance of time: Many leadership for social justice scholars have argues that context is critical to the practice of equity-oriented education (DeMatthews and Mawhinney, 2014). However, we argue that history and the future are just important. In the following sections, we explain the importance of each.\nSupport for immigrants: an ecology of nested systems: While leadership for social justice is important, it is an incomplete theory of leadership by itself. In order for it to have meaning in an educational context, it needs to be coupled to other concepts (Brooks et al., 2008), and in this case we are interested in ascertaining the possibility of leadership for social justice in a manner that supports immigrant students. To do so, we chose to use a framework developed by the American Psychological Association's Presidential Task Force on Immigration (2012). This was especially appropriate, as the framework offered in this report was a synthesis of research on the topic (Figure 2). The framework is a nested model that situates support of each immigrant at the center and suggests that support is manifest at four system levels: microsystems, mesosystems, exosystems and macrosystems. In the following sections, we explore and explain each.\nMicrosystems: Microsystems are those that occur in the immigrant's local context - that is, they are the systems of support that they experience every day in their community (Furman and Shields, 2005). These can be grouped into four categories of support: workplace, school, families and peers. To be sure, some communities will have additional forms of support but for the purposes of this discussion we will use the existing framework.\nMesosystem: As the American Psychological Association (2012) report explains, \"the mesosystem is made up of interconnections between microsystems, and these interactions have an indirect influence on the individual. For example, the relationship between the immigrant child's parents and school, or consultation provided by the local mental health agency to the child's school, may have implications for the kinds of services the immigrant child will receive from the school\" (p. 11).\nExosystem: The exosystem \"is an extension of the mesosystem and is made up of interconnections between settings and larger social structures, such as governmental entities that have influential but indirect effects on the individual's life\" (American Psychological Association, 2012, p. 12). The exosystem is where immigration policy (and other policies) mediates the implementation of services and supports at the local level. In countries like the USA, which has a federal system, the exosystem is also where state and national policies interact with each other to shape implementation (Spring, 2016; Chubb and Moe, 2011).\nMacrosystem: As explained in the American Psychological Association (2012) report, \"the macrosystem is the most distal context but nonetheless has implications for well-being. For immigrants, these include the cultural, economic, and historical contexts in both the sending and receiving nations\" (p. 12). Importantly, issues related to xenophobia and these three other dynamics from both the country of origin and the new country shape an immigrant's experiences.\nLeadership for social justice and immigration: exploring key concepts from two frameworks: In this section, we briefly discuss how concepts form the two frameworks we have discussed are manifest in each of the articles in this special issue of the International Journal of Educational Management.\nLeading schools with migrant children in Shanghai: understanding policies and practices: Qian Haiyan and Allan Walker analyze publicly accessible policy papers and literature to provide a contextualized interpretation about the status-quo and challenges facing leaders of schools that enroll migrant children. In doing so, they explore several issues related to immigration policy and practice, and also touch leadership for social justice. Their article focuses on internal migration in China, particularly recent trends related to those moving from rural environments to urban centers. Their analysis brings to light issues at the exosystem in that national policies were having a tremendous negative influence on immigrant support until they were changed in the early 2000s. They discuss school leadership as policy implementation, but suggest that values related to the rights of children are at the forefront of school leadership. As the paper is focused on specific policies, it does not explain day-to-day implication of the policies for school leadership practice.\nSchool leadership, social justice and immigration: reflection, extension and conclusion: In this paper, we explained and examined two frameworks. The first helped us understand leadership for social justice. We identified an Educational Domain and an Equity Domain that interact with each other and noted the importance of leaders' understanding the ways that their work is facilitated and constrained by history, context and the future. The second framework looked at the ways that immigrants are supported. This was organized around a series of related systems that interact with one another: the microsystem, mesosystem, exosystem and macrosystem. Considered in isolation, each has the potential to help us understand immigration and leadership as complex theoretical constructs that can help guide empirical inquiry and even help improve practice. Considered in relation to one another, these frameworks highlight that there is a theoretical disconnection between these two concepts. That said, looking at the articles in this special issue in relation to each framework suggested promising possibilities, while also highlighting areas where further theoretical and empirical research is necessary.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore theoretical connections between educational leadership for social justice and support for immigration. The authors seek to identify strengths, weaknesses and opportunities for further study and improved practice.\nDesign/methodology/approach: This is a theoretical research paper that introduces, evaluates and expands two frameworks for understanding leadership and immigration.\nFindings: Findings suggested that there is a need for educational leadership scholars to more purposefully investigate issues related to social justice and immigration.\nOriginality/value: This study offers a novel theoretical perspective on leadership, social justice and immigration.\n",
"role": "assistant"
}
] |
Economic growth, public expenditure on health and IMR in India: An econometric investigation of causal linkages
|
[
"India",
"Economic growth",
"Granger causality",
"Panel cointegration",
"Infant mortality",
"Public expenditure on health"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The relationship between economic growth and public expenditure on health and subsequently its impact on health outcomes is an important subject of analysis and debate in recent years. Researchers have investigated this linkage by employing a variety of econometric tests and the results are confounding (e.g. Deolalikar, 2005; Bhalotra, 2007 have opposite results).\n2. Literature review: The literature investigating the relation between economic growth, public spending on health and health outcomes have explored various facets and variables. According to the World Bank (1993), improvements in health situation may contribute to economic growth in multiple ways. This maybe reduction in production laws caused by labor illness; increased in education rate because of healthier children; use of natural resources that would otherwise be inaccessible because of diseases and freeing of alternative uses of resources that would otherwise be spend on medication. Researchers have found evidence for all these linkages in different studies (Gonzalez-Paramo, 1992; Barro and Sala-i-Martin, 1995).\n3. Research gap and the Indian context: To summarize the literature, researchers explained the impact of health on economic growth, economic growth on health spending and economic growth on health outcomes. However, there is only one study to the authors' knowledge that focuses on the tripartite relation between economic growth, public spending on health and health outcomes. Day and Tousignant (2005) explored the dynamic relationship between three variables: an indicator of health outcome, including IMR, real per capita income (as measured by GDP or GNP) and real per capita spending on health, using time series data for Canada. The results derived suggested the evidence of a weak statistically significant relationship between per capita health spending, health outcomes and per capita GDP. The absence of a strong statistical relationship may be due to model misspecification or may reflect the fact that at high levels of population health, the returns to increase in health spending are small. Given the background, the present study aims to address the following objectives: to explore the causal relation between economic growth and public expenditure on health; to investigate the bi-directional causality between public expenditure on health and IMR; and to explore the two way causal linkage between economic growth and one of the health outcome indicators, namely, IMR.\n4. Data set and variables: In order to achieve the study objectives, panel data for 16 major states spanning a period from 1990 to 2010 (20 years) were selected and analyzed. Choice of variables in econometric studies if not a difficult exercise still requires the backing of relevant economic theory without which the forecasting power of the inference would be suspect. In the present context, the literature review suggests economic growth and public expenditure on health are integrated, as are public expenditure on health and IMR. Further, it also suggests a linkage between economic growth and IMR. Accordingly, per capita GSDP of the states, and per capita public expenditure on health at the state level and the respective IMR of those states constituted the relevant variables of the study. The 16 major Indian states are considered that constitutes approximately 83 percent of the total GDP in India. The number of states is confined to the major states since consistent time series data on the study parameters are available that is essential for empirical investigation. The choice of the initial year is 1990-1991 since it was the beginning of liberalization and the introduction of economic reforms.\n5. Methodology: The econometric analysis of the correspondence between economic growth and public expenditure on health and between economic growth and IMR and public expenditure on health and IMR for the period 1990-1991 to 2010-2011 was conducted in three different stages taking two pairs of variables at a time (GSDP and health expenditure; health expenditure and IMR; and IMR and GSDP). In the first stage, the series was checked for stationarity. In the second stage, cointegration analysis was performed to explore the long-run behavior of the variables. In the third stage, in case the variables were found to be cointegrated, the Granger causality test was applied to study the endogeneity/exogeneity of the variables under study.\n6. Results: The results of the econometric analysis are discussed in this section. This section reports the stationarity and unit root tests, Pedroni cointegration tests and the Granger causality results of the panel data in three subsections. Before presenting the detailed time series results, the group correlations and descriptive statistics were analyzed (Table I). Followed by this, a set of cross-sectional independence tests were run to examine whether the data from the cross-sections (i.e. states) were independent. All cross-sectional independence tests indicated that the cross-sections were independent from each other (Table II).\n7. Discussion: The present study aimed at testing the causal relations between economic growth and public spending on health and its corresponding impact on one of the major health outcomes, namely IMR. The study was conducted in the Indian context based on panel data for sixteen major states for the period 1990-1991 to 2010-2011. The results were interesting and to some extent counterintuitive.\n8. Policy implications: The present study has quite a few implications for the policy makers. The first and foremost policy implication is the need for including health capital in the growth policy. In many cases, the growth policy of a nation may be revolving around non-human capital formation that is expected to create a \"trickle down\" effect on human capital. Based on the present study, it becomes evident that funding allocation in health over a considerable period of time is required for human development and thus should be a priority in public policy formulation. The study findings suggest more fund allocation in the government budgets, toward health sector. Since public health expenditure was found to affect health outcomes, more fund allocations are expected to improve the overall health conditions. The finding that public health expenditure does not affect health outcomes in the short run has two implications. First, it implies that the realization of health outcomes may be a long-run phenomenon and thus consistent investments in health and human capital is necessary for long-term positive health outcomes. Second, it calls for a private role in health provisioning and improvement of health status. In this regard, public-private partnership can be considered as an alternative toward health care delivery from the public sector. The finding of IMR nor affecting economic growth should not be taken as a deterrent to allocation of public spending on health and the span of introspection of the effects of growth on IMR and vice versa should be increased that would allow more relevant outcome of the phenomenon. This also implies that the government should try to investigate the reasons why the effects are not realized and take remedial measures to get the desired results.\n9. Conclusions: The present study aimed to investigate the tripartite causal linkage between economic growth, public expenditure on health and IMR in India. The study revolved around the analysis of state-level data for a 20-year period of the given study variables. Major findings indicated a significant short run effect of economic growth only on health expenditure. In the long run, there was a bi-directional causality between economic growth and health expenditure. Health expenditure and economic growth were both found to reduce infant mortality in the long run, but not the reverse. Thereby, the study supported the role of growth and public health expenditure on improving health outcomes. However, it contradicted the notion of health outcomes influencing growth (as per the social rate of return theory).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to employ a two-step approach to investigate the bi-directional causal linkage between: economic growth (measured by GDP) and public expenditure on health; public expenditure on health and infant mortality rate (IMR); and economic growth and IMR in the Indian context.\nDesign/methodology/approach: The present study uses econometric analysis, namely, panel cointegration and Granger causality on 20-year panel data on 16 major Indian states to investigate the causality.\nFindings: The results suggest GDP to Granger cause public expenditure on health both in the short run and in the long run, but public expenditure on health to Granger cause GDP only in the long run. Further, public expenditure on health and economic growth were found to Granger cause IMR in the long run. However, the reverse linkage from IMR to public expenditure on health and/or economic growth was not significant.\nResearch limitations/implications: The present study provides support to the existing literature on the effects of economic growth on health expenditure and health outcomes but also raises a question on the time required to realize the same.\nPractical implications: The findings have implications for policy makers on the time frame and application of health expenditure to achieve better results.\nOriginality/value: The present study is one of the first to test the tripartite linkage between economic growth, public health expenditure and health outcomes at a state-level analysis.\n",
"role": "assistant"
}
] |
Cancellations of elective orthopaedic procedures: Compliance with NHS Modernisation Agency guidelines
|
[
"Orthopaedics",
"Clinical audit",
"Patients",
"Medical care",
"Medicine practice",
"United Kingdom"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Preoperative assessment plays an important part of preoperative planning.. Adequate preoperative assessment helps to identify the medical conditions and gives an opportunity to optimise the medical conditions that may endanger the patients' safety during surgery and also in the postoperative period (Clelland et al., 1996). Cancellation of an elective case is \"wasteful expenditure\". This not only involves another needy patient being pushed behind in waiting list, but also costs time and money. In the event of cancellation of a case, although there is no direct cost to the hospital for the surgery not performed, there is a \"wastage\" of valuable resources. (MacDonald et al., 1992) as equipment and time lies unutilised. Staff is paid without any productive output, and an in-patient bed is unnecessarily occupied for a day or two. Such expenditure is termed as \"perceived\" expenditure. This could be about PS282.25 per patient per day. From the patient's perspective, it brings inconvenience and suffering, as the patients have to reschedule their appointments both at home and work place. There is also a psychological and physical preparation, involved in preparation for surgery, which needs to be repeated at a later date. In last few years, many hospitals across UK, the routine preoperative assessment of an elective case is being done by nurse practitioners. They are specially qualified and trained to do this job. Rai et al. (2003) have shown that a nurse-led pre-assessment system with a targeted referral system can be effective in decreasing the cancellation rate in an elective surgery.We conducted a retrospective study of the adequacy of preoperative assessment for all elective orthopaedic surgeries cancelled after admission, during a 24-month period, at a district general hospital. The results of the study are presented and recommendations are suggested to reduce the rate of such cancellations.\nAims and objectives: The aim of the study was to identify the medical causes of cancellations of elective orthopaedic surgeries and scope of reducing cancellations by improved preoperative assessment.The following objectives were devised to achieve this aim:* To review the medical reasons for cancellation of elective orthopaedic surgeries at a district general hospital.* To see if adequate preoperative assessment has been done in each case whose operation has been cancelled on medical grounds.* To identify specific deficiencies in the local preoperative assessment protocol practiced at the hospital, with respect to the chosen \"benchmark\".* To suggest improvements which are likely to have the maximum impact on the cancellation rates.\nMaterial and methods: The guidelines issued by NHS Modernisation Agency, (2003) National Good Practice Guidance on Preoperative Assessment for Inpatient Surgery were considered as the benchmark against which the hospital protocol and practice is compared. The case notes of elective surgeries cancelled after admission from January 2003 to December 2004 were reviewed. The adequacy of preoperative assessment was ascertained from the pre-assessment forms, which were cross-tallied with the anaesthetic notes. The parameters noted were: the proposed operation, reason for cancellation, person responsible for cancellation, adequacy of preoperative assessment, and the specific aspect of the NHS Modernisation Agency's guideline that was not followed which otherwise, could have potentially avoided cancellation in each particular case. The exclusion criteria were lack of theatre time, cancellation by patient, absence of surgeon, intervention by acute surgical emergencies and any other non-medical reason.\n\"Benchmark\" for this study - NHS Modernisation Agency's guidelines: The salient features of NHS Modernisation Agency's (2003) National Good Practice Guidance on Preoperative Assessment for Inpatient Surgery guidelines are:* A paramedical assessor is used, who is empowered to order basic investigations and make referrals as per protocol. He/she is trained in basic cardiovascular and respiratory examination, as well as airway assessment.* This paramedical assessor acts as a central point of contact for interdisciplinary coordination.* An initial assessment is done immediately after the decision for surgery is made.* The criteria for \"fitness\" are agreed locally.* An explicit protocol for referral is drawn up.* A multidisciplinary team of anaesthetists, surgeons, intensive care specialists, and primary care physicians develops all these guidelines.* The criteria for fitness should be satisfied before adding a patient to the waiting list, or booking a surgery date.* Patients on the waiting list are contacted six weeks before surgery, for reassessment if necessary.* All patients are contacted two weeks before surgery date to confirm their medical and social circumstances, and to arrange tests near the time of surgery, if required.\nResults and discussion: A review of the hospital computer records showed 1,110 elective cases were operated on by Orthopaedic department at our hospital between January 2003 to December 2004. During this period, 44 elective operations were cancelled due to medical reasons following their admission into the hospital for surgery. This constitutes 3.96 per cent of all elective orthopaedic admissions. All of these admissions were classed under CEPOD 4 category, i.e. elective cases to be done at a time convenient for both, the patient and the hospital. These included 27 joint replacements, five forefoot surgeries, three arthroscopies, and nine others (see Figure 1).The cancellation rate for elective orthopaedic surgery in our hospital (3.96 per cent) is small compared to those in other published series (Jones et al., 1997; Wildner et al., 1991)) Of these cancelled cases, 71 per cent of them were cancelled by the anaesthetist. The consultant orthopaedic surgeon was responsible for cancelling in 20 per cent of the cases. In 9 per cent of the cases, the person responsible for cancellation was not apparent from the notes. The Cardiac causes (34 per cent) topped the list of conditions responsible for cancellation of a case. This is followed by other medical conditions such as high blood pressure (25 per cent); high INR (7 per cent), and others (34) per cent. Therefore the cardiac and vascular causes together accounted for more than half the cancellations. Study by Hand et al. (1990) also confirms that cardiovascular condition constituted an important reason for cancellation of an elective case. The other reasons for cancellations in this study were unique to each case. These for example included: acute laceration at incision site, active psoriasis at the operative site, rash in groins with swollen leg in case of elective hip replacement, phlebitis of leg, low platelet count, and hand surgeon referral for expert opinion.The locally agreed policy allows the examining clinician to place the patient on the waiting list based purely on the patient's orthopaedic problem. Therefore most of our 1,110 patients did not have any initial assessment before their name was placed on the waiting list for surgery. The first preoperative assessment is normally done only six weeks prior to the date of intended surgery. Only in exceptional cases, an anaesthetist's opinion is sought prior to listing the patient. A total of 28 out of 44 cancelled patients were admitted to the orthopaedic ward a day prior to the surgery without having any preoperative assessment. In three cancelled cases although a preoperative assessment was done, a relevant past history (which was the medical cause of cancellation) was not recorded. In the remaining 13 cancelled cases although complete documentation of medical history was made, no further opinion was sought from the senior team member in regards to the optimisation of the patient prior to the surgery. All these 13 cases had medical conditions that justified an expert opinion. An appropriate referral for optimisation of patient and an opinion of an anaesthetist would have prevented the cancellation.Our observation showed that, 11 patients in all were cancelled due to high blood pressure, and 15 patients were cancelled due lack of cardiologists opinion. However, we understand the benefit of anaesthetic input in avoiding cancellations due to high blood pressure is limited (Rai and Pandit, 2003; West and Galasko, 1995). This is because the stress of impending surgery could raise the blood pressure even in a well-controlled patient. However in 15 cases due to cardiac reasons, a cancellation could have been prevented. They could have been either optimised prior to the surgery, or not included in the waiting list as they are not fit to undertake a major surgery.Three cases were cancelled due to high INR. All these patients have not stopped their Warfarin prior to admission. Incidentally all three patients did not have any preoperative assessment. One case was cancelled due to inappropriate listing, which was later referred to an expert in another hospital. In this case the listing was done by a non-consultant. In such cases an opinion of a senior surgeon prior to the listing of the patient would have avoided a cancellation. Other reasons like cancellations due to laceration at operative site, phlebitis, rash and active psoriasis cannot be avoided by adequate preoperative assessment as these all occurred with in two weeks of the date of surgery. Though there are locally agreed protocols for referral of a case to senior medical staff for opinion during pre-assessment by nurse practitioners in our hospital, there is a clear deficiency in regards to their use. Awareness, education and empowerment of the nurse practitioners could solve this problem.In 28 cases where there has been a cancellation of a major case, the patients either did not turn up for the pre-operative assessment or a letter of invitation for preoperative assessment had not been sent by the hospital. Although an attempt was made to identify the cause for their non-assessment, this was not possible due to lack of the relevant documents. It is also possible that some of the patients did not have any preoperative assessment, as they are asked to come at short notice when another patient suddenly cancelled at short notice. This is the hospitals policy to optimise the use of theatre time. However such invitations at short notice may not work due to non-availability of patients who are fit for surgery. Since most of the cancellations were done by the anaesthetist, there is a possible role for additional anaesthetic input in the pre-operative assessment alongside the orthopaedic nurse practitioners in pre-assessment clinics. However the drawbacks of this could be increased need of manpower, which means increased expenditure for the hospitals. There may also be reduplication of the work unless the roles are not clearly defined. Alternatively a joint clinic with both a junior doctor the orthopaedic nurse practitioners could be an alternative. It should be remembered that an initial assessment before the inclusion of a patient in the waiting list is not a substitute for second assessment six weeks prior to the surgery. In places with long waiting lists the medical condition of the patient may change significantly meanwhile waiting for the surgery.After comparing the local protocol with the guidelines of the NHS Modernising Agency, five major areas of improvement were identified that could have a potential to decrease the cancellation rates of elective orthopaedic surgeries:1. The local agreed protocols should be developed in line with the NHS Modernisation Agency's guidelines.2. An initial assessment that involves complete medical history and relevant blood investigations should be done as soon as the decision for surgery is taken. The patient should be added to the waiting list only if the patients satisfy the above initial assessment.3. Alternatively the patients with cardiovascular, respiratory, haemopoitic or hepatic or renal conditions undergoing major surgery should have an anaesthetic review from a consultant anaesthetist at initial pre-assessment. Alternatively, an explicit protocol for referral should be made with multidisciplinary input, for use by the trained nurse practitioner.4. All patients irrespective of the medical condition should be contacted two to six weeks prior to the intended date of surgery, to confirm their medical and social circumstances have not changed. Arrangements should be made for reassessment or further investigations if necessary near to the date of surgery.5. Junior doctors have to confirm with the consultant concerned before putting the patient's name on the waiting list (Wildner et al., 1991).\nConclusion: Ideally, cancellation rate for elective operations should be zero. However, our experience shows that certain causes of cancellations were unavoidable. The present review of our series shows that about half of the medical causes of cancellations could have been avoided. For this it is important for the local guidelines to be as close to the NHS Modernising Agency's (2003) National Good Practice Guidance on Preoperative Assessment for In-patient Surgery recommendations as possible. Multi-disciplinary teams consisting of orthopaedic surgeons, anaesthetists, out-patients staff and nurse practitioners should be involved in developing the local guidelines. The nurse practitioners should be able to refer the cases to specialist directly if necessary. Invitations for surgery at short notice to patients who have not been pre-assessed and optimised unless the patient is young do not improve the cancellation rates of elective operations. Cancellations due to medical reasons can in general be avoided by meticulous preoperative assessment, appropriate referrals, and optimisation of the patient before intended surgery.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to find the incidence and medical reasons for cancellations of elective orthopaedic cases following admission for an operation in a district general hospital. The paper also aims to determine the deficiencies in the local preoperative assessment protocol.\nDesign/methodology/approach: - This is a retrospective study. The elective orthopaedic surgeries cancelled following their admission into the hospital due to medical reasons between January 2003 and December 2004, were identified. These cases were reviewed using the preoperative assessment charts and case notes. The NHS Modernisation Agency's guidelines, National Good Practice Guidance on Preoperative Assessment for Inpatient Surgery, are taken as the benchmark for comparison.\nFindings: - The paper finds that 44 elective orthopaedic cases were cancelled due to medical reasons. Of these patients, 64 per cent did not have a pre-operative assessment prior to the admission for the planned surgery; 6 per cent had inadequate documentation of relevant past medical history; and 30 per cent (patients with significant medical problems) were not referred to the anaesthetist for advice although these problems were identified during the pre-operative assessment.\nResearch limitations/implications: - It is possible to stimulate good medical practice through audit.\nPractical implications: - Cancellation of an elective operation is not uncommon in hospital practice. This paper aims to highlight the possible avoidable causes for such cancellations. The paper identifies such deficiencies in the local preoperative assessment protocol and suggests remedies to improve the quality of care. These improvements and close adherence to the guidelines are important as preoperative assessment by nurse-led clinics are being increasingly practised through out the UK.\nOriginality/value: - The paper fulfils its aims of identifying the medical causes for cancellation of an operation and also the existing deficiencies in the preoperative assessment practice. The paper values the importance of adherence to NHS Modernising Agency's guidelines in preoperative assessment. It is useful not only to the Orthopaedic department but also to the department managers.\n",
"role": "assistant"
}
] |
Production data analysis system using novel process capability indices-based circular economy
|
[
"Internet of Things (IoT)",
"Intelligent manufacturing",
"I-cloud",
"Process capability indices",
"Production data analysis"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Companies must take into account both their own growth and environmental protection; many studies have consequently been conducted on green energy and the concept of a circular economy (see Guo et al., 2017; Li et al., 2018; Lin et al., 2016; Wang, 2016; Wang et al., 2018; Wu et al., 2010). Chen, Wang and Tan (2019) pointed out that sustainable production can be achieved by the enhancement of product quality, which has the added benefit of increasing product value and industry competitiveness. Increasing process quality lowers scrap and rework rates, lengthens product life span, and reduces product maintenance, which can in turn reduce environmental pollution. As technologies associated with the Internet of Things (IoT) have gradually matured, the measurement and collection of production data have continued to advance, enabling the collection of big production data. Effective data analysis and application can enhance manufacturing technologies and management technologies, which not only accelerate the development of intelligent manufacturing for Industry 4.0 but are also conducive to the improvement of process quality.\n2. Theoretical background of the production data analysis system: In this section, we first present a theoretical background for process capability indices. We then review literature relevant to statistical hypothesis testing.\n3. Cloud-based quality control case study: With economic progress, the idea of physical exercise as a form of recreation has risen in popularity (Chen et al., 2012). This has enabled a flourishing of the sports bike industry. This has been further aided by increasing environmental awareness. Bicycling is widely recognized as an eco-friendly, economic, and healthy means of transportation, and a number of cities around the world have established bike-sharing mechanisms. Therefore, our case study focused on a bottom bracket bearing manufacturer in Central Taiwan. The manufacturing process of bottom bracket bearings contains eight important quality characteristics: seven NTB quality characteristics and one STB quality characteristic (surface roughness (S)). Details are displayed in Table III.\n4. Case execution, results and discussions: Using the IM-IC-WM-execution model, the mechanical parameters of item IM, and the case data in Table III, we performed case execution, which is explained in Section 4.1. Section 4.2 presents the results and discussion.\n5. Conclusions: The IM-IC-WM-execution model employs a production data analysis model to form a loop feedback system for intelligent manufacturing. With the IM-IC-WM-execution model, the quality control centers of bicycle manufacturers can use the cloud platform and the production data analysis model to assist the parts manufacturers and suppliers in the entire industry chain in identifying CTQs that are in need of improvement and also offer reference for improvement. The improvement achievements and experience can also serve as reference for the entire industry chain. The improvement experience is added to knowledge management, thereby enabling parts manufacturers and suppliers improve incoming quality control and raw material homogeneity while also enhancing tool management, formulating SOPs and proposing reasonable maintenance systems to reduce process variation. Poor product design and tolerance design may also be causes of poor quality, so providing such information as feedback to bicycle manufacturers can help them devise more robust product designs, more reasonable tolerance designs, and selection criteria for better parts suppliers.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to establish mechanisms for process improvement so that production efficiency and product quality can be expected, and create a sustainable development in terms of circular economy.\nDesign/methodology/approach: The authors obtain a critical value from statistical hypothesis testing, and thereby construct a process capability indices chart, which both lowers the chance of quality level misjudgment caused by sampling error and provides reference for the processes improvement in poor quality levels. The authors used the bottom bracket of bicycles as an example to demonstrate the model and methods proposed in this study.\nFindings: This approach enables us to plot multiple quality characteristics, despite varying attributes and specifications, onto the same process capability analysis chart. And it therefore increases accuracy and precision to reduce rework and scrap rates (reduce), increase product availability, reduce maintenance frequency and increase reuse (reuse), increase the recycle rates of components (recycle) and lengthen service life, which will delay recovery time (recovery).\nOriginality/value: Parts manufacturers in the industry chain can upload their production data to the cloud platform. The quality control center of the bicycle manufacturer can utilized the production data analysis model to identify critical-to-quality characteristics. The platform also offers reference for improvement and adds the improvement achievements and experience to its knowledge management to provide the entire industry chain. Feedback is also given to the R&D department of the bicycle manufacturer as reference for more robust product designs, more reasonable tolerance designs, and selection criteria for better parts suppliers, thereby forming an intelligent manufacturing loop system.\n",
"role": "assistant"
}
] |
Learning challenges and preferred pedagogies of international students: A perspective from the USA
|
[
"International students",
"Preferred pedagogies"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: International students add richness and diversity to educational institutions around the world (Andrade, 2006; Kwon, 2009). International students come to the USA to fulfill their dreams for higher education and career opportunities (Lin, 2012). These students usually have temporary visas and English is not their native language (Andrade, 2006; Lin, 2012).\n2. Theoretical framework: Bandura's classic theory on social learning emphasizes that individuals absorb learning values from their early environments. Role models (parents, teachers, family members) play a strong role as powerful transfer agents in any learning process. Thus children initially mimic behaviors from such role models and subsequently retain such learning patterns even as adults. Adults continue to exhibit prior learning methods even in the absence of their role models. The cultural environments that individuals are raised in become defining foundations for their future learning norms. These learning standards are simultaneously reinforced as individuals observe others in their environment being rewarded or punished (Bandura, 1971; Myers, 2008).\n3. Literature review: Kwon (2009) did a study on 165 international students in the USA to identify their problems and transition concerns. Students from Asia experienced maximum anxiety as their proficiency in English was inadequate compared to that of other ethnic groups. The more proficient international students were in English, the less social isolation they experienced. In comparison to other ethnic groups, Asian students rated their professors' teaching efforts positively. This positive acknowledgment could also be cultural as individuals from these cultures are trained to be deferential toward authority. Most of these students preferred medium-size classes as this gave them better opportunities to follow the pace and practice their communication skills.\n4. Methodology: This section will identify various aspects involved in the methodology such as research site, survey instrument, data collection, participants, and response rate. The Institutional review board guidelines were followed throughout the process of this empirical research. International students in the USA are defined as \"students who are enrolled at universities pursuing higher education. These students are not citizens, immigrants, or refugees of the USA. They can hold a variety of visa statuses such as F (student) visas, H (temporary worker/trainee) visas, J (temporary educational exchange-visitor) visas, and M (vocational training) visas\" (Korobova, 2012, p. 16). This study included students with an F visa or a student status.\n5. Results: The main research questions guiding this study were:\n6. Discussion: Scholars suggest national cultures play a powerful role in any form of learning (Bandura, 1971; Deresky, 2011). Some cultures emphasize on rote memorization and knowledge recall, while others underscore on analytical and critical skills (Charlesworth, 2008). Today's higher educational environment in the US reflects a medley of students from different cultures putting an extra burden on scholars to understand their audiences better. The debate often becomes whether faculty should accommodate to the different learning needs of their international students or vice-versa (Rodrigues, 2004, 2005).\n7. Conclusion: This empirical study enhances the body of knowledge on international students. It showcases the perceptions and insights of the 111 international students in a comprehensive manner. The contribution of this study is that it identifies responses to several themes such as: learning challenges; learning needs; students' recommendations; and preferred pedagogies (Table IV-IX provide these results). Empirical or qualitative studies usually identify one or more topic related to international students in their learning journey overseas. For instance, (Rodrigues, 2004, 2005; Soosay, 2009) examined preferred pedagogies, (Kwon, 2009; Lin, 2012) analyzed academic and non-academic concerns, (Woods et al., 2006) showcased learning challenges and teaching strategies and (Angelova and Riazantseva, 1999; Wang and Li, 2008) investigated writing concerns. This research provides a holistic picture in an effort to offer guidance to faculty and support to international students.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to understand the challenges international students face during their higher education in the USA. This research study was conducted at a private university in the Eastern region of the USA from October 2012 through May 2013 (Fall and Spring semesters) with undergraduate and graduate international students. International students for this study are defined as those students who are on an international student visa status (F1).\nDesign/methodology/approach: This research adopts an online survey method to understand the challenges international students encounter in a single university. A total of 111 students responded providing a response rate of 33 percent. The results were analyzed using frequencies and percentages.\nFindings: The findings of this study suggest that the primary learning challenges of international students are inadequate writing skills, non-participation, and comprehension abilities. The predominant learning needs are writing academic papers, enhancing class participation, and having local mentors. The students recommend that professors offer detailed feedback, encourage nurturing environments, and provide informal feedback for assignments. The preferred pedagogies are class lectures, class discussions, and individual assignments.\nPractical implications: This study provides implications for both faculty and administrative staff in managing international students better.\nOriginality/value: This paper provides a holistic picture of the challenges international students face during their higher education in the USA. Most of the research on this population focuses only on one of the elements addressed in this research. The results of this empirical study will help guide both local faculty and global academic institutions in understanding this student population better.\n",
"role": "assistant"
}
] |
Patient satisfaction and resident postgraduate year status
|
[
"Patient satisfaction",
"Burnout",
"Postgraduate status",
"Healthcare quality"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Patient satisfaction has gained importance over time especially as a strong healthcare quality indicator. Patient satisfaction can be defined in several ways: expectations, conceptions and patient assumed outcomes (Hyrkas and Paynonen, 2000). It is considered subjective and related to expectations towards the healthcare process and individual patient value systems. Patient satisfaction is used commonly as a surrogate physician-professionalism and communication-skills measure (American Council of Graduate Medical Education, 2011). Because satisfaction is subjective, it is usually measured by patient surveys despite their potential low-sensitivity problems (Streiner and Norman, 1989). It is especially important since patient satisfaction with healthcare significantly affects their behaviors such as adherence to appointments, pharmacological treatments and other therapies (Francis et al., 1969; Linder-Pelz and Struening, 1985).\nPatient satisfaction and physician training: Studies looking at the relationship between patient satisfaction and physician training reveal conflicting results. Yancy et al. (2001) found that internal-medicine physicians had better patient satisfaction scores than internal medicine residents in a Veterans Affairs hospital clinic. However, Jackson et al. (1999) found no difference in the patient satisfaction scores whether they saw an attending or internal medicine resident. Interestingly, there is limited literature on patient satisfaction differences among doctors caring for the patient. We investigated whether there were any differences in patient satisfaction depending on resident postgraduate year status. With a primary emphasis on learning clinical skills during residency, advancement through residency training may yield better satisfaction scores since higher-level residents would have the requisite skills and knowledge to more effectively manage patients. The questionnaire we used examined patients' perceived satisfaction with their doctors. Thus, it could also be possible that a greater patient-load coupled with greater responsibility at higher residency-training levels may translate into lower patient-satisfaction levels.\nResearch question and method: Our institution is large, urban, tertiary-care, teaching hospital located in New York City. The internal medicine residency program is among the largest training program in the US. Up to 140 internal medicine residents are trained in inpatient and outpatient medical settings by 84 attending physicians. All residents have dedicated time for ambulatory clinics along with regularly scheduled continuity clinics. All resident physicians were indirectly supervised by an attending physician who did not see the patients directly. Total patients seen by the residents increase with training levels (third-year residents seeing most). Spanish interpreters are used for Spanish speaking patient encounters (constituting a large proportion) with residents who did not speak Spanish. Though there have been studies addressing patient satisfaction differences between attending and resident providers, none address the differences between residents' training-levels. Our main research question, therefore, addresses that gap:\nDiscussion: Our cross-sectional survey-based study suggests that patient satisfaction (assessed by questionnaire) varies with residents' training-level. The PGY1 residents achieved the highest scores, followed by PGY3. There are several possible explanations for this finding. One might be residents' time with patients. The PGY1 residents having less experience and also fewer patients might spend more time with their patients. However, in our study, there were no significant differences between groups on the time doctors spent with patients. Patients reported the time spent, so this finding may be unreliable since the measure is subject to recall bias.\n",
"role": "user"
},
{
"content": "Purpose: Patient satisfaction has been recognized as an important variable affecting healthcare behavior. However, there are limited data on the relationship between doctor post-graduate year (PGY) status and patient satisfaction with provider interpersonal skills and humanistic qualities. The authors aims to assess this relationship using an American Board of Internal Medicine (ABIM) questionnaire.\nDesign/methodology/approach: Participants were: patients attending a primary care clinic at a large urban academic hospital; and physicians treating them. The survey questionnaire was the ABIM patient satisfaction instrument; ten questions pertaining to humanistic qualities and communication skills with responses from poor to excellent. Mann Whitney U test and multi-variable logistic regression analyses were used to explore score differences by PGY level.\nFindings: The postgraduate year one (PGY1) had higher patient-satisfaction levels compared to PGY2/PGY3 residents. The PGY1 level residents were more likely to score in the 90th percentile and this remained constant even after adjusting for confounders.\nResearch limitations/implications: The research was a single-center study and may have been subject to confounding factors such as patient personality types and a survey ceiling effect. The survey's cross-sectional nature may also be a potential limitation.\nPractical implications: Patient satisfaction varies significantly with PGY status. Though clinical skills may improve with increasing experience, findings imply that interpersonal and humanistic qualities may deteriorate.\nOriginality/value: The study is the first to assess patient satisfaction with PGY status and provides evidence that advanced trainees may need support to keep their communication skills and humanistic qualities from deteriorating as stressors increase to ensure optimal patient satisfaction.\n",
"role": "assistant"
}
] |
Industrial product pricing: a value-based approach
|
[
"Industrial pricing",
"Price setting process",
"Pricing orientation",
"Pricing decision",
"Pricing policy",
"Product costs"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Of the three main approaches to pricing used in industrial markets - cost-based, competition-based and value-based - the third is considered superior by marketing scholars and pricing practitioners. But only a few industrial firms have adopted it: cost-based and competition-based approaches continue to play a dominant role in industrial pricing practice (Coe, 1990; Shipley and Bourdon, 1990; Noble and Gruca, 1999).The marketing and pricing literature is silent on how organizational and behavioral characteristics of industrial firms may affect pricing orientation (Ingenbleek, 2007) and on how pricing orientations may affect the final price-setting process for industrial products. To address this phenomenological gap, we designed a qualitative inquiry based on semi-structured interviews with managers in small- and medium-sized US industrial firms that have successfully adopted value-based pricing. We also interviewed managers in similar firms that have not adopted this pricing approach. By probing the \"lived worlds\" of these executives, we hoped to generate a grounded theory about the organizational practices that contribute to or hinder the development and implementation of modern pricing practices and how they are used in the price-setting process in industrial markets.Our results reflect similarities and differences in the experiences of managers in industrial firms using all three pricing orientations. They contrast firms and leaders with respect to how they organize for pricing, manage the pricing process, make product-pricing decisions, manage the transition to more advanced pricing orientations, and develop internal capabilities to face uncertain and ambiguous decisions.\nResearch design: Theoretical foundation\nFindings: Our data reveal that the decision-making process and the factors influencing price-setting and price-point definition for existing and new products in US small- and medium-sized industrial firms varies dramatically by pricing orientation, i.e. value, competition and cost. We discovered stark differences in the locus of the pricing function, the nature of the pricing process, the organizational structure, the diffusion of pricing capabilities, and in leaders' behaviors in firms with a value-based pricing (VBP) orientation versus those with cost- or competition-based orientations (CBP and COBP).Finding 1: firms using value-based pricing support their product-pricing decisions by relying on formal market research, scientific pricing methods and expert recommendations, while those using other orientations (cost or competition) rely on experience, prior knowledge, gut and intuition\nDiscussion: Our data contradict the widely held assumption (Ingenbleek, 2007) that value-based pricing can simply be \"adopted.\" Rather, they strongly suggest that implementation and internalization of value-based pricing requires deep organizational changes that transform the fabric of the firm. We expected to find significant differences in how small and medium industrial firms organized their pricing functions and how pricing decisions were made - and we did. But we also found stark contrasts among firms with different pricing orientations, related to the firms' organizational characteristics.We begin by discussing the role of intuition and gut feeling in the product price-setting process; then we examine the transformation to value-based pricing from a learning theory perspective.The role of intuition and gut feeling in the price-setting process for industrial products\nLimitations: The findings presented in this paper should be considered in light of several limitations that may affect their generalizability. Our sample of small and medium industrial firms was small (15) and not randomly selected. The sample included firms in only three industrial sectors: building products, transportation products, and plastics and chemicals. A larger sample, and one that includes other sectors such as IT or pharmaceuticals, may yield different findings.Although we paid special attention to the potential risks of researcher bias, it is important to mention that the principal researcher has significant experience in and knowledge about industrial pricing - in particular, value-based pricing. However, we made great effort to remain self-reflective about these risks (Corbin and Strauss, 2008) by using open-ended questions to elicit rich, unstructured narratives of respondents' experiences (Maxwell, 2005, p. 22), interpretations, and understanding of pricing events and firm activities.\n",
"role": "user"
},
{
"content": "Purpose: - This paper seeks to examine the influence of pricing orientation on the price-setting process in industrial firms.\nDesign/methodology/approach: - The authors designed a qualitative inquiry based on the principles of grounded theory with 44 managers in 15 industrial firms located across ten US states. These managers included CEOs, pricing and marketing professionals, and financial professionals working in three industries (automotive, building products and chemicals).\nFindings: - The study's results reflect similarities and differences in the experiences of managers in industrial firms using all three pricing orientations. It reveals stark contrasts by pricing orientation with respect to how firms organize for pricing, manage the pricing process, make product pricing decisions, manage the transition to more advanced pricing orientations, and develop internal capabilities to face uncertain and ambiguous decisions. The findings also uncover contrasting price-setting processes by pricing orientation and the balanced used of scientific versus intuitive decision-making processes.\nPractical implications: - Pricing is often a neglected element of the industrial marketing mix. This study offers a variety of organizational practices by pricing orientation. The results highlight how best-in-class companies that adopted modern pricing practices to derive product prices are organized and how they reach pricing decisions.\nOriginality/value: - This study studies the commonly accepted pricing orientations and links them to organizational structure and decision-making theory. This study contributes to bridging pricing and organizational theories.\n",
"role": "assistant"
}
] |
Value co-creation and social media: A systematic literature review using citation and thematic analysis
|
[
"Social media",
"Innovation management",
"Thematic analysis",
"Citation analysis",
"Value co-creation",
"Co-creation",
"New product development",
"Co-innovation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The advancement of information and communication technology (ICT) and development of Web 2.0 have increased participation and sharing of information between customers as well as customers and firms. In this regard, social media plays a vital role in enhancing value co-creation from customers (Martini et al., 2012). Kaplan and Haenlein (2010) defined social media as \"a group of Internet-based applications that build on the ideological and technological foundations of Web 2.0, and that allow the creation and exchange of user-generated content\" (p. 61). Such new platforms for connectivity and interaction have encouraged participations of customers in co-creation (Prahalad and Ramaswamy, 2004). These \"co-creation technologies\" facilitates engagement of customers in value creation (Holdgaard and Klastrup, 2014). They have empowered customers, and made them more influential. Customers are not mere recipient of products and services; rather they have assumed pivotal role in new product development (Hoyer et al., 2010). These changes have forced firms to change their orientation from firm-centric view to customer-centric and involve customers through co-creation and joint product development.\n2. Research methodology: Citation is an acknowledgement of a published statement which is mentioned in the bibliography of a publication and it is vital to incorporate the most relevant publications for the progression of research. To select, analyze and interpret the patterns of citations in the literature of value co-creation and social media, this paper has used the bibliometric methodology which is also known as co-citation analysis. Citation analysis is \"quantitative-oriented bibliographic approach\" (Gundolf and Filser, 2013, p. 178), which is used to determine the most influential articles on the subject. Researchers have used this technique to map the structure and development of various research fields, such as organizational behavior (Culnan et al., 1990), strategic management (Nerur et al., 2008), and international marketing (Samiee and Chabowski, 2012). A frequently cited document gives the indication of the quality and influence of the article (Xi et al., 2015). This type of analysis is based on the idea that the intellectual links between publications can empirically be defined by the strength of co-citation (Garfield, 1979) which is in turn used as a measure of similarity of concepts and ideas between two publications. Therefore, such a scholarly work exhibits interconnectedness between researcher and scientific concepts (Gundolf and Filser, 2013; Linan and Fayolle, 2015). Citation analysis is based on the theory that citations are reliable indication of scientific interaction and it makes conceptual connections between scientific ideas while making them visible (Small, 1978). By doing so, it put forth interesting concept which are further elaborated by other researchers (Acedo and Casillas, 2005).\n3. Citation analysis: Since the seminal work of Prahalad and Ramaswamy (2004), there is significant expansion in the literature of value co-creation. The rise of service dominant logic and popularity of social media have augmented this growth. However, the literature eliciting the role of social media and virtual communities in boosting co-creation and joint product development is far from homogenous. This has encouraged the authors to perform a review of the literature for \"systematic and analytical clarification\" and developing a \"shared understanding\" of the different scholarly work on the core area (Saarijarvi et al., 2013). This systematic review begins with the citation analysis.\n4. Thematic analysis: After performing citation analysis, thematic analysis on the entire sample of 56 articles was carried out to identify major themes and sub-themes in the articles. \"Themes and sub-themes represent the core ideas, arguments and conceptual linking of expressions on which an article's research questions, constructs, concepts and/or measurements are based\" (Linan and Fayolle, 2015). The articles are grouped together on the basis of similarity in themes. Four major themes were identified in this analysis which contained different sub-themes. Figure 1 shows the result of thematic analysis.\n5. Conceptual model: Based on the themes identified in the aforementioned citation and thematic analysis, the authors have proposed a conceptual model (Figure 2). This model integrates customer's motivations to engage in co-creation along with firm specific resources to improve quality of value co-creation. Social media is particularly suitable for this joint product development as it \"creates a large virtual ecosystem via which companies and their customers can engage in a common and equal creative effort\" (Martini et al., 2014, p. 427). Social media allows firm to capture previously untapped resource of customer knowledge. Now firms are using skills and creativity of customer in boosting its innovative performance. Firms are providing tools to the customer become active part of idea generation and design stage (Piller et al., 2012). Therefore, social media and online communities have turned the economic relation between firms and customers in more interactive and participative social exchange.\n6. Conclusion: Value co-creation and social media has a natural alliance. With the rise of these concepts, firms are changing their ways to become more customers centric, rather firm centric. Value co-creation especially in the context of social media is also gaining popularity as a budding area of research, as it provides more access to data, and can enable researchers to study social settings in detail. However, the developing literature on this topic is far from homogenous and needed categorization for better comprehension of the previous work. This article aimed to address this issue by conducting a systematic review of the literature on co-creation in the context of social media using citation and thematic analysis. Articles were clustered into different groups based on the similarity of their core area. A number of converging themes and sub themes were identified in these articles. One major theme noted pertained to customer's characteristics, motivation, and experience to engage in co-creation. Equally important, but somewhat lesser research work, focused on firm's capabilities for knowledge sharing and processing. Based on these identified themes, a model with propositions is proposed that shows how characteristics of the customers and firms can complement and interact with one another to augment innovative output generated through co-innovation.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to deliver a summary of the influential work regarding value co-creation in the context of social media. Although, research on the role of social media in co-creation and new product development is growing field; the extant literature is still in developing stage, which needs systematization and categorization to comprehend its current stage and previous research.\nDesign/methodology/approach: For this purpose, existing literature on social media and co-creation was studied. Initially, a citation analysis was conducted of influential papers correlated with the topic to identify three streams of research. Later, thematic analysis was carried out to explore specific themes within these categories.\nFindings: Through citation analysis three research streams namely customer's co-creation on new product development, firm specific capabilities for knowledge sharing, absorption and processing and new opportunities were identified. Later, total four categories were identified through thematic analysis which contains different sub-themes such as test of theories, proposed theoretical frameworks, lead users characteristics, customer's motivation and experience regarding co-creation, online communities of customers and open innovation.\nResearch limitations/implications: This study also categorizes and systemizes the extant literature exploring role of social media in value co-creation. Such systematic review of the extant literature will help the academicians to understand the previous stream of work and pursue a particular line of enquiry in furthering the understanding of interaction between social media and co-innovation.\nPractical implications: This work is particularly useful for practitioners as more firms are moving the business online. These firms are actively using social media and user-generated content to gain insight into customer's preferences. By increasing the participation of customers and lead users through online communities, firms can also increase customer's commitment.\nOriginality/value: The paper adds to the extant literature by identifying research streams and themes in the extant literature on the role of social media in value co-creation. Later, these themes are abductively linked to develop a theoretical framework.\n",
"role": "assistant"
}
] |
Short food supply chain between micro/small farms and restaurants: An exploratory study in the Marche region
|
[
"Restaurants",
"Local food system",
"Logistics service",
"Supply chain"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Over the last few years, there has been an increase in academic studies about the Short Food Supply Chains (SFSCs), and how they attempt to reconfigure relationships between producers and consumers in space and time (Marsden et al., 2000; Renting et al., 2003; Sage, 2003). These initiatives, in contrast to the trend towards standardisation and globalisation that characterises the dominant agri-food system, include the \"re-localisation\" of food production and consumption (Hinrichs, 2000; Hinrichs, 2003; Winter, 2003; Goodman, 2003). The innovative aspect of the SFSC lies in its ability to \"re-socialise\" food, through relationships of \"proximity\" and \"trust\" that are created between producers, consumers and food (Marsden et al., 2000; Renting et al., 2003; Hallett, 2012). In SFSC, various sales models (on-farm sales, farm shops, farmers' markets, pick-your-own and community supported agriculture) are based on the direct contact between producers and consumers and the short distance between them. SFSC initiatives have the potential to move the food system towards a sustainable direction (Torjusen et al., 2008). Wubben et al. (2013) recognise three different business model of SFSC; a SFSC can be driven by initiator-stakeholder that aim at increasing the economic viability of their business, supporting producers or promoting producer-consumer connection. The aim of this work is to estimate the benefits, the obstacles and the interests associated with the implementation of a logistics service that supports the SFSC for local restaurants. The remaining of the paper is organised as follows. Section 2 provides the study background. Section 3 describes the structure and the operational mode of the proposed local food service. Sections 4 and 5 illustrate the research questions and the research method, respectively. Section 6 summarises the survey outcomes, subsequently discussed in Section 7. Section 8 proposes the conclusion of this study including limitation and future research areas.\n2. Literature review: local food and restaurants: In recent years, literature on the SFSC has expanded towards the restaurant industry. Studies reveal an emerging interest in consumers' perception, opinions, behaviours and habits around the inclusion of local food in restaurants' menus (Alfnes and Sharma, 2010; Schubert et al., 2010; Alonso et al., 2013; Campbell and DiPietro, 2014; Lillywhite and Simonsen, 2014; Contini et al., 2017). Local food consumption receives positive judgments in conferring \"values\", above all towards the intrinsic characteristics of products such as quality, freshness and taste, especially for restaurants (Gregoire and Strohbehn, 2002; Inwood et al., 2009; Kang and Rajagopal, 2014). Other benefits are the knowledge of the food source (Gregoire and Strohbehn, 2002; Strohbehn and Gregoire, 2003), the environmental benefits for food travelling shorter distances (Duram and Cawley, 2012), and support for local farmers (Reynolds-Allie and Fields, 2012; Dougherty et al., 2013). Local growers, who sold to restaurants, had the benefit of having a constant revenue stream at a higher price than that of intermediaries (Strohbehn and Gregoire, 2003; Sharma et al., 2012; Alonso, 2010). The high cost of local food to producers does not affect the purchase of local food by restaurants (Schmit and Hadcock, 2012; Sharma et al., 2014); the food quality is more decisive than price in restaurateurs purchasing choices (Starr et al., 2003; Inwood et al., 2009). Although these findings are in contrast with other studies (Green and Dougherty, 2008, Peterson et al., 2010; Dougherty et al., 2013), which highlight issues for local producers due to the low price required by the restaurants. However, direct sales by farmers to restaurateurs may have positive results for the economic viability of the local food network (Sharma et al., 2012), and the use of local food in restaurants, tourist accommodations and food stands can be a promotional vehicle for the territory (Inwood et al., 2009; Blichfeldt and Halkier, 2014).\n3. Local food service structure and operational mode: The literature analysis points out the weaknesses connected to the distribution of local food to restaurants. The main barriers for purchasing local food are: the limited product availability; the lack in information flow between farmers and restaurateurs; and the logistics barriers. Such criticalities can be faced by introducing a coordinated logistics system, a local food service aimed at promoting and enabling local food consumption among restaurant, without undermine the short food supply chain principles. The creation of a local food service designed according to the characteristics of a cloud computing system has been considered. Thus, the actors will directly interact within the system without the interference of additional intermediaries. The customers will have the opportunity to choose the farmers from which to buy, and they will directly interact with them. An online platform will display the farmers' identification data, the types and the quantities of products available for each producer and the products prices defined by the producers themselves. Therefore, the consumer will only have to choose the farm from which to make the purchase and specify the quantities of desired products. The local food service staff, basing on the orders loaded on the online platform, will manage the products collection and distribution following one of the three distribution models shown in Figure 1.\n4. Research questions: The research objective is to identify the potential barriers and enablers for the logistics service implementation. The research questions are:\n5. Research method: The analysis is structured in five main steps:\n6. Results: Producers\n7. Discussion: Producers and restaurateurs' opinions (Table I), connected to the implementation of a logistics service, have been investigated: the attitude assessment has referred to a five-point Likert scale (1-strongly disagree, 3-agree, 5-strongly agree).\n8. Conclusion: Recent literature shows an increasing interest in the consumption of local food. Restaurant's chefs identify in the local food a strategic option to increase their business. However, in most of the cases, communication and logistics barriers complicate the local food purchasing. Our study analyses the introduction of a logistics system facilitating and enabling the use of local food among restaurants. The results indicate the absence of strong barriers in the logistics service implementation. The main users seem to appreciate the services proposed by the new system and to agree to the connected organisational change required, such as the use of IT. Three main groups should be interested in the results of this paper and its future development: producers such as private managers and local politician. Following the successful examples of cooperation (Van der Ploeg et al., 2012; Le Velly and Dufeu, 2016), producers can self-organise and manage the local food logistics service to improve and expand their sales market. The better communication and distribution channel structure would enable micro and small farms to achieve restaurants as clients. Also, due to stakeholder's interest and absence of evident barriers, private managers focussed on the fair economy, could have interest in the implementation of the local logistics service. Finally, for the local politicians, the logistics service would be a way to promote the sustainable development of the territory, valorise the territorial asset enrich the touristic attractiveness of the area and sustain the values of the social economy. Policymakers would give an important and concrete support to the project implementation by organising events and provide informative material to increase the knowledge diffusion on the logistic food service among restaurateurs, producers and community; detecting the more suitable initiator-stakeholders; and exploring the possibility to use EU, national or local fund and financial tools for the project launch.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore the strength and weakness connected to the implementation of a local food logistics services, designed to facilitate and enable the use of local food among restaurants.\nDesign/methodology/approach: A total of 60 restaurateurs located in a small Italian city and 100 owners of micro and small-sized farms, located within a 30 km radius from the city, have been interviewed. The collected data have been analysed through descriptive analysis and Kruskal-Wallis and Mann-Whitney U tests for differences between gender, production type and business size.\nFindings: The results agree with existing literature: poor communication and an unstructured logistics limit the interaction among actors. However, significant for the purpose, the data show that producers are interested in expanding their market and restaurateurs are interested in a broader supply of local products, and they both open to a logistics service that improves the supply of local food.\nResearch limitations/implications: The following exploratory study is based on a sample of farmers and restaurants in a specific area so the results could not be generalised in a national/international context.\nPractical implications: The logistics service objective is the promotion of a sustainable territory development in support of the social economy values.\nOriginality/value: The study investigates this issue focussing on a practical solution of local food distribution system among restaurants. The logistics service tends to solve the logistics criticalities maintaining the principles of the short food supply chain.\n",
"role": "assistant"
}
] |
Constructing the Norm-critical awareness scale: A scale for use in educational contexts promoting awareness of prejudice, discrimination, and marginalisation
|
[
"Equal opportunities",
"Health services",
"Higher education",
"Instrument development",
"Norm criticism",
"Social norms"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Despite the uncontroversial ideal of \"doing good\" that is characteristic of helping and caring professions, inequities in health and care persist, echoing wider structural relationships in society. Due to professionals' self-image of being persons that do not oppress, and that treat everybody fairly, these issues can be difficult to address (see, e.g. Schroeder and DiAngelo, 2010). It is therefore important to promote awareness of healthcare inequity early, among students, in order to improve the care of marginalised groups (Holmstrom et al., 2017). The present study was carried out within bachelor and postgraduate programmes in nursing at a university college in Sweden, with the aim of developing a valid and functional scale for measuring norm-critical awareness. This scale would be useful in studying the level of norm-critical awareness in groups in which interaction and communication are professionally essential for outcomes related to equal opportunities, inclusion and diversity. The scale could identify knowledge gaps and normative beliefs in groups and individuals, as well as evaluate norm-critical educational interventions. However, no such scale exists today.\n2. Background: 2.1 Norm-critical perspectives\n3. Methodology: This study was carried out in the context of a nursing education department at a Swedish university college. The University college offers three-year bachelor programmes in nursing and social psychiatric nursing, as well as one-year programmes in specialist nursing. All Swedish education for healthcare professions is guided by the National Higher Education Ordinance, and this is applicable to these programmes. Within these guidelines, universities and colleges are free to form their curricula. At the university college of this study, an effort was made to develop and to systematically implement norm-critical content in nursing education, something that has not been done previously in Sweden.\n4. Results: The final sample of 209 students had an age range of 19-60 years, with a mean age of 27 years. The majority (87 per cent) of the students were born in Sweden. The gender distribution in the total sample was 88 per cent female students and 12 per cent male students.\n5. Discussion: The results indicate that in our sample of nursing students, the theoretically derived dimensions of function, consequences, identity, resistance and learning measure the underlying construct of norm-critical awareness and that a summary NCAS score does indeed reflect this construct. The statistical analysis indicates that being aware of the function of norms is equivalent to general norm-critical awareness.\n6. Conclusion: This study indicates that the summary score of the NCAS reflects general norm-critical awareness among nursing students, and thus it can reveal attitudes that are linked to prejudice, discrimination and marginalisation. The present analysis supports a 20-item scale measuring norm-critical awareness, building on the five theoretically derived dimensions of function, consequences, identity, resistance and learning. These dimensions can be used as subscales in empirical investigations using the NCAS, but also as a summary score.\n",
"role": "user"
},
{
"content": "Purpose: Healthcare professionals' conscious or unconscious norms, values and attitudes have been identified as partial explanations of healthcare inequity. Norm criticism is an approach that questions what is generally accepted as \"normal\" in society, and it enables professionals to identify norms that might cause prejudice, discrimination and marginalisation. In order to assess norm-critical awareness, a measurement scale is needed. The purpose of this paper is to develop a scale for measuring norm-critical awareness.\nDesign/methodology/approach: The scale-development process comprised a qualitative item-generating phase and a statistical reduction phase. The item pool was generated from key literature on norm criticism and was revised according to an expert panel, pilot studies and one \"think aloud\" session. To investigate the dimensionality and to reduce the number of items of the scale, confirmatory factor analysis was performed.\nFindings: The item-generation phase resulted in a 46-item scale comprising five theoretically derived dimensions revolving around function, consequences, identity, resistance and learning related to norms. The item-reduction phase resulted in an instrument consisting of five dimensions and 20 items. The analyses indicated that a summary score on the scale could be used to reflect the broad dimension of norm-critical awareness.\nOriginality/value: The Norm-critical awareness scale comprises five theoretically derived dimensions and can be used as a summary score to indicate the level of norm-critical awareness in educational contexts. This knowledge is valuable for identifying areas in greater need of attention.\n",
"role": "assistant"
}
] |
Documentation in a community informatics project: The creation and sharing of information by women in Bangladesh
|
[
"Bangladesh",
"Analogue back-up",
"Information continuum model",
"Personal documentation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper reports on the creation and sharing of information in the context of a community informatics project, participatory research and ownership with technology, information and change (PROTIC), which has been working with village women in remote parts of Bangladesh since 2015. It focusses on the production and sharing of information by the women themselves and is based on data collected by the authors during a workshop held in Dhaka in March 2019 and subsequent focus group discussions in a Bangladeshi village, as well as on previous observations of documentation practices in the PROTIC villages by two of the authors.\nThe PROTIC project: Bangladesh is the eighth most populated country in the world with more than 160m people in 147,570 square kilometres. More than 65 per cent of the population live in the rural areas (World Bank, 2018). Socially and culturally Bangladeshi women are marginalised and disadvantaged because of their lower access to economic resources and limited participation in the decision-making process (Guhathakurta and Banu, 2017). Agriculture is the main source of livelihood for the rural population and women play a key role in agriculture, but their efforts and hard work tend to be unacknowledged and underappreciated due to the dominant social perception that sees women as housewives and does not recognise their active involvement in agricultural production (Kabeer, 1991; Rahman, 2000, 2010). Traditionally in Bangladeshi society, women are associated with the private sphere of household activities rather than with the formal labour market (White, 1992). The cultural importance accorded to female honour and seclusion (Kabeer, 1991) impacts on the activities that rural Bangladeshi women can engage in and on their daily life. However, women manage more than 50 per cent of the agricultural activities (FAO, 2011, p. 9). They are involved in each and every step of agricultural production, but particularly in the post-harvest processing which mainly happens in the home areas or courtyards (Rahman, 2010; Lewis, 2011).\nLiterature review: While abundant research has been done on information and documentation practices in academic and workplace contexts, comparatively little research has considered documentation practices in everyday life. Studies that have looked at how households process and manage information have focussed on the adoption of technology and its place in household information practices rather than on the use and management of information per se (Kalms, 2008). Moreover, research on personal information management is usually concerned with the management, use and sharing of information and often does not onsider document creation (McKenzie et al., 2014). A few studies have looked at information and documentation practices in the context of leisure activities, such as Hartel's (2006, 2010) study of the information and documentation practices of gourmet cooks and McKenzie and Davies' study of wedding planning calendars (McKenzie and Davies, 2010), while McKenzie and Davies (2012) and McKenzie et al. (2014) have studied how households keep track in their everyday life through planners and calendars. In many cultures, women are viewed as the guardians of the family memory (Ketelaar, 2005), \"the customary nurturers and keepers of family tradition\" (Davison, 2000, p. 81). Tucker and Bogadottir's (2008) research in Iceland showed that women were traditionally perceived as being in charge of keeping the family's records, mementos and memorabilia because men were occupied outside of the house and women were in charge of all aspects of the home. However, no previous research has been found that studied the information and documentation practices of rural women in developing countries.\nMethodology: Data for this paper were collected as part of the evaluation of the information literacy workshop held in Dhaka on 27 March 2019 and from subsequent focus group discussions organised in Borokupot by one of the authors. During the last session of the information literacy workshop, the participants were asked to comment on the activities that they had engaged in and to answer specific questions about what information they would like to keep from the workshop, in what format, and with whom they would be sharing it. They were asked to choose from a list of proposed formats (their own notebooks, printed sheets with mainly text, printed materials with lots of images, video recording, sound recording, text messages, Facebook posts, information in an app, or brown paper) which one would be their preferred medium to keep information from the workshop and which one was their favourite format in general to keep information that they would need again later. The participants were also asked to comment on what they had written during the workshop in the notebooks that had been handed to each one of them at the beginning of the workshop and on their use of notebooks in general to keep information. In the same way as for the other sessions of the workshop, the discussions were facilitated by four Bengali speakers and audio recorded. In addition, the facilitators were provided with an evaluation sheet in which they recorded some comments made by the participants. The discussions were held with the groups to reflect the collectivist context of the Bangladeshi research setting. During the discussions, the women often concurred with one another. Therefore, in the analysis of the findings, it was considered more appropriate to attribute their sayings to the groups rather than to individuals.\nContinuum analysis of the documentation practices of PROTIC participants: The main objective of the PROTIC project is to empower Bangladeshi women working in agriculture through improving their access to information. Information is provided to them as part of the project in different forms, verbally in training sessions and by the call centre, through printed leaflets, via SMS and in the mobile phone apps that were developed for the project. This section focusses on the documentation created by the women participants themselves and uses the four dimensions of the information continuum model to analyse how this documentation is created, captured, used and shared by the PROTIC participants.\nDiscussion: The findings from the information literacy workshop held in Dhaka on 27 March 2019 and the subsequent focus group discussions in Borokupot show that the Bangladeshi women participants in the PROTIC project wanted more information than what they were currently receiving and that they wanted to keep the information they found useful. They have taken the initiative to create themselves some documents for their own use and to copy some information that was provided to them into other formats that they thought would be more durable and would make it possible for them to access the information in the future. In so doing, they initiated documentation practices that were new to them and to their communities. Using notebooks for recording information or brown paper sheets for recording and sharing information are not traditional practices in Bangladeshi villages. Before they became participants in the PROTIC project, the village women had few opportunities to use writing. They had learnt to write when they went to school, but most of them left school early and did not need to use writing in their everyday life since then. Through their engagement in the project, they have been rediscovering writing and enjoying it. As one of them said: \"After leaving school, I did not write like this for a long time because I did not need to. I thought I forgot how to write! But now, writing in the notebook, I feel like I am rediscovering my writing skills\".\nConclusions: Our research has shown that the Bangladeshi women who participated in our workshop and focus groups want more information that what was currently provided to them, and that some information that was provided to them was not in a format that would guarantee that they would be able to access it in the medium to long term. For this reason, they started creating documents themselves to fill some of the gaps or to capture the information in formats that were more suitable for longer-term preservation in their village context. The women themselves are taking a long-term perspective, thinking of future uses for the information they are receiving and anticipating that digital records may no longer be accessible to them in the future; and for this reason, they have been creating analogue back-ups of information provided to them in digital form.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to discuss the creation and sharing of information by Bangladeshi women participants in a community informatics project and to assess to what extent the information provided to them meets their short and longer-term needs.\nDesign/methodology/approach: The analysis is based on data collected during a workshop with village women in Dhaka and focus group discussions in rural Bangladesh in March and April 2019. The information continuum model is used as a framework to analyse the data.\nFindings: The study shows that the women document their learning and share it with their families and communities and that they are very conscious of the importance of keeping analogue back-ups of the information provided to them in digital format. They use notebooks to write down information that they find useful and they copy information provided to them on brown paper sheets hung in the village community houses.\nPractical implications: This paper raises questions about how information is communicated to village women, organised and integrated in a community informatics project, and more generally about the suitability and sustainability of providing information in digital formats in a developing country.\nOriginality/value: The paper shows how village women participants in a community informatics project in Bangladesh took the initiative to create and preserve the information that was useful to them in analogue formats to remedy the limitations of the digital formats and to keep the information accessible in the longer term.\n",
"role": "assistant"
}
] |
Psychological contracts and performance management in Mexico
|
[
"Psychological contracts",
"Performance appraisal",
"Performance management",
"Mexico"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Global forces have placed performance at the heart of the employment relationship. Simultaneously, performance management has become, at least implicitly, human resources management (HRM) primary goal for strategic purposes. Much research on performance appraisal has focused on structural factors within organizations or on the cognitive side of the superior-subordinate appraisal relationship (Levy and Williams, 2004). Surprisingly, performance appraisal has received little attention from studies of psychological contracts, though it is a key practice in which promises and rewards related to work performance arise (Rousseau and Greller, 1994).The psychological contract is generally defined as an individual's beliefs regarding conditions of a tacit exchange agreement - promises - between him/herself and the organization. Individuals develop this belief system over time based on various factors such as an employing organization's practices, daily interactions with managers, colleagues or co-workers (Rousseau and Greller, 1994), and personal ideals about work. Thus, psychological contracts are subjective in nature yet also culturally determined (Rousseau and Schalk, 2000). So are perceptions of performance appraisal (DeVoe and Iyengar, 2004). We aim to bring these two research streams together using psychological contracts theory to explore performance management practices in Mexico paying special attention to the role of national culture and other social structural and economic factors.Our study examines the content of the psychological contract from the viewpoint of both sides of the employment relationship - managers and employees - in the context of performance appraisals. By psychological contract content we mean the beliefs system held by an individual regarding performance, its evaluation, and the expected rewards and recognition that follow.\nPsychological contracts: Psychological contracts help understand the success and failure of different HR practices and the consistency between individuals' beliefs and organizations' expectations of employee behavior. HR managers, however, rarely frame HR practices in terms of psychological contracts, i.e. managing promises or commitments from both parties - individuals and organizations. Managing all the subjective perceptions and attitudes involved in personnel practices requires sensitivity toward individual psychodynamics and their interplay within organizations (Rousseau and Tijoriwala, 1998). If for some reason behaviors on either management or employees' side violate psychological contracts, organizational awareness does not arise until someone in the relationship perceives that the contract has been broken (Rousseau and Greller, 1994).HR practices often serve as sources of promises such that individuals believe there is a return to expected behavior. In other words, HR practices function as psychological \"contract makers.\" Moreover, each time a new HR practice is introduced, the conditions or terms of the psychological contract change. For example, when organizations introduce a new performance appraisal system, individuals might expect performance to be measured according to standards different from prior agreements and in exchange for different rewards (Rousseau and Greller, 1994). Such changes could challenge existing psychological contracts. What happens if expected behaviors are not explicitly communicated and the new benefits are not consistent with individuals' work beliefs? Employees could face uncertainty and either learn the tacit conditions of the new contract by observing others' behaviors, including those of managers, or turn to other contract makers - such as a different HR practice - in order to make sense of change. Therefore, communication processes such as encoding (through individual's predispositions) and decoding (influenced by social cues) play an important role in the interpretation of performance expectations because of the impact on psychological contracts.Understanding the complexity of these socio-psychological dynamics is important for the management and resulting effectiveness of psychological contracts.The role of HR and front-line managers\nPsychological contracts in Mexico: International research has questioned whether the psychological contract is an appropriate concept to understand employment relationships. Because each society develops its own bases for social relations and contracts, models developed in the USA might not apply to all contexts. In some societies, employment relationships are governed by rules (e.g. France) or by a culture of compromise (e.g. Belgium). Besides, firms and individuals, the social context plays a major role in shaping employment relationships. Therefore, the concept of contract may not apply for understanding and managing employment relationships (Cadin, 2000; Sels et al., 2000).In some countries, the state and other institutional arrangements such as norms governing economic activity (North, 1990) replace the role of legal contract for employment and individuals do not feel the need for engaging in conditional or promised terms. In other words, the social context governs the employment relationship producing in the employee the same commitment to performance that contracts do in other societies. Although most societies have institutional norms such as laws governing employment relationships, they vary in the application of such norms due to cultural and institutional differences (Rousseau and Schalk, 2000). Psychological contractual relationships then become relevant under certain circumstances such as where labor laws do not provide structural conditions for equality or fairness; where there exists mistrust on institutional agencies that govern employment relationships; or generally, where individuals feel the need to rearrange their personal status within a firm through psychological contracts.We have argued elsewhere that work in Latin American countries such as Mexico is paramount. Beyond providing a means of sustenance, work plays a central social and emotional role in individuals' life. This creates an intricate implicit social contract between workers and employers at the firm level (Elvira and Davila, 2005a) that in turn affects individuals' psychological contracts (Elvira and Davila, 2005b).These phenomena relate to the historical development of employment relationships. Mexican society has not followed economic rationality as a development principle but a different rationale: that of solidarity and a sense of community (Ibarra, 2006). Based on this principle we expand prior work on psychological contracts in Mexico (Diaz-Saenz and Witherspoon, 2000) and outline contextual dimensions (cultural, structural and economic) relevant for psychological contracts in Mexican organizations. Then, we link our discussion of contextual factors to performance evaluation.Cultural factors\nMethod: To explore performance management practices in Mexico through the lens of psychological contracts we conducted focus groups interviews at two organizational levels in three different organizations. The two organizational levels included front-line managers vs supervisory and clerical staff. On average, seven individuals participated in each focus group. Only four women were part of the sample. Organizations belong to the manufacturing sector, have been operating in the country for more than 20 years, and employ 2,000 employees on average. Regarding ownership, we studied a Mexican multinational and two Mexican firms recently acquired by a Latin American multinational. We selected organizations based on the following criteria: first, organizations had to be operating in Mexico for a long time in order to have mature HR practices. Second, organizations had to be considered successful in terms of their strategy and market orientation. Finally, the HR executives had to be committed to the objective and methodology of the research.The groups' interviews were conducted on location in a meeting room for each organization. As recommended by Rousseau and Tijoriwala (1998), we developed semi-structured interviews that opened spaces for deeper discussion with an eye to uncover silent interpretations potentially present in the collective mind of organizational members. We used several recording strategies mainly due to limited resources. An interview was tape recorded, another was typed verbatim by a research assistant, and the third was conducted by company staff. Sessions lasted an average of 60 minutes.We analyzed the data by first identifying the terms or conditions individuals perceive to owe and the expected return from their employer in the context of performance appraisal. Then, we clustered single terms into categories arranged by their semantic relationships, in order to develop a broader interpretation of the psychological contracts around performance appraisal (Rousseau and Tijoriwala, 1998). We present below the main themes that emerged from this analysis.\nFindings: We present the study's findings around three key themes:1. performance appraisal systems;2. fulfilled and unfulfilled promises; and3. sources of terms and conditions as the content of psychological contracts for managing and participating in performance evaluation systems.For each theme we compare and contrast the stated viewpoints of managers and employees. Table I presents examples of verbatim statements of managers and employees that illustrate the content of psychological contracts linked to the contextual factors of Mexico in performance appraisals systems.Performance appraisal systems\nDiscussion and conclusions: managing promises in performance evaluation in Mexican organizations: The findings just described can be interpreted in light of cultural factors and the structural and economic conditions in Mexico discussed earlier in this paper. It seems that paternalistic leadership might not suffice to motivate employees who trade good performance for benefits or other perquisites (Martinez, 2005) in the context of HR practices that are inflexibly transferred from other international contexts. For example, performance appraisals that use ranking systems undermine the socio-emotional aspect of the superior - subordinate relationship, diminishing the managers' control over the process. The content of the paternalistic leadership style needed for job security and harmonious organizational climate changed under imported performance appraisal systems. Consequently, new behaviors emerged in the analysis. First, managers could only exhibit a protectionist behavior in limited cases such as the first time the employee is ranked as a low performer. After a second time, he/she has to punish the worker with no merit increase and ultimately dismiss him/her.An important finding is that feedback interviews are perceived as the social space where managers and subordinates arrange exchanges and commitments by developing emotional engagements (Rousseau and Greller, 1994). Giving feedback seems to be the most deficient part of performance appraisals in these three companies. Suppressing this social space could reduce the value of the performance appraisal system and force managers and workers to use alternative social structures that sustain good relationships. Examples of this could be the refusal to evaluate subordinates and a supervisor's bypassing his own merit increase. Thus, vertical social networks seem important alternative structures (Elvira and Davila, 2005a) through which the process of performance appraisal operates. Interestingly, if somewhat puzzling, HR executives indicated that the feedback interview is unwelcome by managers because of the subjectivity involved (Davila and Elvira, n.d.), whereas the managers interviewed rejected the new performance system because it lacked a feedback interview. This contrast in perceptions illustrates the intrinsic complexity of psychological contracts in Mexico. Managers spend the year building good relationships with their subordinates that facilitate the exchange of benefits or perquisites for job improvements (Martinez, 2005). All that effort could be lost if the feedback interview is mismanaged.The uncertainty around the appraisal's outcome jeopardizes organizational resources that managers and subordinates could use to trade for performance and improve the employment relationship. Consistent with prior studies of motivation and performance appraisals (DeVoe and Iyengar, 2004), employees in our study asked for intrinsic motivators such as employee recognition, promotional rewards, and developmental and career opportunities. However, the performance appraisal system used at these companies does not include this type of recognition; it is perceived as tied only to financial rewards.Employees' perceptions of uncertainty could also derive from the changes in appraisal systems and the use of various instruments. Consistent with the literature (Rousseau and Greller, 1994) HR and managers simply do not have the time for managing changes in standards while simultaneously helping employees understand them.Organizations change performance appraisal systems for many reasons. Perhaps, in the long run, cultural traits of the employment relationship tend to deteriorate the effectiveness of appraisal systems. This happens, for example, when managers adjust employees' salary increases according to their own criteria or when employees perceive that a good appraisal depends on the personal relationship with the manager. Here, again, vertical social networks lubricate the rigid performance appraisal system. HR could consider reliance on subjective perceptions as a reason to improve or change evaluation systems. Nonetheless, managers and employees appear unaware of when or why a new system will be implemented, and rarely receive proper training. Therefore, HR could make a contribution by creating a climate where managers and employees develop the terms and conditions of a psychological contract.Managers expect to protect their employees from the lack of rewards of a poor performance evaluation and employees expect their managers to advise them of what it takes to obtain a good evaluation. This happens even in the context of modern performance appraisals such as top ranking differential systems, and provides evidence for the hybrid management approach that characterizes structures in Mexican organizations (Davila and Elvira, 2005).The paternalistic leadership style that generates dependency behaviors in subordinates (Martinez, 2005) could work against some assumptions behind performance appraisals systems. Workers need independence from their superiors and substantive decision-making discretion before committing to their own work objectives and skill development. Performance appraisals systems are designed under the assumption that organizational policies allow individuals to be responsible for their own performance, and therefore, might call for different styles of leadership.Subjectivity is another key factor in shaping the terms and conditions of the psychological contract through performance evaluation. Subjective assessments shape perceptions of performance standards and of rules for attaining future rewards. In this study, both managers and employees rely of their subjective views as a means to understand not only the process but also the outcome of the appraisal. Managers, for example, use their own criteria and personal experience for setting standards and rewards, often aimed to be fair with employees. Employees subjectively analyzed what it takes to obtain a good evaluation: a good relationship with the superior.Although unacknowledged in most performance management studies, the macro-economic instability of the country affects firms' purpose in performance appraisals. Salary increases in the companies we studied depend not only on the performance evaluation but also on the country's financial situation. Employees also see that their salary increase is not a result of their performance but an adjustment to the country's inflation rate. Economic fluctuations impeded one firm to offer differentials in salary increases and merit increases to both managers and employees. As a result managers and employees devalued of the role of performance appraisals in the employment relationship. Still, it seems that managers and employees are aware of the strict norms for dismissal enforced by Mexico's labor laws because it was hardly mentioned that a worker could be dismissed by low performance. Turnover appeared more the result of the lack of intrinsic motivators than of a poor performance appraisal.In conclusion, in this research we used the psychological contract concept to understand the management of performance appraisals in three Mexican organizations. We found that contextual factors such as cultural, structural and economic reasons influence the socio-psychodynamics of the manager - subordinate relationship and affect perceptions of performance evaluation. We also found that HR's role is an important variable to consider in the implementation of the appraisal system.Performance appraisals in Mexico could be better managed when considering the perceptions underlying psychological contracts. These perceptions are influenced by the social contract in which the employment relationship of this country is embedded. In particular, in Mexico individuals understand the intrinsic socio-emotional behaviors that accompany performance evaluation and engage in the process at different levels of commitment. Both parties - managers and employees - respond according to their needs and what they perceive they receive from the company. In terms of structural factors, individuals face the challenge of managing imported rigid bureaucracies that contradict cultural values. Finally, perceptions of the performance appraisal's purpose are altered because the outcome is often more strongly linked to the country's economic situation than that of the firm.\n",
"role": "user"
},
{
"content": "Purpose: - The aim of this paper is to use psychological contracts theory to explore performance management practices in Mexico paying particular attention to the impact of national culture, social and structural factors.\nDesign/methodology/approach: - The paper examines the content of psychological contracts from the viewpoint of both parties in the employment relationship - managers and employees - in the context of performance appraisals. It conducted focus groups and interviews at these two organizational levels in three different organizations.\nFindings: - Findings are organized around three themes: performance appraisal systems, the promises (fulfilled and unfulfilled) and the sources of terms and conditions as the content of psychological contracts for managing and participating in performance evaluation systems. For each theme the paper presents and contrasts the viewpoints of managers and employees.\nPractical implications: - Because of the sensitive cultural nature of performance appraisals, the findings help managers implement this process by explaining the underlying psychological contracts in Mexico. Specifically, employees experience the socio-emotional behaviors that accompany performance evaluation and engage in the process at different levels of commitment. Both managers and employees respond according to their needs and what they perceive to be the company's reward.\nOriginality/value: - Performance appraisal has received little attention from studies of psychological contracts, though it is a key practice in which work promises and rewards related to performance are made. In this study, the paper brings these two research streams together and apply it to a culturally unique setting.\n",
"role": "assistant"
}
] |
The diversity of activation markets in Europe
|
[
"Activation",
"Governance",
"Marketization",
"Quasi‐markets",
"Europe",
"Unemployment"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Undoubtedly, the introduction of markets for the provision of social services is the most salient and contested form of new governance that European welfare states have been confronted with. In the provision of activation services the introduction of market mechanisms has been booming in Europe during the last decades as well. At the same time, several authors point out that differences in the design of activation and, more generally, social services markets are significant (Struyven and Steurs, 2005; Bredgaard and Larsen, 2008; Nyssens, 2010; Bode et al., 2011).We think that exploring the issue of market diversity in more depth is interesting for various reasons. First, it may be interesting in itself in that it helps to avoid rather rigid distinctions between various service provision models by showing that concrete service provision models often do not neatly fit into provision model typologies and that boundaries between various models are blurry rather than clear. Second, exploring diversity is a first step in developing a more elaborate research agenda, involving issues such as:* Explaining diversity: does diversity mainly reflect different stages in the development of service markets, or is a more complex explanatory framework required? For example, does diversity reflect diversity in activation policies? Is it related to the public values that marketised service provision should guarantee in various national contexts?* What are the consequences of market diversity for the effectiveness and outcomes of activation? From the literature we know (see the final article in this issue) that the results of marketised activation are rather mixed, and certainly not unambiguously positive. Do some market models produce different types of results and effects than others? What is the impact of various market models on the quality and accessibility of services? Are political choices regarding the design of markets explicitly intended to pursue specific policy objectives?This article aims to be a first step in developing a research agenda for marketised activation by elaborating upon the issue of market diversity, based on our study of governance reforms in the policy areas of activation and income protection in nine European countries (van Berkel et al., 2011). The article intends to identify dimensions of diversity in the design of national activation markets. Based on the concept of quasi-markets defined by Le Grand (1991) we will assess the development towards market governance in the provision of activation in the nine countries under investigation. The first part of this article will outline the theoretical framework of market governance and the concept of quasi-markets. The second part will discuss the diversity within European activation markets along the identified dimensions. Finally, we summarise the empirical evidence.\nMarket governance and the introduction of quasi-markets: In the literature market governance, the rational goal model of governance or New Public Management governance is often presented as a specific type of governance or governance regime with distinct characteristics compared to other types of governance: hierarchical or procedural governance or \"the old public administration\" on the one hand, and network or open system governance or \"the new public service\" on the other, as discussed by Denhardt and Denhardt (2000), Newman (2001), Considine (2001) and Struyven et al. (2002). These authors characterise types of governance in various ways, but they all make clear that they differ considerably in terms of their theoretical foundations, underlying models of human behaviour, the public values they emphasise, the role of government, etc. Considine (2001) has summarised the main characteristics of the various types of governance as follows[1] (Table I).Whilst typologies such as the one developed by Considine are useful instruments in analysing how models for the provision of social services are developing, two features of governance reforms in general and the introduction of service markets specifically should be underlined. First - as was mentioned in the introduction - rather than replacing one dominant type of governance with another, these reforms may result in governance hybridisation. Several authors point out that the introduction of markets has resulted in mixed delivery models rather than pure market delivery (Warner and Hefetz, 2008). Svensson et al. (2008, p. 514) develop a similar argument: \"[...] the issue at stake is not how to select the best regime but how to manage a fruitful coexistence given the prevailing aims of public policy\". Second, market design changes over time, partly, as has been emphasised among others by Finn (2009), because the development and management of welfare markets involves a sharp learning curve. Hefetz and Warner (2004) show that marketisation is not a one-way street, and that public managers not only contract out but may also contract back-in. According to these authors, these governance dynamics develop while public managers \"balance technical and political concerns to secure public value\" (Hefetz and Warner, 2004, p. 171). This underlines the point made in the introduction that diversity may at least partially be related to differences in the stage of development of service marketisation.To assess the development of the market governance model we will rely on a second concept: quasi-markets. Le Grand (1991) was one of the first to elaborate the concept of quasi-markets in the context of social policy, showing how markets for providing social services are similar to and differ from conventional markets. Quasi-markets are markets because monopolistic state providers are replaced with independent ones competing with each other for customers of purchasers. Ideal-typically, marketisation involves a clear split between purchasers and providers of services (Struyven et al., 2002). In the context of activation services, the monopolistic state providers used to be benefit agencies, local welfare agencies or Public Employment Services (PES) agencies, i.e. the core agencies traditionally involved in the administration of income protection systems for the unemployed and the provision of the predecessors of activation policies: labour-market policies or employment services. At the same time, the traditional state monopoly in social service provision should not be taken too strict: in several countries, actors such as third sector organisations and NGO's have a long tradition in providing social services and this involves services targeted at groups that now belong to the target groups of activation. The qualification quasi refers to a number of characteristics of markets for social services that makes them different from conventional markets. First of all, the providers are not necessarily private, so that marketisation is not synonymous with privatisation (Struyven et al., 2002). In addition, profit maximisation is not necessarily the main objective of these providers as is the case in conventional markets. Second, the direct consumer of the service is not always involved in purchasing decisions, which are often made by the core public agencies mentioned before that, in the context of service marketisation, assume the role of service purchaser. Third, the state remains involved in the financing of the services through budgets or vouchers for social services (Nyssens, 2010).Following the concept of quasi-markets we can identify different potential dimensions of diversity in the implementation of market forces in the provision of activation services. On the one hand we have to consider the three group of actors involved: the providers, the purchasers and the customers/users. Regarding the providers we have to consider their status as private or non-private and/or for-profit or non-profit actors. After having analysed the actors separately we look at the relation between them, focusing on the one hand on the split between purchasers and providers, on the other hand on the split between purchasers and customers. When analysing the purchaser-provider split we are interested to know how the relationship between them is regulated and to what extent they are exposed to competition. When analysing the purchaser-customer split we are interested to know to what extent the roles of purchaser and customer are separated with respect to purchasing decisions and financing.\nDiversity in activation markets: In this section activation services in the Czech Republic, Finland, France, Germany, Italy, The Netherlands, Sweden, Switzerland and the UK will be discussed along the five identified dimensions (purchasers, providers, customers, purchaser-provider split and purchaser-customer split).The purchasers\nConclusion: In this article we explored diversity in the context of the marketised provision of activation services in nine European countries. Although we do not pretend that we have provided a complete overview of all dimensions of diversity, we can conclude that activation markets differ in a wide range of characteristics. Using the concept of quasi-markets as defined by Le Grand we are able to show that marketisation tendencies occurred to varying degrees and in various ways in the countries involved in our analysis. The development towards quasi-markets is characterised by a split between purchaser and provider of activation services, replacing monopolistic state provision with competing providers. The analysis showed that all countries indeed abandoned monopolistic state provision, most of them by introducing privately provided services, although each country involved in the analysis has its specific mix of for-profit and non-profit private and public agencies in the provision of activation services. But the provider-purchaser split was not implemented to the full. In most countries the public purchasers continue to act as providers of activation services. Only the UK, The Netherlands and Switzerland implemented a rather strict split between purchasers and providers for unemployment benefit recipients.The second element of quasi-markets outlined by Le Grand, the involvement of the state in financing activation services, is given in all countries. Some variance was found related to the involvement of the consumer in the purchasing decision. In two countries the decision was partly transferred to the consumer: Germany introduced a voucher system and The Netherlands introduced IROs where consumers (unemployment benefit recipients) choose their own activation trajectory.As a conclusion it can be stated that The Netherlands and Switzerland seem to have marketised their activation services more substantially than others; the UK is currently strengthening the role of marketised service provision. The German voucher system also points towards a considerable degree of marketisation but its impact until now is modest. All in all, the nine countries under investigation are far from the implementation of conventional markets but all of them show tendencies towards quasi-markets. Simultaneously, we also found instances of the opposite tendency of de-marketisation, in the sense of reinforced in-house provision of services. Related to the governance typology of Considine and the question whether the introduction of activation services markets amounts to a shift in types of governance or rather to a hybridisation of governance models, the evidence that we were able to provide in this article seems to point to the latter rather than the former. However, a more systematic insight into countries' activation markets will require further research.In the introduction we argued that studying market diversity is not only interesting in itself but may also be a first step in investigating broader issues related to explaining diversity and to the potentially heterogeneous effects and impacts of various market models on service effectiveness, service quality, service access, etc. We lack the room here to elaborate upon these issues in any detail. But more importantly, the data needed for such an exercise are not available. As is argued in the article on effects of governance reforms by De Graaf and Sirovatka in this issue, our insights into the impact of governance reforms in general are limited. And although we could say that marketisation is the most frequently and intensively researched governance reform, we are not aware of studies that investigate and compare various market models in any depth. Evaluation studies of activation markets often abstract from concrete market characteristics. Given the contested nature of marketisation on the one hand and the diversity of activation markets on the other, it might be useful to reconsider the focus of research into the marketisation of activation services. Rather than asking whether or not markets do a better job than the public sector, it might be more useful to investigate the question what type of markets produces what type of effects for what types of target groups of unemployed people. Against this background, we hope that this article has made a useful contribution to developing a research agenda for comparative activation market research.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this article is to analyse the diversity of markets for the provision of activation services.\nDesign/methodology/approach: - The article is based on the outcomes of a project involving nine European countries. The project investigated changing forms of governance of income protection schemes and activation services for unemployed people. Diversity is investigated by focusing on five dimensions of diversity derived from the quasi-market concept as developed by Le Grand: the purchasers, the providers, the customers, the purchaser-provider split and the purchaser-customer split.\nFindings: - The paper finds considerable diversity in the design of markets for the provision of activation. Diversity is visible in all dimensions involved in the analysis. One interesting finding is that a full split between purchasers and providers hardly exists, although some countries have introduced a stricter split than others. Another finding concerns the voice and choice of service consumers, which seems hardly affected by the introduction of market mechanisms in the provision of activation. Finally, marketisation does not seem to be an irreversible project, as de-marketisation processes were identified as well.\nOriginality/value: - Most current research into activation markets and their effects pays little attention to the issue of diversity in the design and functioning of markets. This article argues in favour of more systematic research of market diversity and of the variety of effects of various market models. Rather than comparing marketised with public service provision, a stronger focus on various market models may strengthen our insight into how service provision models affect the effectiveness of activation services.\n",
"role": "assistant"
}
] |
Are restaurant menus vectors of bacterial cross-contamination? A pilot study in Turkey
|
[
"Turkey",
"Restaurant",
"Environmental food hygiene",
"Menu cleanliness",
"Menu microbiological load",
"Hygiene"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The consumption of food contaminated by foodborne pathogenic microorganisms or their toxins may lead to death, diseases, prolonged hospitalization, and financial burden (Dalton et al., 2004; O'Brien et al., 2002; Velusamy et al., 2010). It is well-established that factors threatening food safety must be precluded at the food production stage (Carraturo et al., 2016). Furthermore, the conditions where food is served, as well as various environmental factors, have been associated with foodborne diseases (Miranda and Schaffner, 2016; Teixeira et al., 2007). Food processors are the leading environmental factors threatening food safety (Byrd-Bredbenner et al., 2015; Green et al., 2006). The World Health Organization and Centers for Disease Control and Prevention (CDC) indicates that contamination of food may result from environmental contamination in the process from food production to consumption (World Health Organiation, 2015; Centers for Disease Control and Prevention, 2015). In a recent CDC study on \"Epidemiology of foodborne norovirus outbreaks, United States, 2001-2008,\" it was demonstrated that infected food handlers contributed to 29 percent of foodborne disease outbreaks by contaminating food (Centers for Disease Control and Prevention, 2014). Apart from the food processors, service equipment, tables, and cleaning cloths used, menus are also among the main factors leading to bacterial contamination and transfer in the service process, and are a critical checkpoint (Choi et al., 2014). It is important to be aware of the hazards of other environmental contamination sources outside the restaurant kitchen, such as the furniture, equipment, and menus, and to implement the necessary control measures. The risk of menus being a particular biological hazard, due to direct and indirect contact with the consumer, is considered to be a source of transmission for foodborne diseases (Choi et al., 2014). For up to 24 h following their contact with the menu, microorganisms can be transmitted onto the fingers (Sirsat et al., 2013); therefore, the risk of contamination should not be overlooked, and standard operating procedures must be established to control this hazard (Choi et al., 2014). The central question, considered by the limited number of studies conducted over recent years, was whether the cleaning of menus in a restaurant should be included in the daily cleaning protocol (Alsallaiy et al., 2016; Gamez et al., 2016; Sirsat et al., 2013).\nMethodology: The present study was conducted between April and June 2015. Inclusion criteria were: the laminated menus contained more than two pages, the first-class restaurants were certificated by relevant regulations (Ministry of Culture and Tourism of Republic of Turkey, 2005); and the restaurants had consented to participate in the study. In total, 41 first-class restaurants located in Ankara, Turkey and which were selected using the systematic sampling method, were included in the study. The investigators provided the restaurant managers and service personnel with a brief questionnaire to survey the high periods of traffic in the restaurant and determine how the menus were presented to the consumers. The menus included in the study were folded and contained multiple pages. They were either kept in a separate room and handed out to the customer at the table on arrival or displayed on the table with table menu holders. Sampling from the surface of the menus was carried out between 14:00 and 16:00 h, following a high period of customer traffic.\nStatistical analysis: The data were statistically analyzed using SPSS v15.0 software (SPSS Inc., Chicago, IL, USA). The variables were investigated using visual (histograms, probability plots) and analytical methods (Kolmogorov-Smirnov/Shapiro-Wilk's test) to determine whether or not the data were normally distributed. Descriptive analyses were presented as means and standard deviations for normally distributed variables. Since log-transformed values of the ATP+AMP hygiene monitoring (RLU) and ACC (log cfu/cm2) assays were normally distributed; means and standard deviations were used. The student's t-test was used to compare mean RLU values between hand-out menus and menus left on the table. The impurity status of the menus was expressed by menu presentation type using cross-tabulations. The Fisher's exact test (because kh2 test assumptions do not hold due to low expected cell counts) was used to compare these proportions in different groups. The criterion for significance was set at p<0.05. The inter-rater agreement between the two assays (ATP+AMP hygiene monitoring and ACC) in determining the dirty and clean menus were investigated using the k test where values <0= poor agreement, 0.0-0.20=slight agreement, 0.21-0.40= fair agreement, 0.41-0.60= moderate agreement, 0.61-0.80= substantial agreement, and 0.81-1.00= almost perfect agreement (Landis and Koch, 1977).\nFindings: The ACC was 1.52 log cfu cm-2, as shown in Table I. The ATP+AMP hygiene monitoring values was 1844.20 RLU 100 cm-2. Enterobacteriaceae, E. coli, and S. aureus were not detected on the restaurant menus.\nDiscussion: The impurity of the menus has been shown as providing information for the customer about the overall cleanliness of the restaurant (Henson et al., 2006) and the habits and attitudes of the employees toward hygiene (Kusumaningrum et al., 2003). The present study is one of the pioneering studies to evaluate menu hygiene and to our knowledge, it is the first topic of its type to be conducted in Turkey.\nConclusion: This study showed a high level of contamination on menus cards even in first-class restaurants, and the contamination status did not differ according to the mode of menu distribution, such as leaving them on the table or handing the menu to the customer. It is considered that pathogenic microorganisms are transferred from the menus by the customers' hand. Therefore, hygiene courses for service personnel should include information about the risk of contamination from the menus and appropriate cleaning and disinfection practices. Additionally, during service, it is necessary to monitor hygiene control of the menus by using objective methods within the framework of operational prerequisite programs. Maintaining health, and hygienic hand washing practices are of utmost importance, particularly following contact with menus by the consumer/customer. Finally, it is necessary to determine the contamination threshold value to assess the effectiveness of cleaning methods.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to evaluate the hygiene of laminated menus.\nDesign/methodology/approach: The cleanliness of 41 laminated menus from first-class restaurant in Ankara, Turkey, was monitored according to the level of bioluminescence measured using a rapid hygiene monitor (Lumitester PD-20 and LuciPac Pen). Aerobic colonies, Escherichia coli, Enterobacteriaceae spp., and Staphylococcus aureus were enumerated.\nFindings: Based on the adenosine triphosphate+adenosine monophosphate hygiene monitoring levels, analyses showed impurity in 82.9 percent of the samples. The aerobic colony count on the restaurant menus showed a mean 1.52 log cfu/cm2. The study results demonstrated a high-level impurity on the menus in first-class restaurants in Ankara.\nOriginality/value: This study is the first in Turkey to assess the microbiological load of menus. This paper adds relevant insights into the assessment of menu hygiene and highlights the high contamination of the restaurant menus. Useful information about including the restaurant menus in the daily cleaning protocol is provided.\n",
"role": "assistant"
}
] |
Hand-held computers to increase accuracy and productivity in agricultural work study
|
[
"Agricultural environment",
"Hand-held computers",
"Time study",
"Work study"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Horticultural crops are \"production systems\" whose products - fruits, vegetables and flowers, are highly sensitive, to mechanical pressure, in particular, and thus, require gentle, accurate and complicate handling. This characteristic makes the replacement of manual work with machines extremely difficult. Therefore, most fruit, vegetable and flower cultivars growing and production tasks, e.g. trellising, fruit harvesting, sorting and packaging are performed manually, which makes manual labor a major cost component in horticulture operations (Eben-Chaime et al., 2011). In an early survey during the mid 1970s, bedding plant growers attributed about 25 percent of plant production costs to labor (Aldrich and Bartok, 1992). Nelson (1991, p. 574) estimate labor contribution to the total costs at 34.81 percent, including 5.58 percent depreciation and 2.5 percent interest. Unlike labor, the last two are not \"operational costs\" and when excluded from the total, labor contribution to the operational cost rises to 38 percent. In greenhouse crops, labor is a key factor for the development and the maintenance of the sector (Manzano-Agugliaro and Garcia-Cruz, 2009). Greenhouse operational costs of roses, tomatoes and cucumbers, in different places on the globe are listed in Tables 1-5 of Hanan (1998, p. 6). In Southern Spain, the labor cost in greenhouse crops such as tomatoes, lettuce, peppers, melons, water melons, courgettes, cucumbers and beans are 36.6-40 percent of the total cost (Manzano-Agugliaro and Canero-Leon, 2010; Montoya-Garcia et al., 2013). Despite all differences and the large variability in absolute magnitudes, labor contributes about 40 percent of the operational cost, in all cases. The portion increases in open fields where maintenance and holding cost are much smaller. In addition, in all these references, labor is the largest single cost contributor. Moreover, the enormous manpower required for the various operations causes bottlenecks that downgrade productivity - reduce the yield and increase costs. This high manual labor requirement, impedes the reduction of costs and improvement of working efficiency (Bechar et al., 2007).\n2. Methods and materials: 2.1 Work measurement - theoretical background\n3. Results: 3.1 Quality performances\n4. Discussion and conclusions: In this paper the use of a HHC platform for agricultural work measurement has been presented, and its merits discussed. The primary advantages of the HHC are its size and weight. These advantages contribute in two ways. First, the removal of the physical barrier to work measurement in agriculture, in a similar way it does in the logging production (Wang et al., 2003). Due to the products - fruits/flowers, being in fixed position, the workers are not standing in fix stations but move all across the field. The size and weight of the HHC enables to follow the workers while performing the measurement.\n",
"role": "user"
},
{
"content": "Purpose: Labor is the largest single cost contributor in agriculture. Accurate estimation of labor requirements is, thus, a key to cost reduction. Work measurement is the professional discipline for this type of estimations, in the industrial engineering domain. Horticulture, however, lays a substantial barrier to work and, thus, to work measurement. Till harvesting, its products - fruits/flowers, are in fixed positions, and for all tasks the workers have to arrive at the plant. The purpose of this paper is to develop, test and analyze a system to perform horticultural work study in agricultural environments in order to accurately estimate the required amount of labor for each activity and to improve productivity.\nDesign/methodology/approach: In this paper, the development of an advanced system for work study on a hand-held computer (HHC) platform for horticultural work measurement is presented and an experimental study was conducted. The methodology consists of characteristics of the system platform including hardware, interface and software, development of a dedicated measuring software, a controlled experiment in agricultural environment and a statistical analysis.\nFindings: The study shows that a single surveyor who uses traditional tools is subject to measurement errors, which can be reduced only with the assistance of a second surveyor. The study further shows that the HHC platform enables to avoid this second surveyor - a single surveyor who uses the HHC platform performs as accurate as two surveyors who use traditional tools. Of course, being computers, the HHC platform maintains the advantage of error free data transfer, in practically negligible time.\nOriginality/value: This paper presents a unique approach to perform work study in agricultural environment and contributes to minimize the errors accumulated in the process and the manpower required to perform the measurements.\n",
"role": "assistant"
}
] |
How internal attributions affect knowledge sharing behavior
|
[
"Knowledge sharing",
"Ability",
"Effort",
"Firm sizes",
"Internal attribution"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Knowledge sharing is an important aspect of knowledge management, which can be an organization's sustainable competitive advantage. Through knowledge sharing, the organization could integrate both internal and external knowledge, detect opportunities and obtain more advantageous positions (Gavirneni et al., 1999). Although knowledge sharing results in improved organizational effectiveness, few firms have been successful in encouraging staff to share their knowledge and facilitating their sharing capabilities to support the related activities.Organizations are confronted by a variety of knowledge sharing problems, which can eventually lead to resistance to knowledge sharing behavior, such as barriers within organizations, a lack of technological resources and training, as well as differences in individuals' skills (Riege, 2005). Moreover, another factor determining knowledge sharing's success or failure is an individual's motivation to participate in community knowledge generation and sharing activities (Jolaee et al., 2014).\n2. Theoretical development: 2.1 Causal analysis of knowledge sharing barriers\n3. Research model and hypotheses: This study developed a research model to examine and compare the effects of organizational and technological barriers as causal factors on employee effort and ability in relation to knowledge sharing in both large and small firms. The research model is based on attribution theory, which is a widely accepted model for validating individual and organizational behavior. Figure 1 shows that perceived lack of effort and low ability are seen as internal attributions that mediate the relationship between organizational and technological barriers and knowledge sharing behavior, including the moderating effects of firm sizes on the relationship. The research hypotheses are discussed in the following sections.\n4. Research method: 4.1 Measurements\n5. Analyses and results: The research models were analyzed using a latent structural equation modeling technique, the partial least squares (PLS) approach. PLS is an advanced statistical method that allows optimal empirical assessment of a structural model together with its measurement model, which links each construct with a set of indicators measuring that construct. For this study, the PLS model was created in SmartPLS 3.0 to test the research hypotheses. PLS was appropriate for the data analysis in this study because it maximizes the variances explained in the dependent variables, and it is less demanding in terms of sample size. Furthermore, the assumption of multivariate normality does not restrict the use of PLS.\n6. Discussion and conclusion: By considering both organizational and technological barriers in its theoretical model, this study has accounted for a substantial portion of the variance in knowledge sharing behavior. Moreover, it illustrates how firm size may moderate the relationship between causal factors (organizational and technological barriers) and internal attributions (perceived low ability and lack of effort). Tables IV summarize the results of this study.\n7. Implications: This study applies causal impediments to explore how internal attributions discourage knowledge sharing behavior. We identify and examine two major causal impediments, which are organizational and technological barriers, and examine how these impact firms of different sizes with regard to how employees share knowledge. A review of self-efficacy theory indicates that an individual's beliefs in their abilities can predict their behavior and attitudes, and that negative beliefs may discourage knowledge sharing behavior in both large and small firms (Kuo and Young, 2008). Furthermore, when a firm grows and there are then multiple groups within the organization, then the efforts made to share knowledge among employees seem to decrease.\n8. Limitations and future research: Although it has new findings, this study has the following limitations that need to be addressed by future research. First, although this study concentrates on exploring the effects of organizational and technological barriers on employees' knowledge sharing behavior at firms of different sizes, it did not address the different situations of employees who work at large firms but distributed into several small units in different locations, and those who work at large firms and have centralized management in same location. In addition, younger employees may use technology to share knowledge more than older employees. The impacts of different employee ages and the average ages at large/small firms on the research model were not discussed in this study. As such, researchers may consider differences in firm size and division location in the sampling process, as well as the ages of employees at large/small firms. Such efforts would help in making a more comprehensive assessments of the issues examined this work.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to compare the effects of organizational and technological barriers to knowledge sharing between large and small firms through the lens of attribution theory.\nDesign/methodology/approach: The structural equation modeling approach was applied to estimate the conceptual model by using survey data from a list of Taiwan's top 1,000 manufacturing and 500 service companies. A total of 229 valid questionnaires were collected.\nFindings: The empirical results show that both organizational and technological barriers have relationships with an individual's effort and ability with regard to knowledge sharing behavior. When organizational barriers occur, the perceived lack of effort has a direct effect on knowledge sharing behavior for large firms, while negative sharing behavior among employees of small firms is influenced by the perception of low ability through the perceived lack of effort.\nOriginality/value: A review of the literature reveals organizational and technological barriers that lead to the negative influences of internal attributions on knowledge sharing. This study, therefore, contributes to a comprehensive perspective on how to encourage knowledge sharing behavior at different sizes of firms.\n",
"role": "assistant"
}
] |
Trends and gaps for integrating lean and green management in the agri-food sector
|
[
"Lean",
"Green",
"Agri-food",
"Environmental sustainability"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The agri-food sector has a great economic and political importance for countries. It has a direct relationship with the fulfillment of human needs, the support of employment and economic growth, and has a great impact on the natural environment (Cagliano et al., 2016). The agri-food industry is of particular interest as it simultaneously attempts to deal with a number of complicating characteristics often unheard of in other industries (e.g. the short shelf life of raw materials, seasonality) (Powell et al., 2017).\n2. Background: 2.1 Agri-food sector\n3. Research method: We conducted a systematic review, as a valid research approach and a necessary step in structuring a research field (Easterby-Smith et al., 2015). This qualitative method has already been successfully used in other similar studies on topics closely linked to the present study, including green supply chain management (Srivastava, 2007) and lean manufacturing, Six Sigma and/or sustainability (Cherrafi et al., 2016; Sunder et al., 2018). As pointed out by Drohomeretski et al. (2015), the main objective of a review is not only to evaluate how much is being published but, rather, to identify trends and gaps in the literature.\n4. Research results: This section describes the data obtained from the sample of papers. After a descriptive analysis of the sample is presented, the integration of lean and green in the agri-food sector is discussed. The method described in the previous section enabled us to define the four lines of research linked to the interrelationships between lean, green and the agri-food sector. To be more specific, the four research topics and number of papers evaluated per topic are:\n5. Discussion: Based on the literature review, the benefits and synergies of the lean and green approach could be extrapolated and applied to the agri-food sector understanding its specific and contextual factors. For example, lean does not, in its original form, address the safety aspects of food production (Engelund et al., 2009). Thus, it is necessary to adjust some lean and green principles to the specific requirements of a production scenario in order to correctly and efficiently implement those approaches in the agri-food sector (Dora and Gellynck, 2015). Furthermore, there is an important gap in the theoretical, prescriptive and quantitative development of LG-A topic. Conceptual frameworks, mathematical models and statistical analysis should be conducted, respectively, to fill this gap.\n6. Conclusions: This paper presented a systematic review with the aim of investigating the contributions of and interrelations between the lean and green approaches in the agri-food sector. Based on a sample with 117 papers, our research identified contributions to and gaps in this body of knowledge. Given the worldwide agenda for improvement of the agri-food systems, it must be stressed that there is a significant gap regarding lean and green integration in this sector. There is extensive literature in the context of green/agri-food, green/lean and lean/agri-food. However, research combining all three topics has not yet emerged. It is hoped that this study will inspire further research and exploration in this area. We have highlighted key aspects, gaps, approaches and methodologies that, in our opinion, set a new and promising research agenda for implementing the lean and green approach in the agri-food sector. The contributions of this paper do not only advance the theoretical knowledge of the lean and green field, they are also beneficial for organizations, especially agri-food companies, which aim to effectively deploy lean in their processes to meet current global politics, as Horizon 2020.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze gaps and trends, as well suggest approaches and methodologies that should be addressed by future studies for implementing the lean and green management in the agri-food sector.\nDesign/methodology/approach: Based on a sample with 117 papers, this paper presents a systematic review on the integration of lean and green in the agri-food sector.\nFindings: Key findings indicate that research on lean and green topics has increased in recent years, an important gap in the integration of lean and green in the agri-food sector has been identified. Two paths that remain open for further research are detected: the lack of theoretical, prescriptive and quantitative research and the possibility of integrating the two most used tools of lean (i.e. value stream mapping) and green (i.e. life-cycle assessment).\nPractical implications: This study does not only advance the theoretical knowledge of the lean and green field, but it is also beneficial for agri-food companies who aim to effectively deploy lean and green in their processes in order to improve both operational and environmental performances.\nSocial implications: No other industry matches in such a complete way the agri-food industry's challenge of sustainability that is due to the amount of resources it consumes and its interrelatedness with the well-being of humanity.\nOriginality/value: There are separate streams of established research on lean and green management in the agri-food sector, yet very few authors have addressed the intersection of these strategic initiatives.\n",
"role": "assistant"
}
] |
Performance, valuation and capital structure: survey of family firms
|
[
"Family firms",
"Corporate ownership",
"Financial performance",
"Corporate governance"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Several recent studies reveal increasing importance given to family relationships and their businesses. The prime motivation for scholars directing academic research toward family firms has largely been their dominance on most nations' economic landscape. For instance, in the United States, 60 percent of the Gross Domestic Product (GDP) is generated by family-controlled businesses (McConaughy et al., 2001). These are often thought to be \"microfirms\", however, family-owned businesses make up approximately 35 percent of the 500 Fortune firms (Anderson and Reeb, 2003a). Indeed, regarding the predominance of family firms in particular regions of the world, family control is often the case in large US companies (Bhattacharya and Ravikumar, 2001; Anderson and Reeb, 2003a; Villalonga and Amit, 2007) as well as among corporations operating in Western European Countries (Faccio and Lang, 2002). Additionally, several studies document the importance of family firms in the East Asian region (Claessens et al., 2000). Despite variations between countries, family business represents a substantial portion of an economy and has a huge impact on the economy as a whole, which contrasts to Berle and Means's image of the ownership concentration corporation, i.e. ownership is dispersed among small shareholders, while control is concentrated in the hands of managers.It is generally accepted that the traditional finance literature has little to say about how family (taking into account specific governance characteristics such as family ownership, family control and family management) affects firms' performance, value and capital structure. Previous studies show that family controlled firms differ from non-family controlled firms (e.g. Daily and Dollinger, 1991; Gallo, 1995). Nonetheless, a little theoretical and empirical research has been done to support and advance understanding of this premise. Thus, for a complete study of family business, it is first necessary to clarify the definition of the family firm.The literature on family businesses is wide-ranging and it is difficult to find consensus on the exact definition of a family firm, and numerous attempts have been made to articulate conceptual and operational definitions of them. The typical family business has been characterized as an organization controlled and usually managed by multiple family members (Anderson and Reeb, 2003a), often from multiple generations (Gomez-Mejia et al., 2001). Despite this fact, research does not provide us with a monolithic picture; so we define a family firm whenever the members of the family and closely connected others have fractional equity ownership and/or when the family serves on the board of directors. Thus, the scope of this survey is related to public quoted firms controlled by a family.We, therefore, divide our literature review in two blocks. First, we review current literature related to the impact of family ownership on firm performance and valuation. Our main objective concerns the question: Does family ownership really affects firms' performance/valuation? Second, we focus on how ownership/governance structure influences capital structure, as a proxy for risk aversion. Our main concern regards the fact that the empirical literature shows family firms to be less leveraged (e.g. Anderson and Reeb, 2003a; Gomez-Mejia et al., 2007), which challenges the traditional capital structure literature. However, before presenting the main points of the literature on how family (ownership, control, management) affects firm performance, this paper engages in a discussion on whether family firms are really differently from non-family firms, and if so, how and why they are different. The main contribution of this paper is to update the literature taking into consideration not only the differences between family and non-family firms, in terms of valuation and performance, but analyzing the impact of the family as an additional organizational variable in a firm's financial issues.The remainder of the paper is organized as follows. Section two analysis whether family firms are differently from non-family firms, and if so, why they are different. The third section reviews how family affects firms' performance and value, concerning three main corporate governance variables: ownership concentration, ownership control and managerial ownership. Section four reviews the literature on the impact of ownership/governance structure on capital structure, regarding family firms. Section five concludes the survey.\n2. Are family firms really different?: 2.1 Definition of family firm\n3. How family ownership affects firm performance and value?: 3.1 Family ownership and firm performance\n4. How does owner risk aversion affect ownership structure in family firms?: While there is a relatively large amount of literature, particularly, related to widely held firms, on the effects of ownership on firm performance and value (e.g. McConnell and Servaes, 1990; Demsetz and Villalonga, 2001), the relationship between ownership structure and capital structure remains largely unexplored. Some review of the extant interdisciplinary literature indicates a complex array of factors that influence the relation of capital structure as a proxy for control risk propensity or risk aversion and ownership structure in family firms.4.1 Capital structure theories: an overview\n5. Conclusion: The objective of this paper is to review the literature on how family (taking into account specific governance characteristics such as family ownership, family control and family management) affects firm performance, value and capital structure.The literature review has revealed that family firms on average perform better than non-family firms. Family ownership is, on average, positive for firms' valuation, and also for firms' performance. Nonetheless, this is a non-linear relationship, which suggests that ownership structure is endogenous. So, it is important to distinguish between ownership concentration (claims on firms' cash-flow) and control (the holding of voting rights at board level). Therefore, family-controlled firms may use mechanisms to enhance their voting control, such as dual class shares or pyramidal structures which allow them to have incentives and power to systematically expropriate wealth from the firm's other claimants, creating severe moral hazard conflicts between the founding family and minority shareholders. So, we are now able to ascertain that the main institutional factors embodied in a corporate governance system affect the relationship between ownership structure and firm value. Thus, further research should examine the relationship between family ownership and firm performance taking in account the legal environment. One policy implications is the necessity to improve the minority shareholders protection from the risk of expropriation by large shareholders.The review presented also shows a significant positive effect of founder management on both valuation and performance, whereas the descendant effect is negative for both performance and valuation, when compared to professional management. This relationship could be explained by the fact that restricting executive talent to a labor pool of family members can be challenging. The family's role in selecting managers and directors can create impediments for third parties capturing control of the firm, creating greater managerial entrenchment and consequently a decrease in performance and value. Since performance and value are lower for such firms, investors impose a premium for the allocation of control to these families. This concession is also a result of a much smaller probability of a takeover, since typically families are a long-term controlling owner.Regarding the influence of family business owners' decisions on the choice of different forms of finance there is a scarcity of literature. Nevertheless, the empirical evidence shows that family firms present low levels of debt. This evidence is interpreted as an external manifestation of a firm's control risk aversion, which challenges the traditional financial theories on capital structure. Indeed, despite families' desire to retain control, the desire to reduce risk is much stronger. This behavior could be explained by the fact that family business owners' desire to protect their monetary and non-monetary benefits induces a business perspective that lasts further into the future (e.g. one or more generations) than the perspective of a more short-term oriented manager with a time horizon from a few years to one working life. Furthermore, controlling shareholders of family controlled firms typically have undiversified portfolios. Thus, further research on the issue of the separation of business and family wealth, or the allocation of private investments to company accounts, is in order.This survey investigates the link between ownership, performance and value by relating ownership to several dimensions of the \"family effect\". The argument that family controlled businesses should be more efficient than professionally run firms, due to lesser monitoring costs in family controlled firms, is controversial. In this way, agency models assume that the effects of concentrated ownership and owner-management will lead to a minimized or even zero level of agency costs. Given significant shareholdings, family owners possess the incentive, power and information to control their managers, thereby reducing free-rider agency costs and boosting returns. Moreover, managers and large family shareholders are often the same persons, and therefore, the residual claimants bear (nearly) all the costs and receive (nearly) all the benefits of their actions. Nonetheless, no separation between ownership and management can offset the advantages in monitoring and disciplining decision agents. It is also true that family-dominated businesses are more likely to be characterized by extraordinary dividend payouts, entrenched managers and a redistribution of wealth from family to employees. Moreover, family differences and role conflict can lead to behavior that does not support the best interests of the firm. Psychological conflict within the family (such as sibling rivalry, autocratic behavior, and nepotism) can offset the benefits of reduced monitoring. Thus, the role of trust, altruism but also stewardship should be further analysed in order to answer the question of whether the agency advantages, as proposed by traditional financial theory, prevail or if the disadvantages are greater. The challenge for such research is not whether a business is family or non-family run, but the extent of the family's involvement and its influence on the firm.\n",
"role": "user"
},
{
"content": "Purpose: - Most countries often have public companies with large controlling owners, typically a family. This empirical evidence aims to contrast with the classical view of the largest dispersed firm presented by Berle and Means and challenge the findings by Bhattacharya and Ravikumar, who predict that the shares held by families will decrease if an efficient financial market is put in place. Therefore, family firms represent an important group in the stock market today. Thus, the purpose of this paper is to analyze the effect of the family as a controlling owner on firms' performance, valuation and capital structure.\nDesign/methodology/approach: - The paper reviews the current literature related to how family (taking into account specific governance characteristics such as family ownership, family control and family management) affects firms' performance and value.\nFindings: - The literature review showed that founder family control and professional (outside) management increase performance, whereas excess control via control enhancing mechanisms (such as dual class shares and pyramidal structures) and descendent management produce both lower valuation and performance. This evidence suggests that families have the incentives and the power to systematically expropriate wealth from minority shareholders.\nOriginality/value: - Previous research shows that family firms on average perform better than non-family firms. But this is a non-linear relation due the fact that the relationship between family ownership and performance cannot be identified without distinguishing between control and cash-flow rights. Thus, the literature review as a whole emphasizes that the incentives for the controlling shareholder to engage in expropriation are a function of the institutional framework in which the firm operates. So, for further research, it is important to investigate how family firms perform in different corporate governance systems. A policy implication is the necessity to improve minority shareholders' protection from the risk of expropriation by large shareholders.\n",
"role": "assistant"
}
] |
Influence behaviors and organizational commitment: a comparative study
|
[
"Organizational commitment",
"A comparative study",
"Influence strategies",
"Influence tactics",
"National culture"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Tactics used by supervisors to influence their subordinates to perform certain tasks often remain salient in the memories of the subordinates for years to come. In some circumstances, praising a subordinate or offering something in exchange for cooperation works best, but in others, managers simply choose to use pressure to get what they need, even though it can strain relationships (Scott, 2004). Research has provided some evidence that sometimes even the least effective tactics such as pressure or coalition building can gain a subordinate's immediate compliance to perform a task (Yukl and Tracey, 1992). However, the questions are: would the subordinate really be willing to take up the job? Will s/he be happy to do the task with her/his best efforts?\nPrior studies of influence tactics: Definition of influence tactics\nTheoretical framework and hypotheses: Managers need to use \"influence strategies to get their subordinates to work toward organizational goals\" (Rao and Hashimoto, 1996, p. 444). Prior studies have shown some evidence that influence behavior is related to organizational outcomes. In this study, the rationale of proposing influence behaviors would appear to affect organizational commitment because influence behaviors are daily occurrences, and because influencing someone else to get something done plays a central role in most organizations (van Knippenberg et al., 1999). Thus, apart from having influence to get a target (subordinate) to commit to an immediate assigned task, the effect of repeated influence behaviors is to strengthen the subordinate's belief in and acceptance of the organization's goals and values. The target individual will thus show a willingness to exert considerable effort on behalf of the organization (Mowday et al., 1982).\nStudy 1: survey study: Sample and procedures\nStudy 2: the follow-up interviews: Sample, procedures and data analysis\nDiscussion and implications: The main strength of this mixed-method study is that I have sought to identify three specific factors (i.e. time, direction and country) for predicting the effects of influence behaviors on organizational commitment. In the following, these factors are discussed.\n",
"role": "user"
},
{
"content": "Purpose: Prior research has identified the outcomes of influence tactics as short-term task commitment, compliance and resistance. This paper argues that leaders' downward influence behaviors should also have an impact on followers' organizational commitment. The purpose of this paper is to examine the effects of three influence strategies (11 downward influence tactics) on organizational commitment, and the moderating effect of national culture.\nDesign/methodology/approach: The study draws on survey data (n=185) and follow-up interviews (n=19) from Hong Kong Chinese employees who work for Hong Kong Chinese or Japanese managers.\nFindings: The quantitative results show that all rational tactics, the inspirational appeal and pressure tactics had effects on organizational commitment. Drawing on the survey and follow-up interview data, three specific factors in the use of influence tactics on organizational commitment are identified. Results suggest that some tactics are more universal and able to provide \"long-lasting\" effects compared with other tactics in enhancing or reducing employees' organizational commitment. The results of the entire study also show that most rational influence tactics seem to be convergent; yet, there is also evidence that other tactics are more culturally specific in generating employee commitment.\nOriginality/value: This is the first empirical study to examine the effects of influence behaviors on organizational commitment with both inter- and intra-cultural samples. The study has also drawn on interview data to demonstrate examples of effective and ineffective influence tactics used by superiors, thereby offering managerial hints to managers on how to exercise their influence behaviors effectively.\n",
"role": "assistant"
}
] |
Measuring the effect of Chinese brand name syllable processing on consumer purchases
|
[
"Purchase intention",
"Product category",
"Brand name syllable",
"Non-store shopping",
"Seller reputation",
"Skepticism"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Brand names play a crucial role in the marketing of products and services and in their acceptance by the public (Charmasson, 1988). Past studies emphasized the importance of brand names in their contribution to brand equity (Keller, 1993), the effect of the gender of the brand on the evaluation of cross-gender extension (Jung and Lee, 2006), but focussed studies on the analysis of brand name syllables appear limited. Since the linguistic aspect is the essence of branding and of particular importance for the success of international brand names (Usunier and Shaner, 2002), this brand name syllable gap needs to be filled.\n2. Theoretical background and hypotheses development: 2.1. Brand name syllables\n3. Method and results: From 2011 to 2013, total 252 papers were published in Journal of Consumer Research. Among them, at least 197 papers used experimental design to find the causal effect and 189 papers used at least three studies separately to test their conceptual model. To avoid multiple-treatment interference, different groups of participants were involved in their studies. We also found that to avoid too many factors made subjects hard to make the right decision, most of the papers did not use four factors between-subjects design. Only one paper (Noseworthy et al., 2011) used four factors (ad contextxgenderxworking memory constraintxproduct congruity) between-subjects factorial design. Actually gender knew already, the experimental decision difficulty level was just like that of three factors design. Following their approach, thus we use three factors design at most in our four studies and analysis of variance or regression analysis statistical method to test our four hypotheses separately and the conceptual model.\n4. General discussion: From these studies, we find all the hypotheses are supported (Table I).That is, customers' purchase intention is more favorable for 2-syllable brand names than for multi-syllable ones. The product category will moderate the effect of the brand name syllable on the purchase intention. Multi-syllable brand names will decrease the purchase intention to opt for search goods/services more than experience goods/services and decrease the purchase intention for experience goods/services more than for credence goods/services. Moreover, multi-syllable brand names will decrease the purchase intention to opt for sellers of low repute more than for sellers of high repute. We also find consumers' skepticism will moderate the interaction effect between brand name syllable levels and seller reputation outlined in H1. That is, among low skepticism consumers, brand name syllable levels will not affect the purchase intentions; among high skepticism consumers, multi-syllable brand names will decrease purchase intention to opt for sellers of low repute.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine consumer reactions to product categories, online seller reputation, and brand name syllables.\nDesign/methodology/approach: This paper uses four experimental designs to explore the seller reputation, product category, and brand name syllable effects in internet shopping. The authors chose sellers of (low/high) repute from Yahoo Mall. ANOVA is used to evaluate the results.\nFindings: Seller reputation moderates the effect of the brand name syllable level on purchase intention and product category moderates the effect of the brand name syllable level on purchase on internet (experiment 1). Consumers take the longest time to make purchasing decisions when buying credence goods or buying from sellers of low repute and that the response time mediates the moderating role of the product category (experiment 2) or reputation (experiment 3). Moreover, the effect of brand name syllable levels chosen/assigned by sellers of low repute is weakened for consumers with low (vs high) skepticism toward non-store shopping (experiment 4).\nPractical implications: This study is helpful to online sellers if they can identify their reputation, product category and those consumers have skepticism, they can create extra profit through brand name syllable practice.\nOriginality/value: This paper extends the literature on consumers' brand name syllable processing by identifying important moderators and probing into the decision process. The results allow us to substantiate prior research and suggest prescriptive strategies for internet retailers.\n",
"role": "assistant"
}
] |
The COVID-19 crisis and sustainability in the hospitality industry
|
[
"COVID-19",
"Crisis",
"Sustainability",
"Hospitality industry",
"Sustainable consumption"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Sustainability within the hospitality industry has attracted considerable attention from researchers (Jones et al., 2016), although Ertuna et al. (2019) claimed that research on sustainability in hospitality management and tourism was underdeveloped. Research has focussed on the sustainability of a range of activities and operations within the industry, including the hotel sector (Guix-Navarrete et al., 2019), cruising (Bonilla-Priego et al., 2014), tourism (Jorgensen and McKercher, 2019), restaurants (Cavagnaro and Gehrels, 2009), theme parks (Milman et al., 2010) and events (Boggia et al., 2018). Research has also explored the origins of sustainability within the hospitality industry (Mackenzie and Gannon, 2019), and a number of the themes (Jones and Wynn, 2019) and frameworks (Boluk et al., 2019) were used to analyse sustainability. Within the hospitality industry, the concept of sustainability has also been applied to a range of natural disasters and emergencies (de Sausmarez, 2007; Loperana, 2016; Seraphim, 2018), but the focus has usually been on the local, regional or the national level. Gossling et al. (2020), for example, argued that previous work on pandemics has focused on their impact in individual countries, rather than on systemic challenges. However, the Coronavirus Disease 2019 (COVID-19) crisis is widely seen as a truly global problem. More specifically, the crisis has posed wide-ranging problems for the hospitality industry and has steered it, to use a metaphor from the cruising industry, into uncharted waters. There are fears that following the COVID-19 crisis, corporate financial support for sustainability initiatives will be dramatically reduced and possibly totally withdrawn, as companies concentrate their strategic thinking and resources on returning to some sense of normal business operations. That said, counter arguments suggest sustainability will continue to be important in re-establishing, and potentially strengthening, the relationship between the hospitality industry and both the natural and social capital on which it ultimately depends.\nCOVID-19 crisis: COVID-19, a newly discovered coronavirus, is an infectious disease. Coronaviruses, which can affect humans, as well as a range of other mammals and birds, were first identified in 1965, and since then, they have been responsible for a number of disease outbreaks. Severe Acute Respiratory Syndrome (SARS), for example, emerged in China in 2002-2003 and then spread to almost 30 countries in Asia, Europe and both North and South America. The most common symptoms of COVID-19 are fever, a cough, shortness of breath and general fatigue, and many people infected with the disease do not require any specialist medical treatment. However, people with over 70 years of age and those with underlying medical conditions may be prone to serious illness, and in some cases, the virus may prove to be fatal.\nCameo literature review: A cameo literature review, focused on sustainable development, crisis management, consumption and neoliberalism, provides an academic context and a set of reference points for the paper. Defining sustainable development is not straightforward but the following definition, \"development that meets the needs of the present without compromising the ability of future generations to meet their own needs\" (World Commission on Environment and Development, 1987, p.43), first coined in the late 1980's, is still widely used. More specifically, there are a variety of definitions of sustainable development, and by way of a summary, there are definitions based in ecology, and there are broader definitions, which embrace social, economic, environmental, ethical issues and governance. Some previous work on epidemics can be seen to be associated with the broader definitions of sustainability. Hung et al. (2018), for example, examined the SARS 2003 and the H1N1 (Swine Flu) 2009 outbreaks in Hong Kong. In both situations, the hospitality industry looked to adopt a comprehensive recovery plan to sustain businesses. In a short review, Pine and McKercher (2004) analysed the impact of the 2003 SARS epidemic on Hong Kong's tourism industry and, more specifically, on hotels and airlines and concluded that while the industry quickly recovered following the outbreak, contingency planning was vitally important in managing the crisis. Gossling et al. (2020) provided an overview of the early events of the COVID-19 crisis, outlined its potential impact on sporting events, restaurants and cruises and suggested that the challenge for the hospitality industry was to learn how to accelerate the transition to a sustainable future. More generally, Di Marco et al. (2020) suggested that the interactions between changes in the environment and the establishment of infectious diseases had received limited attention from the scientific community, and that it had not been integrated into research work on sustainable development. Looking forwards, Di Marco et al. (2020) recommended that it would be wise to include the study of the emergence of infectious diseases within sustainable development planning.\nRole of the hospitality industry in economic and social life: The hospitality industry is a major part of the world economy. Broadly defined, the hospitality industry embraces a wide variety of accommodation facilities, from resorts and hotel complexes to Airbnb homestays; food and drinks services and facilities, including restaurants, fast food outlets, bars and cafes, coffee shops and nightclubs; and travel and tourism, such as package holiday tours and ocean and river cruises. At the same time, a hospitality offer is often an important component of a large variety of sporting, entertainment and leisure, events and venues. Traditionally, many of the elements within the hospitality industry have essentially been considered as luxuries, which people can enjoy when they have fulfilled the basic needs of food, clothing and shelter. Indeed Revfine (2019, webpage), a knowledge platform for the hospitality industry, suggested that \"a defining aspect of the hospitality industry is... that it focuses on ideas of luxury, pleasure, enjoyment and experiences, as opposed to catering for necessities and essentials\". However, the COVID-19 crisis has called such a characterisation of the hospitality industry into question.\nSustainability and the COVID-19 crisis: Several changes can be identified in the relationships between sustainability and the hospitality industry following the onset of the COVID-19 crisis. Environmentally, the collapse in the number of air flights and the decline in the volume of motor vehicle traffic, along with the closure of large swathes of manufacturing industry, soon led to marked reductions in the emission of greenhouse gases and in pollution levels and to reports of a range of wildlife emerging into the public domain. However, it is unlikely that such environmental gains will be sustained if, or when, some sense of normality returns to the economy and to the hospitality industry. On the one hand, more optimistically, some commentators have emphasised the importance of recognising the environmental improvements generated by the COVID-19 crisis and the need to look to build strong structural environmental measures into any new sense of normality. On the other hand, there are concerns that the COVID-19 crisis has led to relaxations in environmental regulations and policies. In the USA, for example, fuel efficiency standards for cars have been reduced, and environmental monitoring and reporting standards have been relaxed; in Brazil, environmental monitoring has been scaled back in the Amazon, and, more generally, there are concerns about the growth of the carbon footprint associated with the tremendous increase in e-commerce during the crisis.\nConclusions and discussion: Conclusions\n",
"role": "user"
},
{
"content": "Purpose: This paper offers some reflections on changes in the relationships between sustainability and the hospitality industry following the onset of the Coronavirus Disease 2019 (COVID-19) crisis. (Covid-19 is officially a pandemic, but the term \"COVID-19 crisis\" is used throughout this paper because the authors feel that it captures the wider impacts of the crisis, rather than just focussing on the disease itself).\nDesign/methodology/approach: The paper describes the COVID-19 crisis, emphasises the role of hospitality in economic and social life and reviews how the crisis has changed the relationships between sustainability and the hospitality industry.\nFindings: The paper reveals the dramatic effect the crisis has had on sustainability in the hospitality industry. That said, though the crisis has offered a vision of a more sustainable future, this vision may pose a major challenge for the industry and for many of its traditional customers.\nResearch limitations/implications: The paper outlines some of the theoretical, operational, strategic and research implications of the crisis for the hospitality industry and for hospitality scholars.\nOriginality/value: This paper provides a reflective review of changes in the relationships between sustainability and the hospitality industry following the onset of COVID-19.\n",
"role": "assistant"
}
] |
Synthesizing moderate upscale restaurant patrons' dining experiences
|
[
"Structural equation modelling",
"Behavioural intentions",
"Comprehensive hierarchical modelling",
"Moderate upscale restaurants"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Malaysia's rising per capita income and increasing population, especially in urban areas, coupled with a more modern and busy lifestyle have created a greater demand for full service restaurants. Full service restaurants make a substantial contribution to Malaysia's food service sector (Agriculture and Agri-food Canada, 2014). Moderate upscale restaurants were established in Malaysia in the early 1990s, and many of these restaurants were foreign branded. Moderate upscale restaurants are now very popular in Malaysia and many international brands such as TGI Friday's, and the Hard Rock Cafe are common throughout the country (Euromonitor International, 2013).\nLiterature review, hypotheses development and conceptual research model: Service quality\nResearch method: A review of the relevant literature, focus group discussions, and an examination of the measurement items applied in previous studies provided the basis for the development of the closed-ended and self-administered questionnaire used in the data analysis. An extensive literature search generated the scale items used in the questionnaire and then the domain of the constructs were specified for the focus group discussions to refine the instrument (Churchill, 1979; Hair et al., 2010). Three focus groups were conducted to gain an in-depth knowledge of the research topic, refine the constructs and aid in developing the correct items used in the research instrument. In order to gain greater understanding and more insight in developing the questionnaire, three focus groups were conducted. The focus groups had the further role of providing a more in-depth knowledge of the service quality primary and sub-dimensions. Members of the focus groups were asked to identify and discuss the all of the factors that contributed to the satisfaction or dissatisfaction of their recent dining experiences at a moderate upscale restaurant. The members were provided with copies of the draft survey instrument and asked to comment on the relevance of the measurement items. The members were also asked to comment on the overall formatting of the instrument and the length of time they thought it would take to complete the survey. Two groups consisting of five Malaysian post graduate students were recruited from Lincoln University and the University of Canterbury and one group consisting of eight Malaysian professionals to discuss their dining experiences at moderate upscale restaurants in Malaysia. Minor amendments were made to the survey instrument as result of the focus group discussions.\nResults: In total, 546 questionnaires were returned and 11 questionnaires with missing data were excluded from further analyses. The 535 usable questionnaires were subjected to the preliminary screening requirements; the data satisfied the normality and linearity assumptions for all measured items. Non-response bias was assessed (Armstrong and Overton, 1977). Mean scores from the first 275 responses returned and the last 260 questionnaires returned were calculated. No statistically significant differences (p>0.005) were present between the two groups on the means for all of the constructs indicating the absence of non-response bias. In order to validate the EFA results and conduct the SEM analyses, the sample was randomly split into two data sets and each sample met the minimum size requirements (Kline, 2005; Hair et al., 2010). The profile of the sample is shown in Tables IV and V.\nDiscussion and implications: The ten sub-dimensions confirmed in this study differ in number and in content from those identified for other service industries. This result is supported by the findings in extant studies that have confirmed different dimensional structures and report that the precise dimensionality of service quality depends on the service industry under investigation and its cultural setting (Clemes, Brush and Collins, 2011; Dagger et al., 2007; Ryu et al., 2008). The ten sub-dimensions also differ in content and number from the sets identified in other studies on fast food restaurants (Aigbedo and Parameswaran, 2004; Brady and Cronin, 2001). The standardized b coefficients of the sub-dimensions also vary according to the perceived relative importance of the pertaining sets of sub-dimensions for each primary dimension (see Table VII).\nLimitations and recommendations: A relatively large response (n=535) was used in the data analysis. However, the cross-sectional sample was collected from patrons who had their lunch or dinner at four different moderate upscale restaurants in the Klang Valley. Generalising the results to other types of restaurants in Malaysia, or in other cultural settings, should be done with caution. Future research employing comprehensive hierarchical modelling should focus on the particular cultural setting and on the type of restaurant under investigation.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to develop and test a comprehensive hierarchical model of the interrelationships among five higher order marketing constructs (service quality, customer satisfaction, perceived value, restaurant image and behavioural intentions) for moderate upscale restaurants in Malaysia. A third order conceptualisation of service quality is also included in the empirical analysis.\nDesign/methodology/approach: The data were analysed using exploratory factor analysis, confirmatory factor analysis and structural equations.\nFindings: Service quality, customer satisfaction and restaurant image all have a strong, positive impact on the behavioural intentions of moderate upscale restaurant patrons. Service quality, perceived value and restaurant image are also important determinants of customer satisfaction. Behavioural intentions to re-patronise a restaurant and recommend it to friends are influenced more strongly by restaurant image than by customer satisfaction. Three primary dimensions and ten pertaining sub-dimensions are modelled in the third order conceptualisation of service quality. The primary and sub-dimensions also vary in importance amongst the customers in the sampling frame.\nOriginality/value: This is the first empirical research that develops and tests a comprehensive hierarchical model for moderate upscale restaurants to provide a complete and integrative analysis of a service setting. The interrelationships among service quality, customer satisfaction, perceived value restaurant image and behavioural intentions are assessed. A third order conceptualisations of service quality is also included in the modelling framework.\n",
"role": "assistant"
}
] |
A revised method for allocating the optimum number of similar machines to operators
|
[
"Queuing",
"Assignment",
"Binomial model",
"Machine interference problem",
"Numerical partition",
"Shortest path problem"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: There are many factories in the food, textile, plastic, electronic, and rubber industries where one operator runs several almost identical machines. In order to evaluate the productivity and performance of these production systems the average cycle time should be calculated. The cycle time length is the sum of the machine run time, the service time, and the interference time. Interference happens when an operator is servicing or otherwise tending a machine, causing other machines to stay idle, and wait until the operator is available again.\n2. Literature review: Managers of production systems are eager to maximize their profits by increasing yields, while at the same time reducing the cost of production (Tirkel and Rabinowitz, 2014). One of the key methods used to accomplish those aims is to establish an optimum number of operators needed to run the production lines. This paper proposes a method to help management determine the optimal number of operators to be assigned to a given number of machines, and the number of machines that should be run (or controlled, or operated) by each operator (a numerical partition). This should be done with the objective of minimizing production costs or maximizing profits. However, the number of operators that should be employed for minimizing production cost per unit is not necessarily the same number that is needed for maximizing profit. This will be illustrated later in this paper by an actual case study.\n3. Model development and the adjusted cycle time: This section presents a binomial machine interference model and its assumptions.\n4. The optimum number of machines to be assigned to the operators: This section describes the stages to be taken in order to assign the optimum number of machines to each operator, in order to minimize the TCU or to maximize the hourly profit. TCU includes operator cost, machinery ownership cost, machinery operating cost, and material cost (see Equation (5)). Hourly profit is computed by determining the difference between the selling price of a unit and its TCU, then multiplying that figure by the number of units produced in one hour by one operator and N machines (see Equation (7)). This section also determines how to divide a given number of M machines among operators, that is, to specify how many operators should be employed in order to operate M machines.\n5. A real-life case study: A factory for making plastic products has M=23 identical machines. The machines need service that is given by operators when random breakdowns occur. These machines are currently run by four operators, three of whom operate six machines while the fourth operates five machines. The purpose of the case study is to determine the optimum number of operators K*, and how to divide the machines among the operators, in order to reach the objectives of minimizing the average TCU (Equation 48), or maximizing the hourly profit. (Equation 49). A work study finds out that the run time is T=12.30 minutes and the average service time is t=1.03 minutes. It was found that the hourly fixed cost is C\n6. Conclusions: This paper presents a unique method for measuring the performance and productivity of several operators operating several identical machines. This is the first method that enables determination of the period of time needed to produce one unit (cycle time), given the number of machines that each operator runs. The proposed method also enables practitioners to determine the number of operators to be assigned to a given number of machines and the number of machines that will be run by each operator (a numerical partition). This should be done with the objective of minimizing production costs, maximizing profits, increasing productivity, setting a given load on the operators (e.g. 85 percent), or setting other performance measures. Moreover, the method enables calculating various performance indexes such as machine utilization, workload, and yield per hour.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to propose a method to determine the optimal number of operators to be assigned to a given number of machines, as well as the number of machines that will be run by each operator (a numerical partition). This determination should be made with the objective of minimizing production costs or maximizing profits.\nDesign/methodology/approach: The method calculates the machines interference rate via the binomial distribution function. The optimal assignment is calculated by transformation of a partition problem into a problem of finding the shortest path on a directed acyclic graph.\nFindings: The method enables the authors to calculate the adjusted cycle time, the workload of the operators, and the utility of the machines, as well as the production yield, the total cost per unit, and the hourly profit for each potential assignment of operators to machines. In a case study, the deviation of the output per hour of the proposed method from the actual value was about 2 percent.\nPractical implications: The paper provides formulas and tables that give machine interference rates through the application of binomial distribution. The practicability of the proposed method is demonstrated by a real-life case study.\nOriginality/value: The method can be applied in a wide variety of manufacturing systems that use many identical machines. This includes tire presses in tire manufacturing operations, ovens in pastry manufacturing systems, textile machines, and so on.\n",
"role": "assistant"
}
] |
Age-specific survival in prostate cancer using machine learning
|
[
"Prostate cancer",
"Metastasis",
"Medical",
"Machine learning",
"Survival prediction",
"Data mining",
"Ensemble"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: From the past decades, cancer is one of the most leading causes of death worldwide. According to GLOBOCAN (2018), there were 18.1 million new cases of cancer and 9.6 million deaths due to cancer in 2018. Prostate cancer is the second most occurring cancer in men in both developing and developed countries, while it accounts for 1.3 million new cases, the fourth most commonly diagnosed cancer overall (GLOBOCAN, 2018). It more commonly occurs in developed countries like North America, Australia, New Zealand and Western and Northern Europe (Risk Factors, 2017). In developing countries like India also, the incidence rate of prostate cancer is increasing from past years, and more commonly in the urban population (Hariharan and Padmanabha, 2016).Machine learning techniques have been used widely in the past decades by researchers for diagnosis (Alickovic and Subasi, 2015; Wang and Wang, 2017), survival prediction (Jajroudi et al., 2014; Park et al., 2013; Cheng-Min Chao et al., 2014; Walczak and Velanovich, 2018) and recurrence prediction (Andjelkovic Cirkovic et al., 2015; Tseng et al., 2017) in different cancers. In this study, we focus only on the survivability of cancer. Survival is difficult to be estimated due to various environmental, genetic and biological factors. The survival of the patients after diagnosed with cancer is essential for making an informed decision to the patient, as well as for determining the best treatment possible for the patient. Due to the expensive treatment procedures for cancer, survival prediction is essential. Earlier, clinicians used to estimate the survivability of the patient based on their experience. This estimation may be biased due to their limited experience. The prediction modeling techniques have proved to perform better in survival prediction than clinical opinions (Ross et al., 2002) (Walz et al., 2007).Data mining has played an essential role in various areas of medical research. Cancer research is one such area. The availability of some online databases has allowed the researchers to create models that assist the clinicians in better treatments for the patients. SEER database (SEER database) is one such example of an online database that covers the maximum number of cancers and patients. However, some studies have used the hospital-based datasets also to analyze the research and predict the survival on a regional basis (Tseng et al., 2015). Also, not all types of cancers are entirely covered in online databases. Bray et al. (2018) also analyzed the incidence of thirty-six cancers in different countries, and the researchers do not explore most of them due to the nonavailability of data.There have been many studies that analyze the survival of prostate cancer based on various factors like treatment (Teoh et al., 2018), Gleason score (Rusthoven et al., 2014), metastasis (Pond et al., 2014), but machine learning has not been explored much in prostate cancer survival. The Kaplan-Meir and Cox proportional regression methods are the traditional statistical techniques for the survival analysis, but machine learning techniques give an insight into the survival, considering the various attributes of the patients. It does so by first training the model using a part of the dataset and then validating and testing the classifier built from the training. Nezhad et al. (2019) applied deep learning combined with active learning for the survival prediction in prostate cancer, where the data was collected from the SEER database. The proposed approach by the authors gave better results than other baseline machine learning algorithms.This study is an attempt to predict the survivability in metastatic prostate cancer, keeping in mind the age group to which a patient belongs. The TNM staging is determined for the patient after being diagnosed with prostate cancer. The metastatic stage is the advanced stage of prostate cancer, where the cancer cells spread to distant organs. Although a very small percentage of the men diagnosed with prostate cancer are metastatic, the survival in the metastatic stage is essential for the patient, as well as the oncologists, as the survival rate is meager in this stage. The 5-year survival rate is almost 30% for Stage 4 prostate cancer patients (SEER database). But the SEER dataset covers only American patients, and the survival rate is higher in Europe and North America, and comparatively less in Asian and African countries (Bray et al., 2018).It has been proved in the previous studies that age has a significant role in the survival of the patient (Moreira et al., 2017). This study tries to use the fact and thus divide the dataset into clusters of patients sharing the same age group. The oncologists also tend to consider age as a significant factor while deciding the appropriate treatment a patient can bear. Prostate cancer is also more common in older patients. Not only incidence, but the mortality rate also depends on age in almost every disease. Thus, we have used age as a deciding factor, instead of just an attribute for the survival prediction. The oncologists can then evaluate the survival of a patient based on the age group that a patient belongs to. This study can be further extended to different ailments where age can play a crucial part in the survivability of the patients.The rest of the paper is organized as follows: Section 2 discusses the related works done by the researchers. Section 3 explains the dataset and methodology used in the study, while the results are presented in section 4. Last, the conclusion of the study is given in section 5.\n2. Related work: As already discussed in the previous section, there have not been many studies exploring the survival prediction in prostate cancer. The literature survey of the research thus covers two aspects: survival prediction using machine learning and prostate cancer survival analysis. 2.1 Survival prediction using machine learning\n3. Prostate cancer: According to GLOBOCAN, prostate cancer is among the top ten leading cancers in India (Bray et al., 2018). The cancer data shows that the incidence of prostate cancer will be doubled by 2020. It generally occurs in aged men above 65 years of age (Hariharan and Padmanabha, 2016). According to SEER 2012-2016, the median age at diagnosis, data is 66 (SEER database). Prostate cancer is more prevalent in developed countries than in developing countries. In the USA, it is the second occurring cancer in men (Jemal et al., 2009). Although it has a high incidence in developed countries, the survival rate is lower in developing countries due to delayed diagnosis. Geography, drinking and smoking history, family history and race are major risk factors for prostate cancer, apart from genetic changes (Sadeghi-Gandomani et al., 2017). Obesity is also a contradictory risk factor for cancer, but the results are not consistent for all the studies.Prostate cancer is staged using the TNM system. Table 2 gives a detailed analysis of the prostate cancer stages. Stage 1-2 is mostly confined and has higher survivability, while cancer cells in stage 3 spread to neighboring body parts. The metastatic stage refers to the one where the cancer cells have grown outside the particular organ (prostate in this case) and advanced to other body parts like bones or some other organs via blood vessels or lymph vessels. Stage 3 and 4 cancer patients have relatively lower survival than the earlier stages. PSA (Prostate-Specific Antigen) is the most common test to diagnose prostate cancer. PSA levels higher than 4ng/ml are not considered standard. However, just PSA results do not give any conclusive results for cancer. If the PSA levels show abnormal values, a biopsy is generally the next step, which may confirm the presence or absence of prostate cancer. The prostate cancer has a better prognosis for localized cancer and with a lower grade. The grade is defined by the Gleason score, which tells the abnormality and pace of the cancer cells in the body.\n4. Materials and methods: 4.1 Dataset\n5. Results and discussion: 5.1 Analysis of each age group\n6. Conclusion and future work: Metastatic prostate cancer has a very low survival rate, compared to the early stages. The survival information of the patients is important for the patients, as well as the clinicians, to make an informed decision. In this study, the survival of metastatic prostate cancer patients has been estimated concerning the age group to which a patient belongs. Data of 407 patients who were diagnosed with metastatic prostate cancer between the years 2011-2015 was collected from a cancer hospital. The dataset was divided into five clusters based on their age group, and the survival prediction was compared with the complete dataset. Ensemble bagged trees achieving the highest accuracy in the case of the combined dataset was used for further analysis. It was found that the age-specific model gave better overall accuracy than the combined model. The study also explored the oversampling measure for the age group with >= 80 age, and the sampled group showed better results than the original group. Also, the information gain of various attributes was analyzed for the models, and it was proved that the sequence of treatments given to the patients plays a significant role in the survival of the patients in any of the models. The performance status of the patients also proved to be good predictors for all the models.In the future, we will apply some feature selection techniques to reduce the number of features set for the prediction and better survival results. We will also try to improve the outcomes for 60-79 age groups, which present the larger population of the patients who have prostate cancer. Also, the sequence of the treatments will be considered for the survival prediction of prostate cancer patients. This may be achieved by applying sequence mining techniques or time-series framework of deep learning and generating some frequent sequence of medications given to the patients and analyzing their survivability.\n",
"role": "user"
},
{
"content": "Purpose: The incidence of prostate cancer is increasing from the past few decades. Various studies have tried to determine the survival of patients, but metastatic prostate cancer is still not extensively explored. The survival rate of metastatic prostate cancer is very less compared to the earlier stages. The study aims to investigate the survivability of metastatic prostate cancer based on the age group to which a patient belongs, and the difference between the significance of the attributes for different age groups.\nDesign/methodology/approach: Data of metastatic prostate cancer patients was collected from a cancer hospital in India. Two predictive models were built for the analysis-one for the complete dataset, and the other for separate age groups. Machine learning was applied to both the models and their accuracies were compared for the analysis. Also, information gain for each model has been evaluated to determine the significant predictors for each age group.\nFindings: The ensemble approach gave the best results of 81.4% for the complete dataset, and thus was used for the age-specific models. The results concluded that the age-specific model had the direct average accuracy of 83.74% and weighted average accuracy of 79.9%, with the highest accuracy levels for age less than 60.\nOriginality/value: The study developed a model that predicts the survival of metastatic prostate cancer based on age. The study will be able to assist the clinicians in determining the best course of treatment for each patient based on ECOG, age and comorbidities.\n",
"role": "assistant"
}
] |
Support, training readiness and learning motivation in determining intention to transfer
|
[
"Organizational support",
"Training",
"Supervisor support",
"Learning motivation",
"Training readiness",
"Intention to transfer"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Training is a strategic tool for organizations to improve employees' knowledge, skills, and attitudes as part of the critical function of human capital management. To gain a competitive advantage, organizations spend billions of dollars on training programs every year with the expectation that there will be a returns on their investment with improved organizational performance (Beer et al., 2016; Kirwan and Birchall, 2006; Velada et al., 2009).\nLiterature review and hypotheses: Intention to transfer\nMethods: Sample and research context\nResults: Descriptive statistics and correlations\nDiscussion: This study examined the associations among transfer-enhancing support, employees' personal characteristics, and intention to transfer. The results of this study showed the role of the proximal and distant support for training transfer from organizations on employees' intention to transfer. Specifically, the findings of the current study also provide evidence that:\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this study is to augment knowledge of how work environment and personal characteristics affect intention to transfer in a work context. This study aims to investigate the factors that can influence intention to transfer training in a professional development training context. The study examined the predictive capacity of organizational support, supervisor support, training readiness and learning motivation on transfer intention among the study respondents.\nDesign/methodology/approach: Data were collected from teachers in secondary schools in the USA. The structural equation modeling method was adopted to analyze 216 responses using a self-report survey.\nFindings: We found that (a) organizational support was directly associated with supervisor support for training, (b) supervisor support for training significantly influenced training readiness and motivation to learn, (c) training readiness positively affected motivation to learn, and (d) motivation to learn positively influenced intention to transfer. In addition, supervisor support mediated the relationships between organizational support and training readiness and between organizational support and motivation to learn. Training readiness linked intention to transfer and motivation to learn. Motivation to learn also played a mediating role in the relationship between supervisor support and intention to transfer.\nOriginality/value: Our findings add to the academic work on training transfer by empirically analyzing how both the environment (e.g. organizational support) and individual factors (e.g. learning motivation) influence employees' intention to transfer. In particular, we investigated the potential impact of both organizational support and supervisory support on intentions to transfer, compared to previous studies emphasizing only supervisory support to improve training outcomes.\n",
"role": "assistant"
}
] |
Underpricing and market timing in SEOs of European REITs and REOCs
|
[
"Europe",
"Market timing",
"Underpricing",
"SEO",
"Equity offering",
"Offer price discount"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: As stated by Ghosh et al. (2000), seasoned equity offerings (SEOs) are major events in the lifetime of a listed property company that are essential to ensure profitable growth and sustainable development. Especially, real estate investment trusts (REITs) need to regularly raise money through capital increases and therefore access the capital market more often than industrial firms due to their limited funding possibilities with restriction on debt ratio and retained earnings (Boudry et al., 2011). Although SEOs play an important role in listed companies, issued shares are regularly offered at an offer price significantly lower than the price the shares are traded on the offer day - defined as underpricing - or the day before the offering - defined as discounting. Besides the direct cost of raising capital, the issuing firm is consequently accepting additional expenses as \"money left on the table\" at equity offerings. Because the extent of underpricing represents potential equity capital, the issuer decides to forego convincing arguments that must exist to justify such losses.\n2. Listed real estate markets in Europe: REITs have shown to be an attractive investment alternative in the capital market in the past decades (Laopodis, 2009), and they have gained increasing significance especially in Europe (Newell and Marzuki, 2018). Although several countries in Europe adopted the REIT regime in the late 2000s, emerging real estate markets in southern and eastern Europe are still into extending their spectrum of real estate investment vehicles, with Poland and Portugal being the most recent to establish the REIT regime in 2019. Whereas in 1999, only one REIT existed in Europe, the value of listed real estate in the European Union reached $387.6bn, with 226 REITs accounting for $186.9bn in market capitalization (EPRA, 2019b). Table I provides an overview of listed real estate markets in Europe. Regarding total market capitalization of property companies by country, Germany (EUR92.1bn), the UK (EUR77.0bn) and France (EUR55.5bn) have shown to be the largest listed markets in Europe. Whereas Germany is characterized by REOCs (e.g. Vonovia with EUR22.8bn and Deutsche Wohnen with 11.5bn) and further investment vehicles like open- and closed-end funds with only a small proportion being invested in REITs, the UK REIT market is highly matured, comprising more than 80 percent of UK's total market capitalization of listed real estate companies. The same appears for France, where REITs account for EUR51.0bn out of EUR55.5bn. Whereas developed listed real estate markets like Sweden and Switzerland (EUR53.7bn and EUR46.0bn) have not adopted the REIT status yet, REITs play an important role in Belgium, the Netherlands and Spain, with high proportions being invested in the tax-exempt investment vehicle.\n3. Theoretical background: Different theories have evolved to understand the phenomenon of setting the offer price at a discount at equity offerings.\n4. Literature review: Underpricing in SEOs\n5. Data and summary statistics: Sample design\n6. Empirical results: Methodology, model specifications and regression results\n7. Conclusion: As stated by Ghosh et al. (2000), SEOs are major events in the lifetime of a property company that are essential to ensure profitable growth and sustainable development. Although REITs need to regularly raise money through capital increases and therefore access the capital market more often, the issuing firms decides to forego potential equity capital by setting the offer price significantly lower than the price the shares are traded on the offer day - defined as underpricing. With a sample of 470 SEOs of European REITs and REOCs from 2004 to 2018, this study contributes to the literature on SEO pricing in several ways.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to contribute to the literature on seasoned equity offerings (SEOs) by examining the underpricing of European real estate corporations and identifying determinants explaining the phenomenon of setting the offer price at a discount at SEOs.\nDesign/methodology/approach: With a sample of 470 SEOs of European real estate investment trusts (REITs) and real estate operating companies (REOCs) from 2004 to 2018, multivariate regression models are applied to test for theories on the pricing of SEOs. This paper furthermore tests for differences in underpricing for REITs and REOCs as well as specialized and diversified property companies.\nFindings: Significant underpricing of 3.06 percent is found, with REITs (1.90 percent) being statistically less underpriced than REOCs (5.08 percent). The findings support the market timing theory by showing that managers trying to time the equity market gain from lower underpricing. Furthermore, underwritten offerings are more underpriced to reduce the risk of the arranging bank, but top-tier underwriters are able to reduce offer price discounts by being more successful in attracting investors. The results cannot support the value uncertainty hypothesis, but they are in line with placement cost stories. In addition, specialized property companies are subject to lower underpricing.\nPractical implications: An optimal issuance strategy taking into account timing, relative offer size and the choice of the underwriter can minimize the amount of \"money left on the table\" and therefore contribute to the lower cost of raising capital.\nOriginality/value: This is the first study to investigate SEO underpricing for European real estate corporations, pricing differences of REITs and REOCs in seasoned offerings and the effect of market timing on the pricing of SEOs.\n",
"role": "assistant"
}
] |
Development of an integrated real time dispatching system: A case study at a semiconductor assembly and test factory
|
[
"Integration",
"Computer aided production management",
"Dynamic scheduling",
"Real time dispatching",
"Production scheduling"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In today's competitive and dynamic business environment, semiconductor companies as high-tech concentrated firms face heightening customer expectations. They need evaluate the effectiveness of their supply chain, shrink product inventory level, increase demand for product varieties. As one of the most complicated manufacturing systems, the scheduling problems of semiconductor manufacturing system (SMS) have attracted much attention in recent years (Hassoun et al., 2008). Multiple performance objectives of production scheduling are developed continuously, such as adaptive dispatching strategies, cycle time, on-time delivery rate (ODR), work in process (WIP) level, demand fulfillment, etc. (Lee et al., 2009). A qualified scheduling approach should provide the holistic management of the healthy WIP level, demand fulfillment rate, entities effective utilization based on the real time dispatching management.\n2. Literature review: Semiconductor manufacturing contains hundreds of jobs need to be scheduled, dispatched, and delivered in an automated environment. In recent years, several real-time scheduling approaches have been developed to investigate the optimal production scheduling and demand control policies. Zhang et al. (2006) presented EOPN-based real-time scheduling simulation platform for wafer fab system with object-oriented technology. Kumar and Nottestad (2006) developed a simulation model of an integrated short-term production scheduling problem at the manufacturing operation. Their key initiative was to develop the scheduling application that could react quickly to the ever-changing manufacturing floor. Liu et al. (2007) developed an infrastructure of the timed EOPNs-based multiple-objective real-time scheduling system to tackle complex 300 mm wafer fabs. Four specific performance objectives pursued by contemporary semiconductor manufacturers were integrated into a priority-ranking algorithm, and then all wafer lots would be dynamically dispatched by the real-time state-dependent dispatching system. Sabuncuoglu and Kizilisik (2003) studied the reactive scheduling problems in a dynamic and stochastic manufacturing environment. A simulation based scheduling system was developed and several reactive scheduling policies were proposed under various experimental conditions. The experiments showed that the online scheduling was more robust to uncertainty and variations in processing times than the optimum seeking offline scheduling. Ellis et al. (2004) focussed on solving a common wafer test scheduling problem in semiconductor manufacturing. Four heuristics were developed to solve the test scheduling problem with the aim of minimizing the make span required to test all wafers on a set of test stations. In the case study, the proposed solution approaches decreased the make span by 23-45 percent compared with the make span of the actual schedule executed in the manufacturing facility. Wang et al. (2005) developed a hybrid knowledge discovery model for selecting dispatching rules of a semiconductor final testing factory. A combination of a decision tree and a back-propagation neural network were employed to determine an appropriate dispatching rule and predict its performance. The experiment had shown that proposed decision tree could find the most suitable dispatching rule. Wang et al. (2008) proposed a real time distributed shop floor scheduling framework by using an agent-based service-oriented architecture. The distributed scheduling was conducted through web services facilitated by the service-oriented shop floor integration. A case study was conducted to validate the presented dynamic distributed scheduling mechanisms accordingly. Jia and Mason (2009) investigated the identical parallel machine multiple orders per job scheduling problem to minimize total weighted order completion time in a 300 mm semiconductor manufacturing operations. A number of polynomial-time heuristic approaches were proposed to solve the NP-hard problems. Experimental results identified appropriate heuristic techniques for analyzing the practically motivated scheduling problem. Ham et al. (2010) studied the real-time scheduling problem of a multi-stage flexible job shop floor, the binary integer programming model was formulated with a real time scheduling decision for the flexible job shop scheduling problem. A real-time scheduling procedure was presented to generate a schedule with competitive quality by using IP. Tsai and Huang (2010) constructed a real-time scheduling and rescheduling system based on radio frequency identification information for fully automated semiconductor fabs. Three modules were embedded into this system, such as a real-time decision-making scheduler, a rescheduling mechanism for machine breakdowns and bottleneck, and a simulation platform. From the simulation results, the problems of machine breakdowns and bottleneck were well solved through the proposed system and method. Min and Yih (2010) developed a real-time multi-objective scheduler for the selection of decision rules for decision variables in order to obtain the desired performance measures at a certain production interval. Accordingly, a system control strategy based on a simulation technique and a competitive neural network was suggested. Sun et al. (2011) developed a composite allocation rule-based decision policies to allocate the products lots for customer order fulfillment in semiconductor supply chains. Different lot allocation decision policies were evaluated and compared using representative data sets. Computational test results indicated the presented policies could improve the quality of the solution and reduce the cost of the solution significantly. Phanden et al. (2012) introduced an approach to integrate the process planning and scheduling. The presented approach could quickly integrate these two functions with process plan selection module, scheduling module, schedule analysis module, and process plan modification module. The experiments showed that it was easily implemented in a company with existing process planning and scheduling, and achieved a better performance than hierarchical approach.\n3. The case company introduction: 3.1 Company profile\n4. Background to IRTD development in S company: 4.1 Problem faced by S company\n5. IRTD system design: 5.1 Integrated model design\n6. DOI-based integrated dispatching algorithm: The proposed IRTD system for semiconductor ATM factory is designed to realize real-time dispatching for the complex multiple-objective targets. The IRTD system that depends on the heterarchical real-time scheduling control may incorporate different kinds of customized dispatching policies. A DOI-based dispatching algorithm is designed to integrate the multiple indicators of dynamic scheduling and ordering in the system to generate the \"optimal\" dispatching policies. According to the algorithm, the \"right\" raw materials or product segments will be dispatched out from TRDI or SFGI to fulfill true order. The proposed DOI-based algorithm aims to achieve the reducing inventory costs, optimize ODR simultaneously.\n7. IRTD platform application: IRTD platform employs the Microsoft Distributed Component Object Model (DCOM) and client/server (CS) mode which provides a user-friendly interface and debugging environment. The development of the main module is applied common object request broker architecture (CORBA) for developing the IRTD system architecture. CORBA is an open standard that provides comparable services for building distributed CS infrastructures. Numerous implementations of CORBA are programmed by Java. It can be widely available across many platforms, such as Microsoft-based systems and UNIX environments.\n8. Experiment results analysis: The infrastructure of IRTD has been specified detailed in previous sections, and the DOI-based integrated dispatching algorithm is proposed to facilitate the priority dispatching procedure. To verify the effectiveness of the proposed module and algorithm, the simulation is conducted in six continuous weeks. Several performances of scheduling and ordering are compared accordingly. Some assumptions are established in the simulation platform:\n9. Conclusion: Dynamic production scheduling and real time dispatching approaches for modern SMS should possess the following characteristics: combinational performance objectives, which includes the healthy WIP level, ODR, highly equipment utilization; real time dispatching and control, adaptive order fulfillment; software compatibility, supporting software implementation and easy modification and expansibility. Thus, there is a need for a system with real time dispatching functionality.\n",
"role": "user"
},
{
"content": "Purpose: In a complex semiconductor manufacturing system (SMS) environment, the implementation of dynamic production scheduling and dispatching strategies is critical for SMS distributed collaborative manufacturing events to make quick and correct decisions. The purpose of this paper is to assist manufacturers in achieving the real time dispatching and obtaining integrated optimization for shop floor production scheduling.\nDesign/methodology/approach: In this paper, an integrated model is designed under assemble to order environment and a framework of a real time dispatching (IRTD) system for production scheduling control is presented accordingly. Both of the scheduling and ordering performances are integrated into the days of inventory based dispatching algorithm, which can deal with the multiple indicators of dynamic scheduling and ordering in this system to generate the \"optimal\" dispatching policies. Subsequently, the platform of IRTD system is realized with four modules function embedded.\nFindings: The proposed IRTD system is designed to compare the previous constant work in process method in the experiment, which shows the better performance achievement of the IRTD system for shop floor production dynamic scheduling and order control. The presented framework and algorithm can facilitate real time dispatching information integration to obtain performance metrics in terms of reliability, availability, and maintainability.\nResearch limitations/implications: The presented system can be further developed to generic factory manufacturing with the presented logic and architecture proliferation.\nOriginality/value: The IRTD system can integrate the real time customer demand and work in process information, based on which manufacturers can make correct and timely decisions in solving dispatching strategies and ordering selection within an integrated information system.\n",
"role": "assistant"
}
] |
The many aspects of anticipation
|
[
"Forecasting",
"Complexity theory",
"Philosophical concepts"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Anticipation as a shift of paradigm?: The mainstream wisdom claims that causes move things forward. As solid and reassuring as this statement may be, it nevertheless runs into trouble as soon as phenomena of self-organization or network causality are taken into account. What kinds of causality are they? Matters become even worse when the emergence of hierarchies - i.e. levels of organization - are considered. Even if hierarchies may emerge from the bottom up, the higher levels usually exert some kind of top-down constraining influence on the lower levels of the hierarchy. To say the least, downward causation is far from being part of the received wisdom. The hierarchical loops emerging from the cycles of up and down causations between hierarchical levels are even farther away from the mainstream. When hierarchies further assume the form of different, possibly tangled, levels of reality between different types of entities - atoms, molecules, organisms, minds and societies - it is obvious that something important has been missed by mainstream theories of causation (Poli, 2001, 2007).In this paper I shall discuss anticipation, which possibly the most important aspect missing from mainstream theories of causation. Behaving in an anticipatory way means adjusting present behavior in order to address future problems. In other words, an anticipatory entity (system or whatever) takes its decisions in the present according to forecasts about something that may eventually happen.The best-known definition of anticipation is Rosen's: \"An anticipatory system is a system containing a predictive model of itself and/or its environment, which allows it to change state at an instant in accord with the model's predictions pertaining to a later instant\" (Rosen, 1985, p. 341).The most obvious mistake committed by almost everyone when first confronted with the idea of anticipation is to think that anticipation is a feature that we possess because we are such highly complex and wonderfully sophisticated cognitive agents. This is not what the theory of anticipation claims. Indeed, the major surprise embedded in the theory of anticipation is that anticipation is a widespread phenomenon present in and characterizing all types of realities. Life in all its varieties is anticipatory, the brain works in an anticipatory way, the mind is obviously anticipatory, society and its structures are anticipatory, even non-living or non-biological systems can be anticipatory. All this comes as more than a surprise.If all this is true, and providing that the necessary supporting evidence can be accumulated, it implies that a proper understanding of anticipation requires the adoption of an innovative conceptual framework. Moreover, this new framework will have to be innovative in many different ways, some of which will be mentioned by this paper.\n2. A minimal overview of theories of anticipation: As soon as one starts collecting data on anticipation, the first unexpected surprise is the finding that over the past century many scholars from many different disciplines and fields have worked on anticipation. The unwelcome result is that nobody has to date collected and compared the various proposals. It may well be that the same phenomenon has been discovered times and again. Even so, it would be interesting to know the differences, if any, among the theories. It may be that different scholars have seen different aspects of anticipation, and a thoroughgoing comparison between them may help develop a more rounded-out theory. The following notes are nothing more than a preliminary and still rather idiosyncratic survey. Much more systematic work must be done.2.1 Philosophy\n3. The phenomenon of anticipation versus the case of anticipatory systems: After the rapid overview conducted in the previous section, it is now time to delve deeper into the nature of anticipation. I shall start by distinguishing the capacity of anticipation from the nature of systems able to exhibit anticipatory behavior (i.e. anticipatory systems). I contend that studying anticipation as a capacity is very different from studying anticipatory systems. The former endeavor adopts a descriptive attitude, conducts experiments and collects data, whilst the latter seeks to understand what it is that makes anticipation possible.Clear understanding of the difference between anticipation as a property that some class of systems exhibits, on the one hand, from the internal structure that a system should possess so that it can behave in an anticipatory fashion, on the other, requires anticipation to be distinguished from its external modeling or eventual simulation. The difference between internal and external modeling is relevant here. External models seek to capture the outcome of a specific system. Given a system S characterized by behavior B, the system's model tries to find an algorithm producing values as close as possible to behavior B. External models are only interested in the equivalence between their output and the system's behavior; they do not address the problem of how the system actually produces its own behavior. Internal models, on the contrary, do consider the internal working of the system. An internal model endeavors to capture the system itself, or better its internal determinants. We shall see that anticipatory systems are self-referential or impredicative systems. They are therefore such that no algorithm can ever model their behavior.Two remarks are in order:1. the fact that no algorithm is impredicative does not entail that we must give up formal modeling; the theory of rote formal machinery is but one part of contemporary mathematics: however important it may be, it is still only a part; and2. although it is impossible to completely model an impredicative system within an algorithmic framework, this does not rule out the possibility of \"freezing\" the system and of modeling the specific configuration and behavior of the \"frozen\" system.One can also break the system's evolution down into a number of main steps, model them, and ask which properties a transition functor should have in order to model the system's passage from phase n to phase n+1 (two recent studies have started to work out the relevant details: the idea of variable topology by Brown et al. (2007), and memory evolutive systems by Ehresmann and Vanbremeersch (2007).\n4. The phenomenon of anticipation: Anticipation comes in many different guises. The simplest distinction is between explicit and implicit anticipation. Explicit anticipations are those of which the system is aware. They may be used as synonyms for predictions or expectations. Implicit anticipations, by contrast, work below the threshold of consciousness. They may be active within the system without the system itself being aware of them. Implicit anticipations are properties of the system, intrinsic to its functioning.As far as explicit anticipation is concerned, the reflexive side of explicit anticipation becomes visible as the difference between looking into the future and taking account of the consequences of that looking, i.e. as the impact of an anticipation on current behavior. The types or aspects of behavior that can be modified through anticipation are exemplifications of normative behavior. If the system evaluates its own evolution as positive (according to its own criteria), it will maintain its behavioral patterns; conversely, if the system evaluates its own evolution as negative, it may seek to change its behavioral patterns in order to prevent the occurrence of the anticipated negative results.From an evolutionary point of view, explicit anticipation is an advantage because it enables more rapid goal-directed processing. The other side of the coin, however, is that focused goal-oriented behavior usually gives rise to inattentional blindness, i.e. the incapacity to perceive things that are in plain sight (Simons and Chabris, 1999). Patterns constrain attention, govern the boundary of relevance, and they direct attention to pre-established foci. The more efficient the patterns, the more likely is the outcome of an over-restricted focus of attention. The more efficient the behavioral patterns are, the more rigid they become.Anticipation exhibits a variety of temporal patterns, from microanticipations embedded in perception to usually longer forms of social anticipation, ranging from seconds to years and decades.A major question is whether explicit anticipations depend - or to what extent they may depend - on implicit ones. As Riegler (2003, p. 11) aptly asks: \"Are we consciously creating anticipations on basis of which we plan and make decisions, or are anticipations and decisions making made for us?\". Finding the correct answer to this question is far from being a trivial undertaking. The apparently obvious answer that explicit anticipation depends - at least to some extent - on implicit anticipation may beg the question. For it may well turn out that the two forms of anticipation are based either on entirely different enabling conditions, or on different subsystems, or even on entirely different systems. This last case makes sense as soon as one envisages a system composed of different systems (as opposed to the more conventional idea of a system composed of functional subsystems).None of the best-known theories has yet explicitly addressed the problem that systemic behavior may be the result of processes unfolding at different levels of reality, including the biological level (perception, brain processes), the psychological level (cognitive processes), and the social level (social interactions) (for the theory of levels of reality, see Poli, 2001, 2006, 2007).None of the above mentioned questions can be answered without moving from anticipation as a detectable behavior to anticipatory systems.\n5. Anticipatory systems: With the remarkable exception of Rosen's work, the great majority of the studies conducted to date address the problem of anticipation within some specific field of inquiry. Only rarely have scholars discussed the problem of the features which a system should possess so that it can behave in an anticipatory fashion.Analysis of the enabling features of anticipation can be conducted at two different levels: firstly by considering the set of controllers a system may possess, and secondly by asking what abstract features a system must possess in order to be an anticipatory system.5.1 Controllers\n6. Higher-order complexity: Complexity, as usually understood, refers to chaotic systems, i.e. to systems that are deterministic and sensitive to their initial conditions. Thus understood, complex systems are entirely past-governed. For obvious reasons, mainstream complexity is utterly unable to address the challenge raised by anticipatory systems.The concept of super-complexity has been introduced in order to distinguish anticipatory systems from entirely past-governed ones (Baianu and Poli, 2009; Poli, 2009).The rather technical distinction among simple, complex and super complex systems can be approached by considering the possibly simpler distinction among different types of wholes:* Decomposable wholes (simple and chaotic systems). The system can be fragmented into constituent parts without losing relevant information. This implies that the system can be reconstructed from its parts. Similarly, the dynamics of the system can be reconstructed from the dynamics of its parts. Mechanisms are cases in point. Analytic, algorithmic methods work well with wholes of this type.* Quasi-decomposable wholes (complex systems). Fragmentation into constituent parts loses information, irreversibly. Individual cases can still be modeled, even if in an ad-hoc way (the most famous case being represented by the three-body problem). Universal analytic methods are not available for these cases. It is in this that most engineering knowledge consists.* Indecomposable wholes (super complex systems). These wholes are irreducible: their fragmentation loses information. Analytic methods fail to work even for individual cases. Since indecomposable wholes are not (entirely) understandable from their parts, the manipulation of parts may produce unexpected consequences.Anticipatory systems are indecomposable wholes in the sense specified. They are indecomposable because of the hierarchical loop(s) that they contain. The difference may eventually be assumed between properly super-complex systems, some of whose functions are mutually implied, and hyper-complex systems, all of whose functions are mutually implied (as for organisms). One may then claim that super- and hyper-complex systems are the two cases of higher-order complexity theory.\n7. Conclusion: Even if most of the details concerning anticipation and anticipatory systems are still unknown, the following partial observations summarize the current state of knowledge.* Anticipation comes in different guises, such as explicit as opposed to implicit anticipation. Furthermore, different types of anticipation may be at work contemporaneously, such as biological, psychological and social types of anticipation. They have their own temporal patterns, and may be distinguished by other properties as well. Moreover, when different types of anticipation are simultaneously active, they may work harmoniously together or they may interfere with each other.* Anticipation has been a major evolutionary discovery. If Rosen is correct, anticipation does not necessarily depend on previous memories. This amounts to saying that anticipation lies deeper than memory in the functional structure of organisms. It apparently also applies to both psychological and social systems. This is an important contention which requires firmer supporting evidence.* The abstract structure of anticipation depends on hierarchical loops, also known as self-referential, or impredicative loops. This imposes very severe constraints on the modeling of anticipation systems, because, almost by definition, mechanical or algorithmic procedures cannot be self-referential.Anticipation and anticipatory systems are one of the challenges that science is starting to address. At this stage of development, it is simply impossible to know where anticipation will take us. What anticipation shows, however, is that reality has still many surprises in store.\n",
"role": "user"
},
{
"content": "Purpose: - After summarizing the theories of anticipation proposed over the past century, the paper aims to distinguish between anticipation as an empirical phenomenon and the conditions that make anticipation possible. The paper's first part seeks to show that many scholars from various research fields worked on the many nuances of anticipation. The paper's second part seeks to discuss the difference between the capacity of anticipation and the nature of systems able to exhibit anticipatory behavior. The former endeavor adopts a descriptive attitude, whilst the latter seeks to understand what it is that makes anticipation possible.\nDesign/methodology/approach: - The paper presents a theoretical and experimental analysis of anticipation and anticipatory systems.\nFindings: - Anticipation is a widely studied phenomenon within a number of different disciplines, including biology and brain studies, cognitive and social sciences, engineering and artificial intelligence. There is a need for relying on at least two different levels of analysis, namely anticipation as an empirical phenomenon and the idea of an anticipatory system or the study of the internal structure that a system should possess so that it can behave in an anticipatory fashion.\nResearch limitations/implications: - The literature summarized by the paper is only part of a substantially larger body of documents. More extensive analyses are needed to firmly establish the conclusions suggested.\nPractical implications: - The paper allows better understanding of the complexity of anticipation and the differences between types of anticipation (e.g. between explicit versus implicit anticipation).\nOriginality/value: - For the first time, the distinction implicitly present in the surveyed literature between anticipation as an empirical phenomenon and the idea of anticipatory system as the study of the conditions that make anticipation possible is raised explicitly.\n",
"role": "assistant"
}
] |
Blurring the borders between B2B and B2C: a model of antecedents behind usage of social media for travel planning
|
[
"Perceived enjoyment",
"Information quality",
"Behavioral intention",
"B2B B2C convergence",
"Travel planning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: 1.1 Debating the currency of the B2BC dichotomy\n2. Theoretical background, research model and hypotheses: 2.1 Sources of information and social networking sites\n3. Research method: 3.1 Survey instruments design\n4. Analysis, results and discussions: To accurately measure results, composite scores were created, and the subsequent data were put through validity and reliability tests, including the assessment of common method bias, reliability and validity. A common method bias (CMB) occurs when the measured variance is attributable to the method used, rather than the constructs represented by the measures (Podsakoff et al., 2012). Harman's single factor test was used to measure CMB (Podsakoff et al., 2003). All variables were loaded onto a single factor and the unrotated solution was analyzed using SPSS to determine the number of factors needed to account for the variance in the variables (Podsakoff et al., 2003). Ideally the single factor should explain less than 50 per cent of the variance. In the present study, the single factor explained 49 per cent of the variance. It is concluded that common method bias does not occur in this instance.\n5. Implications, limitations and further research suggestions: The stated intent of this paper was twofold. First, to draw attention to the question posed earlier by Wilson (2002), whether a strict parallel conceptualization of marketing into B2B and B2C is as valid today as it may have been before. We contend that such distinctions are no longer paramount, particularly in industries increasingly driven by consumer-generated content and where customers are empowered through erosion of information asymmetry between the firm and the customer. Wilson (2002) had identified the supposed professionalism of organizational buying as a factor that conventionally has set the domains of B2B and B2C apart, yet with the tools at their disposal, the level of professionalism displayed by consumer buying is no longer trivial. Protracted decision-making, which is another distinctive factor of B2B buying (Webster and Wind, 1972), is not necessarily a function of professionalism but rather one of bureaucratic expectations (Wilson, 2002). This may well be an even less potent divider in industries with higher information symmetry between firms and customers, such as the travel industry. As Kietzmann and Canhoto (2013) showed in their integrative model of eWoM, the creation and dissemination of eWoM is a multi-stage multi-level appraisal, reaction and coping process. They made the distinction that where social media is concerned, it is the consumers who pull the firms into social media and not the other way round. This puts further impetus to place the individual customer as the centerpiece and enable us to view the industrial and consumer marketing paradigms as increasingly intertwined.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to call for a scrutiny of the dualist approach to business-to-business (B2B) and business-to-customer (B2C) marketing in industries driven by consumer-generated content. It posits that individual consumer-centric factors are influential for B2B marketing as well in sectors such as the travel industry and investigates the determinants of tourists' intention to use social media websites for travel planning.\nDesign/methodology/approach: Integrating constructs from IS and marketing literature, the paper proposes information quality and perceived enjoyment as antecedents of perceived usefulness, attitude and intention to use. The research model is tested using data from social media users with experience in travel planning.\nFindings: Results show that perceived usefulness and information quality are stronger predictors of attitude and behavioral intention than perceived enjoyment. Enjoyment was not found to be strongly influential. Relevancy and reliability of information and its usefulness concerning travel-planning needs were found more influential.\nResearch limitations/implications: Data were collected from social media users, raising possible issues of representativeness.\nPractical implications: The paper offers clarity regarding antecedents of downstream user behavior which can be of significant value. Demarcations in B2B and B2C perspectives blur in the context of social media, enabling more effective integration.\nOriginality/value: The paper brings in and validates the roles of information quality and enjoyment as influencers of behavior. Identifying the travel industry as a sector having greater likelihood of B2BC convergence, the paper extends IS adoption research to user-interactive sites in the travel-planning context, which can benefit the consumer as well as the supply side.\n",
"role": "assistant"
}
] |
Adult online education: Investigating learning theories and their application to distance learning
|
[
"Online learning",
"Adult education",
"Theories",
"Learning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: With the advent of technology, it has become possible for education to develop from the traditional classroom setting to online degrees and courses for adult learners. Learning theories have been developed to look at how students learn and understand. Such theories allow educators to have a better understanding of learning, which can help them to design more engaging and effective courses, which in turn leads to better and increased learning and development for the student.\nLearning theories and model: Behaviorism\nOnline learning: Online learning can present opportunities for reflection that are not present in the immediate setting of a traditional classroom. Online discussions can allow a student to read their colleagues' posts and reflect on them before posting themselves; they can also review their own writing before posting. There is more time available for reflection online than in a classroom where debate may take place there and then. This reflection may allow for better and deeper understanding of the material, and also of the self as a learner.\nComment: The review is based on \"Adult learning theories: implications for online instruction\" by Arghode, V., Brieger, E. and McLean, G., published in European Journal of Training and Development.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to review the latest management developments across the globe and pinpoint practical implications from cutting-edge research and case studies.\nDesign/methodology/approach: This briefing is prepared by an independent writer who adds their own impartial comments and places the articles in context.\nFindings: This paper looks at four learning theories and how they can be applied to adult online and distance learning. No single theory encapsulates adult learning, but using the theories in conjunction can aid instructors to design effective and engaging courses which will then lead to improved and successful learning.\nPractical implications: The paper provides strategic insights and practical thinking that have influenced some of the world's leading organizations.\nOriginality/value: The briefing saves busy executives and researchers hours of reading time by selecting only the very best, most pertinent information, and presenting it in a condensed and easy-to-digest format.\n",
"role": "assistant"
}
] |
Age-related cognitive changes and distributed leadership
|
[
"Diversity",
"Leadership",
"Cognition"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Ageing trends, deteriorating economic conditions and skills shortages suggest that older workers are likely to remain in the workforce for some time. Research on managing an age-diverse workforce has, accordingly, proliferated, developing around themes (to mention a few) such as motivation, conscientiousness, engagement, attitudes and so forth - usually with a focus on \"challenges\" rather than \"opportunities\". Managing diversity requires a broad range of skills and presents challenges to existing leadership approaches. Age-based cognitive changes may require managers to take into account the psycho-physical requirements of jobs, learning and skills acquisition, attentional processes and so forth, to get the best from older workers and provide a satisfying and productive work experience for the employee. Opportunities exist for tapping into the creative abilities of individuals, through effective, inclusive, leadership. This paper examines age-linked changes to the components of intelligence (crystallised and fluid) and the implications for leadership. It is argued that distributed leadership will improve innovation and competitive sustainability by adapting leadership styles to create an inclusive and more participative workplace that recognises and encourages the different cognitive contributions across the age spectrum.\nAge-related components of intelligence: The components of cognition have been the subject of research for over a century. Cassel (1969), for example, reviewed 11 theories of intelligence, dating from 1900 through to 1964. Cassel defined intelligence as \"being cognitive, embracing the notion of ability to learn or to manipulate objects, and as being adaptive, embracing the notion of adjusting to one's total environment\" (p. 39). While knowledge has advanced on the components of cognition (Aleman et al., 2001; Avolio et al., 1990; Burgess, 2006; Hunt, 2000), the work of Cattell (1963) and Horn (1968) is seminal in conceptualising and measuring crystallised (Gc) and fluid intelligence (Gf), age-linked changes in cognition, neural physiology and performance correlates.\nDistributed leadership encourages employee contribution: Leadership research has traditionally focused on the individual and the large number of published definitions generally conveys the sense of influencing others to achieve organisational objectives (Lawler, 2008). More recently there have been calls to broaden leadership conceptualisations to include \"followers, peers, supervisors, work settings/context, and culture [...] depicted in various models as dyadic, shared, relational, strategic, global, and [as] a complex social dynamic\" (Avolio et al., 2009). Further, the modern leader should recognise that they do not necessarily have all the answers and seek to encourage the strengths and contributions of others (Ancona et al., 2007; Sadler, 2001).\nImplications of age-based cognitive change and leadership: The literature on distributed leadership suggest that effective leadership will look beyond the individual leader, recognising the varieties of expertise available in all the workforce, emergent leadership, shared cognition and concertive action, then encourage permeable boundaries necessary to tap into the creative ideas and problem solving, related to high fluid intelligence, as well as the wisdom and experience of older workers. The skill sets for leadership, implied by these elements, call for continual personal development, insight into the strengths and weaknesses of self and other employees, recognising that creativity, innovation, and problem solving are essential for competitive sustainability (and are not exclusively the domain of senior leadership) and facilitating effective networks and integrative mechanisms (Ackerman and Beier, 2006; Roskos-Ewoldsen et al., 2008).\nConclusion: This paper examined age-related cognitive changes and argued that managing an age-diverse workforce will benefit from examining leadership styles through a distributed leadership \"lens\". The needs and contributions of individuals, across the entire age spectrum, will vary and a challenge exists in managing and nurturing individuals to continue to acquire and use their skills and knowledge to contribute to competitive sustainability. An awareness of age-based changes in cognition should inform the manager of the strengths and development opportunities for individuals. A more inclusive style of problem solving and concertive action, whilst not abdicating authority, responsibility and guidance, may tap into individuals with high levels of fluid intelligence for creative problem solving under conditions of impoverished information, irrespective of the individual's formal position in the hierarchy. The relationship between fluid intelligence and crystallised intelligence also suggests that managers need to provide developmental opportunities to younger staff in order to build a store of knowledge and experience, which will contribute to the firm's intellectual capital.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the evidence for age-related changes in cognition and the implications for leadership styles. In particular, a case is argued for distributed forms of leadership that encourage contribution across the age spectrum and hierarchical levels.\nDesign/methodology/approach: This paper takes a conceptual approach, combining the psychology and management literatures in arguing the case for newer leadership forms, appropriate to an ageing workforce.\nFindings: Three principal components of intelligence (fluid, and crystallised intelligence and working memory) are considered and it is argued that high levels of fluid intelligence, generally higher in younger employees, should be accessed while being balanced by crystallised intelligence (experience). Distributed leadership has been mainly applied in educational settings. This paper argues for distributed leadership to maximise creativity and innovation.\nPractical implications: Leadership forms that maximise creative input from staff across all age levels are likely to contribute to firm innovation and sustainability. Additionally, job satisfaction and turnover among junior staff may be positively influenced through opportunities for greater participation.\nSocial implications: The elements discussed in this paper address important leadership issues for managing a multigenerational workforce.\nOriginality/value: Distributed leadership has been discussed in educational and health literatures for some time; however it is only recently that this approach to leadership has appeared in mainstream management literature. The discussion of age-related changes and distributed leadership introduces and important topic for further research in newer forms of leadership.\n",
"role": "assistant"
}
] |
Joint stock company with variable capital (JSCVC)
|
[
"Credit union",
"Mutual fund",
"Pension fund",
"Rastin PLS banking",
"Rastin sharing accounting",
"Variable capital company",
"Saving and loan association"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Change of assets of a company is a nature of financial institutes. Change of capital and the number of shareholders of conventional joint stock companies are due to specific formalities established by Commerce Law and are not so easy to be done repeatedly in small periods. For having quick varying capital capability, new legal institution should be defined. To cover this shortage, we define joint stock company with variable capital (JSCVC) which can prepare the necessary background for financial profit and loss sharing (PLS) and other financial sharing (Mosharekah) activities.\n2. Joint stock company with variable capital: JSCVC is a joint stock company in which the capital and shares vary by new entrance or withdrawals of stockholders or by lapse of time through changing sharing proportions because of changes in shares' durations. Amount of nominal capital of the company at any point of time is calculated by summing up the nominal values of the shares of shareholders at that time, but company's yields are distributed by time-duration weighted of the shares. The time points and periods for clearing accounts are predetermined according to the article of association of the company. Time points or financial periods can be orderly cyclic/periodical or disordered but predetermined.\n3. Account clearing arrangements: Calling the nominal value of each share q and the number of shareholders m and the nominal value of the capital of the ith person: Di and the duration of the participation of ith person: ti, the number of shares of each shareholder and total number of shares of all shareholders of JSCVC can be calculated as follows:\n4. Application in financial institutions: JSCVC can prepare sound background for many financial sharing activities. JSCVC can be used by majority of investment funds, credit unions, saving and loan associations, pension funds, thrift saving plans as well as Islamic banks and banking and non-banking financial sharing activities. Concluding from JSCVC definition, deposits at bank will have the same properties as shares in JSCVC. The latter justification can change the bank's legal and operational performances to a faired financial discipline and market clearance and, moreover, tightly links money market to capital market which can dissolve both markets inherent volatilities and fluctuations (Bidabad, 2010, 2011a, 2011b, Bidabad and Hassan, 2011). Some applications of JSCVC are as follows.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to propose joint stock company with variable capital (JSCVC), as financial sharing funds and banks necessitate that their capital and number of shareholders be instantaneously variable. Legal personality and accounts clearing of this type of corporations are different from conventional companies.\nDesign/methodology/approach: JSCVC is a corporation in which capital and shares of shareholders vary by new entrance or withdrawal of shareholders at any point of time.\nFindings: Interest rate-based calculations were removed and Rastin Sharing Accounting was applied for JSCVS. Shareholders of JSCVC share the company's nominal capital proportional to nominal values of their shares. Financial outcome of JSCVC is proportional to values of shares weighted by shares duration of participation.\nResearch limitations/implications: To prevent spoiling of shareholders' rights, legal procedure of issuing shares for JSCVC should be defined in compliance with domestic commerce laws in any country.\nPractical implications: JSCVC can be used by majority of investment funds, credit unions, saving and loan associations, pension and provident funds, thrift saving plans as well as Islamic banks and financial sharing activities. In JSCVC, deposit at a bank is treated as a share of the company (bank).\nSocial implications: JSCVC has fair profit distribution and accounts clearing arrangements.\nOriginality/value: Different variable capital companies have been defined in many countries' laws, but essential modifications are presented in JSCVC definition to regulate financial sharing arrangements and bank's performances.\n",
"role": "assistant"
}
] |
Minority community willingness to pay for earthquake insurance
|
[
"Disaster risk",
"Willingness to pay",
"Risk perception",
"Disaster insurance",
"Ethnic minority communities",
"Government aid",
"Housing earthquake insurance",
"Insurance experience"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Background statement\n2. Methodology: Study areas\n3. Results: Descriptive statistics\n4. Conclusion: In summary, this study analyzes how WTP for HEI is influenced by potential factors in the first pilot EMC in China. Wang et al. (2012) and Tian and Yao (2015) studied the same topic, but their bases were theoretical assumptions instead of HEI in practice. Others also studied EMC from different angles, such as ethnic and racial differences (Senkbeil et al., 2014), poverty vulnerability (Cao et al., 2016), flood hazards (Chan et al., 2017) or fire hazards (Du et al., 2017) in EMC. This study addresses certain EMC in earthquake-prone areas of Dali. The results show that WTP for HEI is significantly affected by risk perception, prior experience with insurance, available government aid, age and out-migrating labors.\n5. Limitations: This study provides meaningful information toward the WTP for HEI in EMC in Dali; however, there are some limitations which require further study. First, the conclusion may not be directly applied to other EMC, since other ethnic groups are not represented to a large extent, and the sociodemographic characteristics vary from region to region. Moreover, although the study results reveal that ethnic-culture factors are statistically insignificant to WTP for HEI, the potential influence that ethnic-culture factors have on WTP should not be underestimated. Further research could focus on ethnic-culture factors in disaster risk reduction to a greater degree. This topic is a burgeoning research area in disaster risk management; nevertheless, it is still in need of further exploration and analysis. In the future, study on other EMC could be conducted to compare with this study. In addition to this, in this study only 9 percent of respondents were non-EMC, thus, a comparative study between EMC and non-EMC regarding WTP for HEI is worthy of further investigation.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the relationship between potential affecting factors and the local communities' willingness to pay (WTP) for housing earthquake insurance (HEI) in the context of ethnic minority communities.\nDesign/methodology/approach: A literature review was done to identify possible factors affecting WTP for HEI. Fieldwork was conducted in 2017 in Dali Minority Autonomous Prefecture, where the first Chinese HEI was launched in 2015. Interviews were done in two earthquake-prone counties, as the main ethnic minority communities in the area. A total of 536 questionnaires were collected and used as empirical data for testing the impacts mechanism.\nFindings: Respondents' risk perception, risk exposure, self-prevention behaviors, government aid, insurance experience and sociodemographic characteristics were hypothesized as theoretical indicators correlated to WTP for HEI. Empirical analysis results predict that WTP for HEI is significantly influenced by risk perception, insurance experience, government aid, and age and out-migrating labors. It is evident that higher risk perception and more insurance experience lead to stronger desire for HEI coverage. However, dependency on government aid negatively affects WTP for HEI. Moreover, WTP for HEI is negative in relation to age and out-migrating labors. Surprisingly, ethnic-culture factors were not statistically significant to WTP for HEI.\nOriginality/value: This paper is an attempt to identify and verify factors affecting WTP for HEI, bridging the gap of inadequate research on WTP for HEI in ethnic minority communities.\n",
"role": "assistant"
}
] |
Nurses' shop stewards and their collaboration with management
|
[
"Trade unions",
"Nurses",
"Hospital management",
"Public sector reform",
"Change management",
"Denmark"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nBackground: Trade union membership is very high in Denmark. This is also true of nurses, about 96 per cent of who are union members. The Nurses' union (Danish Nurses Organisation) is generally regarded as a powerful player in the terms of influence on the individual workplace. The union also has a strong voice in policy processes in the field of health and care services, which in Denmark are predominantly publicly organised and financed. Nurses are well educated, to bachelor level, and they enjoy a wide respect as professional counterparts for medical staff (Spitzer and Perrenoud, 2006). A strong union is a result of the fact that nurses, in Denmark as in other countries, are key and irreplaceable staff in most segments of the health sector (Bellman, 2003).Shop stewards liaise between common members on the workplace floor and union representatives. They also serve as links between the union members working in hospitals and in services caring for the elderly etc. on the one hand, and the decision makers at national, regional and managerial levels on the other hand. Generally, we can find three different types of approaches (Kessler and Heron, 2001):1. management-facing;2. member-facing; and3. union-facing.Over the past years, the role of unions in general has been contested to some extent (Bild et al., 1998). Despite this membership loyalty is consistently high. However, nurses raise questions about the union's work at all levels from the chairman to the single shop steward. The portfolio of union operations, the issues raised, strategies launched, and action taken are under continuous debate, and the union is increasingly obliged to justify its actions. The future tasks and profiles of shops stewards in the health sector are also under discussion, not least as a consequence of major changes in the public sector (Vallgarda et al., 2001; Clark and Clark, 2003).In 2007, a comprehensive structural reform in Denmark led to the merger of municipalities and counties into fewer units. Responsibilities for health and care services were simultaneously reassigned to the new municipalities and regions. Many nurses changed employers during this process. The reform will be a stepping-stone towards further reorganisation, particularly in the hospital sector. In 2007, the Government also launched a \"Quality Reform\" which is expected to affect many categories of jobs and not the least in the nursing field.The way in which health services in Denmark are organised has developed gradually over the past 30 years. With some delay, hospital and elderly care management has, inspired by the private sector, and presumably reforms in other countries, implemented managerial reforms and more decentralised and participatory decision structures (Arrowsmith and Mosse, 2000; Dahl and Eriksen, 2005). Through these reforms, more focus has in Denmark been placed on collaboration between differing professional groups, which is also a challenge for the relationship between shop stewards representing the different professions (Westenholz, 2006). At the same time, nurses feel the classical dilemmas of production and efficiency versus care for human beings.A stronger reliance on the involvement and participation of daily management is paralleled with a centralisation of strategic issues. New Public Management inspired forms of navigation are gradually being implemented with the aim of ensuring a better control of expenses, increased efficiency and transparency (Jespersen, 2005). These modernisations are however, not received uniformly enthusiastically in the workplace (Nygard, 2006). Nevertheless, the pace and pressure for changes is increasing, and under such circumstances, conflicts between management and employees are likely to escalate. Under new service paradigms (Scott et al., 2000; Zimmermann, 2006) in the health sector shop stewards and managers alike will be required to reassess the collaborative behaviours and structures.This study examines the collaboration between the nurses' shop stewards and management in the Danish health sector. Numerous workplace studies have demonstrated that democratisation and participation in decision-making is of major importance for job satisfaction (Miller and Monge, 1986) and for the smooth operation and development of the workplace (Westenholz, 2006). Employee involvement may take variety of forms. In a Danish context, employers are usually able to approach management directly, either in operational meetings and other formal participatory settings, or individually. However, shop stewards continue to have a facilitating role in many workplaces. He/she is the managers' short cut to information about the total workforce's \"average\" opinion. On the other hand, the shop steward may be a channel for disseminating management information to employees. In many cases, the shop steward will be the person responsible for negotiating agreements on behalf of the members, regarding for example issues such as working hours, work plans, policies in the field of training, and to some extent wages.Several Danish studies have explored the relationship between management and union representatives. Generally, there is a low power distance in Danish workplaces. In her historical account, Rasmussen (2005) shows that the union have traditionally been strongly opposed to the management in the workplace, as a consequence of a basic conflict of interests between employees and employers. However, in more recent years, a shift of paradigms has taken place. Union representatives take an increasingly proactive role in the continual development of the workplace. Lund and Hvid (2004) refer to survey results that point at closer and better collaboration with the management as a satisfactory outcome for both sides. Recently altered negotiation procedures for wages, and more incentive and performance based resource allocations have forced this collaboration into new dimensions. However, as demonstrated by Navrbjerg (2004) and Navrbjerg and Madsen (2004), management still often outmatch employee representatives in democratic assemblies in terms of insights and power.Closer relations (Bild, 1997; Jensen et al., 1998) between employees, management and the unions go hand in hand with a diminishing number of strikes (Scheuer, 2005). The unions' attention to the needs of the employees has been supplemented by an increased recognition of the importance of the viability and competitiveness of the workplace. The threat of outsourcing has contributed to a wider strategic perspective in many unions. The development has led to further dialogue and search of compromises. Occasionally, union representatives have been extraordinarily proactive in terms of offering assistance to ensure the survival of enterprises in crises. Kristensen and Zeitlin (2004) claim that a \"new deal\" is emerging where part of the managerial responsibility is being transferred to the shop stewards, and where employees are willingly to \"pay\" for deeper involvement in decision-making with increased loyalty. From this perspective, the role of the shop steward is plausibly shifting away from \"rights control\" and is developing to become an \"entrepreneurial\" entity in the workplace.\nPurpose of the study: Relationships between the union representatives in the workplace are not stable. They are affected by incidental events and structural changes. Accordingly, labour relations are a matter for continual investigation. This study addresses the extent and nature of the collaboration between nurses' shop stewards and their workplace managers in the Danish health and care sector. There is particular focus on the structural factors in the managerial hierarchy, the power of the shop steward in terms of experience and member support, and on the nature of the changes that the workplace has been through over the past years.The following main hypotheses will be investigated:H1. Shops stewards maintain closer relations and exhibit higher degrees of loyalty towards the nearest managers rather than to managers at higher levels in the hierarchy.H2. The most professional shop stewards exhibit more positive relations to management than less experienced and dedicated shop stewards.H3. Shop stewards with colleagues, who do not support their union representative, have less favourable relations with management.H4. Shop stewards with colleagues, who are dissatisfied, have less favourable relations with management.H5. Shop stewards with colleagues under intense working pressure have less favourable relations with managers.H6. The intensity of the relationship with the management is related to the practical changes that the workplace has experienced.\nResearch methodology: This research is performed in cooperation with the Danish Nurses Organisation. It is based on a 2006-survey of all nursing shop stewards in Denmark, in total 1,270 persons. The study has a very high response rate (87 per cent), which indicates the importance placed on the issues raised in the survey by respondents.The study is comprehensive. Main focuses are on the changes in the structure of the health sector and shop stewards' responses to these changes. The study also includes a qualitative segment. This paper reports only on selected findings from the quantitative research.Data analysis is primarily based upon cross tabulations and binary correlation analysis.\nThe relationship between the shop stewards and managers: H1: Shops stewards maintain closer relations and exhibit higher degrees of loyalty towards the nearest managers rather than to managers at higher levels in the hierarchy\nThe relevance of a shop stewards background for their relationship with management: H2: The most professional shop stewards exhibit more positive relations to management than less experienced and dedicated shop stewards\nBackup from colleagues: H3: Shop stewards with colleagues, who do not support their union representative, have less favourable relations with management\nColleagues' satisfaction with working conditions: H4: Shop stewards with colleagues, who are dissatisfied, have less favourable relations with management\nWorking conditions and collaboration with management: H5: Shop stewards with colleagues under intense working pressure have less favourable relations with managers\nChanges in the workplace and relations to the management: H6: The intensity of the relationship with the management is related to the practical changes that the workplace has experienced\nConclusion and discussion: This study demonstrates in a variety of ways that the nurses' shop stewards hold close and pragmatic relations with the nearest as well as the upper management in the work place. Collaboration is clearly result oriented. It is also important to notice that disagreement does not discourage contacts or collaboration.The intensity of collaboration is affected to a marginal extent only by the situation in the workplace. Negative and dissatisfied colleagues or colleagues under pressure do, however, make life somewhat more difficult for the shop steward. Hardship, in terms of working conditions at the workplace, has some influence on collaboration with management, and favourable working conditions enhance this relationship. Changes at the workplace do not appear to significantly affect the degree to which the shop stewards are invited to cooperate with the management.This research provides considerable input to the debate regarding the unions' support to and backing for their representatives in the workplace in a situation where significant changes are made to working conditions. It is noteworthy, that shop stewards with greater experience, dedication and a power base based upon many members are significantly more oriented towards the processes of dialogue with the management. This might lead the union to consider how to continually raise the qualifications of shop stewards. In addition, since an average shop steward allocates only a relatively modest amount of time to union tasks, there is need for the union (centrally and locally) to consider very time-efficient supportive measures and learning opportunities for the shop stewards. At the same time it is important to recognise and enhance the stabilised, workable alliances between shop stewards and management in the workplaces.Managerial initiatives are plentiful, and the scenes and agendas in the health sector are rapidly changing. It is a demanding undertaking for shop stewards to keep apace with and ensure continual influence on the issues arising. Other segments of the study demonstrate that a competence gap is particularly significant in topics related to New Public Management strategies and methods, while shop stewards are mostly sufficiently well trained to handle traditional issues such as wage negotiation, working conditions etc.The relationships between shop stewards and managers are regarded as constructive and without deep hostilities. For this reason it is likely that the management will welcome and take on board improved competences for the shop stewards, even where no full agreement exists.\n",
"role": "user"
},
{
"content": "Purpose: - This study investigates the collaboration between Danish nurses' shop stewards and workplace management. The aim of the study is to track changes in workplace climate after a major structural reform of the health sector.\nDesign/methodology/approach: - The data source for the study is a comprehensive survey among union representatives in the health and care sectors.\nFindings: - Generally, and not surprisingly, shops stewards maintain closer relations and a higher degree of loyalty to the nearest managers rather than management at higher levels in the hierarchy. It can also be demonstrated that more experienced shop stewards, those who have been employed in this position and in the workplace for the longest terms have more affirmative relations to management than less experienced shop stewards with shorter tenure. Those shop stewards who spend much time on the entitled duties are rewarded with positive collaboration with management. Hard times at the workplace and dissatisfied colleagues, who do not support their union representative, often result in less rewarding relations with management. Quite unexpectedly, the intensity of relations with management is not significantly related to structural or other changes that the workplace has experienced over the past two years. Changes are therefore accepted as inevitable and regular occurrences in the health sector.\nResearch limitations/implications: - The response rate is very high in the survey. Further qualitative research may reveal details about the background and implications.\nPractical implications: - The study suggests that many shop stewards may suffer from a competence gap in terms of more advanced new public management strategies and tools. This gap has not yet been successfully filled by the services and training activities offered by the Danish Nurses Union.\nOriginality/value: - Results from the study are being taken on board in the union's strategies. The evidence is also helpful for the managers in the health sector, as they are seeking to develop a constructive the collaboration with the unions.\n",
"role": "assistant"
}
] |
Why the youth are so eager for university education? Evidence from Iran's labor market
|
[
"Iran",
"Wages",
"Propensity score matching",
"Average treatment effect",
"Overeducation",
"J45",
"J31",
"I26",
"I23",
"C54"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Like many developing countries, Iran has experienced a rapid increase in the tertiary education enrollment since 2000. As a result the number of young university graduates has also sharply increased in the past decade. The sharp growth in supply of educated job seekers eventually exceeded the availability of suitable jobs and the unemployment rate for university graduates in many fields of specialization has been higher than high school graduates for nearly a decade. Recent studies have also revealed that a large portion of university graduates who have found employment are overeducated for their current jobs and do not utilize their university skills. While the information about poor job market prospects for university graduates is widely available to high school students and their parents, the demand for university education remains strong and most Iranian youth still believe that without a university education they will have less chance for finding a successful job.\n2. Statistical overview: The rapid growth of student enrollment in tertiary education institutions of Iran is demonstrated in Figure 1. We observe larger enrollment growth rates after 2000. This sharp growth was partly facilitated by the rapid expansion of the semi-private Islamic Azad University and Payam Noor Long-distance Learning University. Overall we observe that total enrollment rose from 1.2 million students in 1996 to 4.9 million in 2013. This rapid and unprecedented increase is the main cause of the high unemployment and underemployment rate among university graduates. While total enrollment and graduation continued to increase in 2010-2013, the economy of Iran suffered a severe recession as a result of the international sanctions. This economic decline led to the elimination of a large number of professional and skilled jobs.\n3. Measuring overeducation: Overeducation (also referred to as over schooling) is usually defined as the difference between a worker's attained level of schooling and the level of schooling required for the tasks that she/he must perform in her/his current job. To measure years of over schooling, one must start by measuring the required level of schooling in a job, or even a group of jobs. Works on overeducation have applied three methods to measure overeducation: first, employee self-assessment which relies on survey questions. Each worker is asked to report the level of schooling requirements for her/his current job. This method is subjective and vulnerable to errors in perceptions and assessments of participants. Hartog (2000) argues that respondents may have a tendency to overstate the requirement of their career. Second, expert analysis of the schooling and skill requirements for various occupations. This method is based on information contained in occupational classifications. In spite of its simplicity and objectivity, the method usually cannot capture the dynamic changes in skill requirements of each occupation (Leuven and Oosterbeek, 2011).\n4. Literature review: Empirical studies in many developed and developing countries have revealed that supply of university graduates has a tendency to catch up with demand for skilled graduates in most countries even when the average economic growth is positive and unemployment rate is low. In most western industrial countries this catch up was realized in 1970s and 1980s and was followed by a surplus of university graduates in many fields of study (overeducation). One of the earliest studies on overeducation at national level was Freeman's (1976) book Overeducated American. This controversial book generated a large amount of academic interest in the topic in the USA and Europe.\n5. Methodology: For our econometric analysis we have selected the PSM method, which has been applied in several empirical studies to estimate the causal treatment effects of overeducation on wages (e.g. McGuinness, 2008; Ordine and Rose, 2014). The PSM is a nonparametric technique that pairs treatment and comparison units that are similar in terms of their observable characteristics. It leads to a significant reduction in selection bias by comparing the outcome of interest between the units in the control and treatment groups which, otherwise, have very similar characteristics. As a result, this comparison method controls for workers' heterogeneity. Conditioning on all relevant observable characteristics Xi makes the comparison algorithms too complicated for a high dimensional vector. Rosenbaum and Rubin (1983) define propensity score as probability of receiving treatment as a function of Xi, independent of assignment into treatment. In our terminology, treatment units are the overeducated workers and the well matched adequately educated workers are defined as the control group.\n6. Data set: The data on wages and socio-economic characteristics of workers that we have used in our econometric analysis are extracted from the Household Income and Expenditure Survey (HIES) data. This data set is generated by the Iran Statistics Center on an annual basis and the sample size is optimized for urban and rural population in all provinces of Iran. The HIES includes detailed data on income, expenditure, occupation, age, gender and many other characteristics of the households that are included in the sample. Based on data availability, we selected the HIES data sets for five years during a 13-year interval (2001, 2005, 2009, 2013). It was during this interval that Iran experienced a rapid increase in university enrollment and the number of university graduates. This database includes ISCO occupation codes for adult working individuals. We have limited our sample to individuals who are employed in semi-skilled and unskilled occupations. These occupations lie in ISCO one digit categories of 4 and above. A list of these categories is available on the website of the International Labor Organization.\n7. Empirical findings: Matching on the propensity score is basically a weighting scheme in which the weights are applied on matched comparison workers j for each overeducated worker. Then estimated treatment effect as an empirical counterpart for ATT is given as follows:(3)\n8. Concluding remarks: In this paper, we have tried to shed light on the reasons behind the strong demand for university education in Iran, despite the fact that many university graduates are not able to find suitable professional jobs that match their university education. We hypothesize that one of the underlying causes of this strong demand for university education is that even if a university graduate cannot find a professional job and accepts a semi-skilled or low-skill job, she/he will enjoy a wage premium over less-educated workers because of her/his university degree.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to estimate the wage difference between overeducated and adequately educated workers in a sample of semi-skilled and low-skill occupations in Iran's labor market. The objective is to see if overeducated employees in these occupations enjoy higher wages in comparison to their adequately educated co-workers.\nDesign/methodology/approach: The authors use the propensity score matching (PSM) model to estimate the wage difference between overeducated and adequately educated workers in a sample of semi-skilled and low-skill occupations in Iran's labor market. The PSM method allows the authors to prevent selection bias by comparing each group of overeducated workers with a matched group of adequately educated workers with similar socio-economic characteristics.\nFindings: The results show that in Iran's labor market, the overeducated workers enjoy a wage premium in the range of 10-25 percent for their excess education when they have to work in semi- or low-skill occupations. While this relative advantage has gradually declined for private sector employees over the period 2001-2013, it has remained stable for public sector jobs. The result is attributable to the fact that salary and benefits for public sector employees are directly linked to education attainment and their work experience. The findings show that the relative wage advantage of overeducation is larger for the younger employees with ten or fewer years of experience who have more education than older workers. Overall these findings offer an explanation for the strong desire of Iranian youth for university education. If a university graduate finds a job that matches her/his specialization she/he will enjoy a higher salary than a high school graduate. If she/he has to accept a semi-skilled or low-skill job for which she/he is overeducated, she/he still enjoys a wage premium over her/his co-workers who are not overeducated.\nOriginality/value: The analysis makes three unique contributions: first, the authors use a unique and detailed micro-data for an economy (Iran) in which the public sector dominates the private sector. The authors investigate the hypothesis for private and public sectors separately. Second, the authors divide the sample of workers into market newcomers and experienced workers. The authors analyze the impact of overeducations on each group separately. Third, the authors use the treatment effect of being overeducated on the individual's monthly wage by the PSM. The advantage of the PSM method is that it eliminates the selection bias in the wage effect of overeducation, whereas the traditional regression-based techniques may result in this type of bias.\n",
"role": "assistant"
}
] |
How market orientation contributes to innovation and market performance: the roles of business analytics and flexible IT infrastructure
|
[
"Market orientation",
"Responsiveness",
"Business analytics capability",
"Flexible IT infrastructure",
"Intelligence dissemination",
"Intelligence generation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Within the past decade, firms have competed to attract customers' attention and attempted to generate superior value regarding market changes (customer preferences or competitors' action) (Terawatanavong et al., 2011). Market orientation (MO) has therefore become an invaluable approach to achieve market-related information (i.e. customers' needs), offer effective responses (Tippins and Sohi, 2003) and highlight the role of innovation as a crucial concept in creating competitive advantage through developing unique products or services (Bellamy et al., 2014). Yet, the understanding of the different perspectives of MO and how it affects innovation and market performance need further studies. Thus, the present study intends to find out the influence of MO on innovation and market performance through considering MO from the market intelligence perspective.\n2. Theoretical framework and hypothesis development: The way that organizations can effectively cope with turbulent environments has been a topic of great interest to both practitioners and academics. In response, firms incredibly rely on external sources of information to introduce new services or products and meet customers' needs (Kostopoulos et al., 2011). Despite the significant role of MO and innovation as two key strategic resources for a business success, less attention has been paid to them in the extant literature (Wang et al., 2013b; Olavarrieta and Friedmann, 2008; Pekovic et al., 2016). Section 2.1 presents the theoretical underpinnings of MO, innovation, and market performance. Besides, specific attention is devoted to the constructs of flexible IT infrastructure, BA capabilities and market turbulence and their moderating effects on the proposed conceptual model.\n3. Research methodology: 3.1 Sample and data collection\n4. Results: Partial least squares (PLS) was used for data analysis. PLS is a powerful second-generation multivariate technique that uses a component-based approach to produce the estimations. SmartPLS 2.0 (2005) was used as a data analysis tool which supports two levels of assessment:\n5. Discussion: Despite the important influence of MO on innovation and market performance, empirical evidence are limited and have only addressed this issue in a direct relationship form without considering mediator/moderator variables. In response to the mentioned research gap, the present study has focused on the role of MO from market intelligence perspective and its impact on innovation and market performance. We also consider the moderating roles of IT infrastructure, BA capability and market turbulence and their influence on the proposed research model. We believe that analyzing the effect of these constructs can contribute to our understanding of how they influence firm responsiveness and indirectly improve innovation and market performance. The outcome of this study confirms the role of all moderating variables.\n6. Conclusion: Despite the significant attention about MO and innovation performance, little research has been done to explore factors may influence this relationship. The present study contributes to the body of knowledge by presenting a new conceptual model in order to investigate the relationship between MO, innovation and market performance by considering three main moderators, namely, flexible IT infrastructure, BA capabilities and market turbulence. In this study, MO defined as a three-step concept including intelligence generation, dissemination and responsiveness. The outcome of this study proposes that flexible IT infrastructure positively moderates the link between intelligence generation and dissemination and BA capability moderates the link between intelligence dissemination and responsiveness. This study also draws several future research directions to better understand both marketing and BA contexts.\n",
"role": "user"
},
{
"content": "Purpose: Market orientation (MO) (intelligence generation, intelligence dissemination and responsiveness) is known as one of the key concepts in marketing literature. Although prior research has widely focused on the meaning and application of MO, few attempts have been made to explore how market-oriented firms lead to innovation and market performance and what factors actually moderate this relationship. To fill this gap, the present study aims to explore the relationship between MO, innovation and market performance. This study also attempts to examine the intervening role of IT infrastructure, business analytics (BA) capabilities and market turbulence in the proposed model.\nDesign/methodology/approach: In this study, a questionnaire-based survey was undertaken to test the proposed hypotheses. To verify the proposed theoretical model, partial least squares (PLS)/structured equation modeling (SEM) was performed with 114 valid survey data.\nFindings: Despite prior studies which postulated innovation performance as the final outcome of MO (Han et al., 1998; Song et al., 2015), this study focused on innovation performance as a mediating outcome which finally leads to market performance. The statistical results approve the putative relationship which means managers would be able to realize the paramount role of innovation as an integral part of achieving higher market performance. In addition, no support was found for the relationship between intelligence generation and responsiveness. This finding shows that not all obtained information can help managers in the decision-making process.\nOriginality/value: This study aims to enrich literature by developing a conceptual model to test the link between MO, innovation and market performance. The value of this study is to investigate the roles of flexible IT infrastructure, BA capabilities and market turbulence as the potential moderators in the proposed model. The results advance the understanding of the influence of BA capabilities on the link between intelligence dissemination and responsiveness. Findings also show innovation performance as remarkable and deemed valuable capability, leading to higher performance in marketing-related activities, particularly in highly turbulent markets.\n",
"role": "assistant"
}
] |
Modeling information risk in supply chain using Bayesian networks
|
[
"Information systems",
"Information management",
"Bayesian networks",
"Risk factors"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: A supply chain consists of all the parties involved, directly or indirectly, in fulfilling a customer request (Mentzer et al., 2001). The supply chain not only includes the manufacturer and suppliers, but also transporters, warehouses, retailers and customers themselves. Within each organization, such as manufacturer, the supply chain includes all the functions involved in receiving and filling a customer request. In nutshell, a supply chain is the system of organizations, people, technology, activities, information and resources involved in moving a product or service from suppliers to customers. Vital and relevant information availability to a company in the supply chain, would help in optimizing the entire supply chain rather than just sub optimizing on a local interest. Accurate and timely information sharing across the supply chain improves supply chain visibility and that leads to enhanced supply chain performance (Caridi et al., 2014). This would result into better planned overall production and distribution which can cut down costs and give a more enticing and catchy final product leading to higher sales (Sharma and Bhat, 2013).\n2. Literature review: Information management plays an important role in the supply chain (Closs et al., 2005; Gunasekaran et al., 2008). Daugherty et al. (2006) considered information as a source of competitive advantage, in which the business data process is shared in a controlled way and an integrated and coordinated supply chain can be achieved (Wang et al., 2008; Boulesnane and Bouzidi, 2013). \"Information is the substance from which the managerial decisions are made\" (Forrester, 1962) and one of the pillars which supports a solid supply chain (Ballou et al., 2000; Ketikidis et al., 2008). In fact, delayed, scarce or distorted information can create serious problems in the supply chain (Chow et al., 2008; Handfield and Nichols, 2002; Power, 2005). One of the most serious effect is known as the bullwhip effect and was first identified by Forrester (1962). Globalization has caused increased complexity in the supply chain (Craighead et al., 2007). Information technologies designed to manage complex information flows within or between firms helps in creating value in supply chain by lowering the costs or increasing the service level (Biehl, 2005; Papadakis, 2006; Pandey et al., 2010).\n3. Research methodology: The study focusses at developing a BN for analyzing the various information risks within a supply chain. A thorough literature review has led us to identification of various risk factors in information security as well as risks associated with information sharing in a supply chain. These risks have to be incorporated in a model establishing relationship between them. For risk assessment, BNs can be used to create information risk profile of a supply chain. Subsection 3.1 provides a brief description of BNs.\n4. Research model: This research study employs a risk assessment model for quantifying information risks in a supply chain. The model consists of the following risk factors: information breakdown factors, information leakage factors and reluctance in information sharing factors. Information breakdown factors are related to information security factors. Similarly Information leakage factors and reluctance in information sharing factors are related to information sharing risk factors. These risk factors are developed based on literature review and expert interviews. Five senior-level managers in supply chain/IT domain were consulted and based on their feedback, the following model structure has been proposed for information risk analysis. Profile of respondents in this study is explained in next Section 4.1. Model also shows the relationship between the various variables (information risk factors). Each node represents a risk factor and direction of the arrow signifies the relationship between them. This diagram showing structural relationship is also known as influence diagram.\n5. Data analysis: The BN deals with the various information risks involved in a supply chain. The nodes in the BN represent risk factors. The BN was tested for a set of data obtained for various input nodes assigning normal distribution to the rest. The BN was modeled using Agena Risk software. Figure 3 shows the distributions and results obtained after simulating the model using input data.\n6. Managerial and theoretical implication: In this research paper an information risk analysis model has been developed using BBN. In Bayesian models, reasoning can be done in both ways. The query variable (net impact on revenue) is affected by information risk factors. The impact of these risk factors on business revenue can be studied by managers using BBN. BBN can be used to disseminate understanding about supply chain information risk factors among industry professionals. For some of the risk factors that cannot be expressed in objective data form, their probability can be inferenced using subjective data (expert's judgment). Propagation analysis using BBN helps managers to update the probability, when new data arrives. In risk analysis, probability updating can be defined as the task of computing all posterior marginal's of non-evidence variables given the evidence.\n7. Limitations: The study was conducted in automotive industry, therefore, the results could be industry-specific in nature. In addition, the study examined only five companies in the Indian automotive industry and for confidential reasons calculation were not performed for a specific firm, thus limiting the generalizability of information risks in this sector. A limitation related to the use of the BN methodology presented in this study is the ability to access the necessary data needed to construct the BNs. Depending on the established relationship, some companies may be reluctant to share risk profile data with their customers.\n8. Conclusion and future scope of research: In this research paper, we have proposed a risk assessment model for supply chain information risk using BNs. As we have discussed the capability of BNs while modeling in uncertain conditions, this provides a prefect platform for analyzing the models providing a more robust method for studying the impact or predicting various risk factors at play. The data analysis shows result obtained for a case study and the changes observed in the values of probabilities when certain data sets are known with full certainty. As mentioned earlier the probability distribution can be made more reliable and accurate if filed data are provided to us. One of the positive feature of the BN is its ability to incorporate new data to change probability distribution. Hence, to improve the predictions made in the case study the model need to be fed with more reliable data, which also remains a limitation for the project. Risk profiles for companies and supply networks in other industries should be examined using the methodology illustrated in this study to determine, if industry dynamics significantly influence supply chain risks.\n",
"role": "user"
},
{
"content": "Purpose: Information sharing enhances the supply chain profitability significantly, but it may result in adverse impacts also (e.g. leakages of secret information to competitors, sharing of wrong information that result into losses). So, it is important to understand the various risk factors that lead to distortion in information sharing and results in negative consequences. Information risk identification and assessment in supply chain would help in choosing right mitigation strategies. The purpose of this paper is to identify various information risks that could impact a supply chain, and develop a conceptual framework to quantify them.\nDesign/methodology/approach: Bayesian belief network (BBN) modeling will be used to provide a framework for information risk analysis in a supply chain. Bayesian methodology provides the reasoning in causal relationship among various risk factors and incorporates both objective and subjective data.\nFindings: This paper presents a causal relationship among various information risks in a supply chain. Three important risk factors, namely, information security, information leakages and reluctance toward information sharing showed influence on a company's revenue.\nPractical implications: Capability of Bayesian networks while modeling in uncertain conditions, provides a prefect platform for analyzing the risk factors. BBN provides a more robust method for studying the impact or predicting various risk factors.\nOriginality/value: The major contribution of this paper is to develop a quantitative model for information risks in supply chain. This model can be updated when a new data arrives.\n",
"role": "assistant"
}
] |
A typology of knowledge management: strategic groups and role behavior in organizations
|
[
"Knowledge management",
"Strategic groups",
"Organizational structures"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The nascent field of knowledge management (KM) has largely focused on three major streams: the nature of knowledge; the organizational and managerial aspects of its implementation; and the ways and means of creating and utilizing knowledge management systems (KMS). The stream of the nature of knowledge has received attention in view of the distinction between explicit and implicit knowledge (Nonaka and Takeuchi, 1995; Day, 2005). A good portion of the research in knowledge management has concentrated on the ways in which organizations can extract and use implicit knowledge.The organizational and managerial implementation and applications of knowledge in organizations have also received attention from researchers (Jashapara, 2005; Gupta et al., 2005). A few models have been proposed depicting the flow of knowledge in organizational settings. Holsapple and Jones (2004, 2005) have advanced a knowledge chain model which portrays primary and secondary activities of knowledge. The primary activities include knowledge acquisition, selection, generation, assimilation, and emission. Secondary activities include knowledge measurement, control, coordination, and leadership. In this model, the combined effort in the two categories of activities benefits the organization by increasing its competitiveness in the environment.The third stream focused on the creation, implementation, and utilization of knowledge management systems (Rubenstein and Geisler, 2003). Viewed primarily as an organizational topic of adoption and adaptation, this stream of research also included the examination of the value accrued from the adoption and utilization of KMS (Muthusamy and Palanisamy, 2004).In much of this prevailing literature there is a tendency to consider the organizational agents who transact in knowledge as fixtures of the organization's processes. The distinction made between explicit and tacit knowledge, originally proposed by Polanyi (1966) has had the effect of concentrating much of the research effort into methods and cultural barriers to attain and extract tacit knowledge. Organizational actors are thus largely perceived as \"standard\" human subjects whose cognition one wishes to target - regardless of their role in the knowledge process in the organization. The motivation to partake in such a knowledge process has also been lightly studied, particularly as part of the research into barriers and facilitators to the effectiveness and performance of KM systems.A more detailed model is thus needed to configure who are the actors transacting in knowledge in the organization: what are the factors that motivate them, and how would such typology impinge upon our understanding of the process of knowledge and the outcomes and benefits that may be derived from it.This paper proposes a model of knowledge transaction and processes in which a simplified three-stages framework links the generation of knowledge with its users. The model is then shown to consist of three types of actors in the processing of knowledge. The final section of the paper discusses the variables that motivate the actors in the processing of knowledge to produce, diffuse, and use knowledge in organizations.\nMethodology: The model proposed in this paper was derived from interviews conducted with 37 managers in three large companies. Two companies are very large global firms in the medical technology market, and the third is a global multi-product, multi-divisional company. The study was conducted in the construction products division of this corporation. This study was aimed at the organizational needs and barriers to the installation of knowledge management systems. The managers from the President to functional vice-presidents and heads of divisions had been actively involved with the establishment of knowledge management systems in their respective organizations.Managers were asked to define and explain the role they play in knowledge management in the firm, the factors that drive them or inhibit their effort in this regard, and the link of this effort to perceived or actual benefits from knowledge in their organizations. This study was extended beyond the usual arguments in favor of knowledge and its obvious benefits (Stewart, 1997). The focus of these inputs to the generation of the model of knowledge processing was therefore the individual actors. Similar models have focused on the approaches to knowledge management (Earl, 2001) and activities that animate the knowledge management systems (Holsapple and Jones, 2004).This study is based on qualitative research, where a model is derived from selected interviews with corporate managers in the three companies in the research sample. Qualitative research is an important and acceptable method when the emphasis of the research is not on testing hypotheses but on the development of a conceptual framework and the identification of critical factors and other key variables (Estabrooks et al., 2005; Geisler, 2000; Michell, 2004). This approach has been used, for example, to study the implementation of knowledge management (Nicolas, 2004), and in various topics of healthcare management (Sandelowski, 2004).Since the seminal work by Campbell and Stanley (1963), it has long been recognized that qualitative research generates findings that may be applicable beyond the selective group thus studied - to other and similar groups and situations. Although these findings cannot be generalized to the extent possible with quantitative research (based on a random sample and statistical analysis), the validity and reliability of qualitative studies is strong enough to engender an emerging model, framework, or theory (Campbell and Stanley, 1963; Kirk and Miller, 1985).\nA model of knowledge processes in organizations: The model that emerges from interviews with managers who were actively involved with knowledge management in their organizations has four modes or stages of knowledge processing: generation, transfer, implementation, and absorption. Table I shows these modes.The mode of generation of knowledge is the effort on the part of actors to assemble knowledge from all available sources, including tacit knowledge embedded in their personal and collective experience. Although developed uniquely from the empirical study described above, this mode is similar to stages or activities in other models (Holsapple and Jones, 2005; Kankanhali and Tan, 2004). In the generation of knowledge the actors translate the assembled array of tacit and explicit knowledge into a form suitable for transfer to others. This step includes verbalizing what they know, creating displays and visual formats, and establishing or following standards for codified content, so that the receivers of this knowledge downstream will be able to decode, understand, and use what they receive (Heinrichs and Lim, 2005; Xiogiannis et al., 2004).The mode of transfer of knowledge is a component of the process in which actors in the organization transfer, share, and diffuse the knowledge they possess. Transfer is undertaken by means of debriefings, audits, reports, statements, and the like. Recipients of such transfer may be other organizational members, databases, knowledge systems, archives, and a variety of external people and entities.This component of knowledge processes has been the subject of some interest by scholars, particularly with regard to the barriers that impede the transfer and sharing of knowledge by organizational members (Grant, 1996; Nonaka and Takeuchi, 1995; Polanyi, 1966; Rubenstein and Geisler, 2003).Transfer processes are only one component of an effective mechanism of sharing and diffusion of knowledge. When actors receive knowledge transferred to them, they still need to implement such knowledge in organizational procedures, policies, codes, activities, methods, and practices. The degree to which implementation is feasible and successful will determine the usefulness of the transfer effort. Current research has not yet established a figure of how much knowledge transferred within the organization is actually implemented in the processes, procedures, and workings of the organization.In the study that generated the model described in this paper, interviewed managers suggested that less than half of the knowledge they receive from all sources is actually implemented in their work. They also tend to distinguish between types of knowledge they receive, transfer, and implement. Interviewees identified five types of knowledge: technical; administrative; experiential; inherent attitudes and beliefs; and strategic. The prevailing view is that technical and administrative knowledge is more easily transferred and implemented, whereas experience, attitudes, and strategic thinking are more difficult to articulate, hence to transfer and implement (Rubenstein and Geisler, 2003).Absorption is the final component of the processing of knowledge. It is more than a nuance of implementation. Knowledge is absorbed in the realities of organizational workings so that it becomes an integral part of the cultural foundations and the shared direction of the organization. Managers who participated in our study had indicated that although some knowledge implemented may also be absorbed, it is difficult to isolate the precise nature and elements of knowledge that reach this point of absorption. Nevertheless, they agree that this component of knowledge processing is vital to the organization.\nTypology of actors in knowledge processing: Any effort designed to understand why people in organizations produce, disseminate, and use knowledge must first address the issue of actual and perceived benefits and impacts derived from knowledge. Organizational members are impacted by knowledge and may benefit from it in several ways. In the study reported here, managers had identified five categories of impacts/benefits. These are shown in Table II.In addition to benefits accrued to them in the form of improvements in what they do and how they perform, organizational members may also gain personal and psychological benefits from processing knowledge. Table III shows examples of such benefits as contributions to satisfying intellectual curiosity and thirst for knowledge, mastery of the environment, and personal growth.But, as Table I shows, there are different processes of knowledge (from generation to implementation and absorptions by users). There are also different benefits and impacts from knowledge, as shown in Tables II and III. Organizational members are therefore involved in all modes of processing knowledge, and are recipients of a variety of potential benefits.There emerge three distinct types or roles of knowledge processing that organizational members play. As actors transacting in knowledge, they will have a unique role at any given time, as: generators, transformers, and users of knowledge.Actors acting in the capacity of generators of knowledge are collecting, assembling, and storing quanta of knowledge. Actors playing the role of transformers prepare the knowledge they possess, share it, diffuse it, and transfer it to other people and organizations. A third type of processor is the users. There are organizational members who implement, absorb, adopt, adapt, and otherwise use and exploit the knowledge they receive through the mechanisms of transfer and diffusion. Table IV shows the types of processors and the factors that motivate them.Multiplicity of roles\nStrategic groups: The types of organizational actors who transact in knowledge can be considered strategic groups, in a mode similar to the Miles and Snow (1978) typology of defenders, prospectors, analyzers, and reactors. In this manner, generators, transformers, and users are distinct categories of approaches to the strategic adaptation to environmental changes (Zahra and Pearce, 1990). When organizational members transact in knowledge in one of the types described in this paper, they engender a strategic approach to address organizational problems of performance and competitive behavior (Holsapple and Jones, 2004). The generation, transformation, and use of knowledge can be considered means to strategic behavior (Connant et al., 1990).Over time the distinct transactions in knowledge become stable behaviors in the life of the organization. Rules, norms, procedures, and perceived relationships to performance and competitiveness become an integral part of the organization's culture and its strategic approach to solving problems. Knowledge is a crucial component of strategic management, hence the need for the organization to support unfettered flow among its members and structural units and work groups.Central to accomplishing the flow of knowledge is the organization's ability to motivate its members to participate in knowledge transactions. The different types of transactions are driven by different incentives. Thus, the organization must tailor its offering of incentives to the specific type of knowledge transaction (Harrigan, 1985).Multiple roles and motivators\nConclusions: This paper proposed a typology of people and organizations who transact in knowledge. The paper identified three distinct types of behavior: generators, transformers, and users. These types have unique factors that motivate and drive the organizational actors who transact in knowledge. The typology contributes to the analysis of the management of knowledge in organizations in the following ways. First, the typology helps to distinguish between the roles that people and organizations play in managing knowledge, and the complexities associated with these roles and their behavioral manifestations. Second, the typology provides an analytical framework in which organizational strategies and policies regarding knowledge management can be better studied and explained. Thus, policies on incentives to the generation, transfer, and use of knowledge can be now examined through the prism of types of transaction in knowledge and the roles played by organizational actors. Finally, the typology offers a conceptual framework in which organizations may tailor their policies and procedures regarding KMS and the links between knowledge transactions and organizational performance.The limitations of this study and its findings are threefold. First, contextual variables such as size of the organization may contribute to difficulties in distinguishing among the various transactors in knowledge. In smaller-sized organizations, members may interchangeably engage in the generation, transfer, and use of knowledge. The distinct roles that make the typology may thus be masked by the multitasks assumed by members of small and emerging organizations.A second limitation would be in view of the effect of the structure of the organization. For example, in a structure by functions (referred to as \"departmentation by function\" or \"unitary form\" of organization) the units or individual members will transact mostly in knowledge limited to their functional area. Conversely, in a structure by product, customer, or region, the members will transact in a much more varied spectrum of knowledge as generators, transformers, and users.Third, the model may be limited by the degree to which the roles of transactors in knowledge impact the utility of knowledge transactions and the performance of members and the organization in which such knowledge is transacted. Although the latter limitation is true for any typology, in this case it is important to identify the nature of the utility of the typology of knowledge transactions. One reason for this need is the link between the types of transactions and the management of knowledge systems. Another reason is the link between the typology and strategies, policies, and procedures regarding knowledge management and knowledge systems.Further research would empirically examine the linkages between the types identified in this paper and value created from knowledge transactions. Another research topic would be the study of organizations as distinct types of knowledge transactors. For example, future research may study universities in their roles as generators and transformers of knowledge, and the benefits and value they provide to society. Similarly, business organizations may be studied in their roles as generators, transformers, or users of knowledge and the relationship between the effectiveness of the role with their competitive position and commercial success.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to develop a typology of people and organizations who transact in knowledge.\nDesign/methodology/approach: - Based on structural interviews with 37 managers in three large manufacturing companies a model is proposed which describes the processing of knowledge in organizations.\nFindings: - Four stages are identified: generation, transfer, implementation, and absorption. Similarly, three types of transactors in knowledge are also identified: generators, transformers, and users. The findings from the interviews are the different motives that animate the different transactors in knowledge, and the distinct behavioral roles that these transactors assume in their organizations.\nOriginality/value: - This paper contributes to the literature by proposing a new way of classifying the roles of people and organizations in their transaction in knowledge.\n",
"role": "assistant"
}
] |
Does patient satisfaction affect patient loyalty?
|
[
"Patients",
"Hospitals",
"Quality assessment",
"United States of America",
"Customer satisfaction"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Managers in diverse industries have long been interested in the link between consumer satisfaction with a product or service and the decision to purchase (Keiningham et al., 2007). Marketing researchers have analyzed consumer behavior in sectors such as airlines, banking, gasoline, personal computers and other consumer products, focusing on customer satisfaction effects on market share, cash flow, profits and stockholder returns.Satisfaction's impact on hospital choice is particularly important. Research shows a link between patient satisfaction and healthcare quality (Naidu, 2009). Hospitals with higher patient satisfaction have lower mortality and better process measures (Jaipaul and Rosenthal, 2003; Edlund et al., 2003; Gesell et al., 2005). Thus, evidence that patient satisfaction actually predicts provider choice suggests a pathway through which individuals naturally gravitate toward higher-quality care, despite the difficulties inherent in evaluating service quality. Satisfied patients are also more likely to adhere to recommended treatment (Finkelstein et al., 1999) and less likely to sue for malpractice (Stelfox et al., 2005). Yet, even though patients say they would consider changing hospitals in response to satisfaction information (Boshoff and Gray, 2004; Sofaer et al., 2005), few researchers have explored whether such intent translates into actual decisions. One study examined patients from a single hospital and found a small, positive relationship between patient satisfaction and the probability of returning to the hospital for treatment (Garman et al., 2004). As the authors points out, however, its reliance on data from a single hospital and from only those patients who responded to the hospital's satisfaction survey might limit its generalizability. Consequently, we use a unique dataset to investigate whether patient satisfaction affects decision to return. We match information from three sources:1. Patient satisfaction data on 678 hospitals nationwide from Press Ganey, a healthcare survey research and quality improvement firm.2. Publicly-available information on the frequency with which end-of-life patients return to seek treatment at the same hospital - their loyalty - from the Dartmouth Atlas project.3. Publicly-available information on process-based quality measures that affect patients' immediate hospital environment and market characteristics from the Dartmouth Atlas and Medicare cost reports.Based on a regression analysis, we estimate satisfaction effect on patient loyalty, holding constant other quality measures and hospital and market characteristics.\nMethods: Data and study sample\nResults: Table II presents results from our three regression models. The left column reports results from the first model, which specifies loyalty as a function of a univariate satisfaction measure. It shows that patient satisfaction is significantly positively correlated with end-of-life patients' propensity to return to a hospital. According to Table II, a one-standard-deviation increase in a hospital's average patient satisfaction leads to a 0.1 standard-deviation increase in loyalty, holding other factors constant. High intensity terminal patient care and poor physician coordination are both significantly negatively associated with loyalty. For example, according to the model, a one-standard-deviation decline in a physician's end-of-life care coordination score leads to a 0.28 standard-deviation decline in loyalty, holding constant other patient satisfaction measures.Other hospital characteristics also affect patient loyalty (not reported in Table II). Holding all else constant, teaching and government hospitals have lower patient loyalty than their counterparts. Hospitals in counties with ten or more hospitals, not surprisingly, also have lower loyalty than residents with fewer choices; patients are less likely to revisit a hospital when there are several alternatives. Table II's middle column reports results from the second model, which specifies loyalty as a function of satisfaction's four principal components. Results from Model 2 show that satisfaction with process is the most economically important patient care experience for determining loyalty; satisfaction with facilities is second (although not statistically significant at the 5 per cent level). A one-standard-deviation increase in patient satisfaction with hospital process leads to a statistically significant 0.232 standard-deviation loyalty increase, holding other factors constant. This is consistent with the strong relationships between loyalty and process quality measures reported above. Table II's right-hand column reports results from the third model, which specifies loyalty as a univariate satisfaction function, but allows satisfaction's effect to vary for high- and low-satisfaction hospitals. Table II confirms that the \"insulation\" hypothesis holds for hospital services. Satisfaction's effect on loyalty is positive and statistically significant for hospitals in the bottom half of the patient satisfaction distribution. However, for high-satisfaction hospitals, satisfaction's incremental effect on loyalty is insignificantly different from zero.\nConclusions and recommendations: We show a statistically significant link between patient satisfaction and loyalty. This result persists, even holding constant hospital characteristics and process-based quality measures likely to be highly salient to satisfaction. Although this result is statistically significant, it is relatively small in magnitude. In practical terms, a one-half point increase in satisfaction on a five-point Likert scale - a large change - leads to a 1.2 per cent increase in end-of-life patient loyalty. A one-half point change in satisfaction is approximately a three standard deviation change, which leads to a 0.3 standard deviation change in loyalty - approximately 1.2 per cent.For several reasons, this basic result may understate patient satisfaction's importance to loyalty in hospital markets. First, the effect of satisfaction with certain hospitalization experience is larger than overall satisfaction's effect. An increase in satisfaction with admissions, discharge and other processes, comparable in magnitude to a one-half point increase in overall satisfaction, leads to a 2.5 per cent increase in loyalty. This is because a three standard deviation change in satisfaction with process leads to a 0.696 standard deviation change in loyalty - approximately 2.5 per cent.Second, end-of-life patients represent only a small (although significant) share of hospital use. In 2003, there were 34,783,000 admissions to non-federal short-stay inpatient hospitals; each had a 6.6 day average stay. This implies that total inpatient days that year were 229,567,800 (Centers for Disease Control, 2006). According to calculations based on the Dartmouth Atlas (not in any table), the average inpatient days consumed by end-of-life patients from 1999-2003 was 22,256,355 per year, or 9.7 per cent of the total.Third, satisfaction may be a less important determinant of end-of-life patients' propensity to return than ideal measures based on broader patient populations. Because hospital choice is for most patients so infrequent, constructing a broad-based hospital loyalty measure is difficult, and other loyalty measures lack external validity.Fourth, our models estimate satisfaction's effect, holding constant process-based quality measures. This assumes that patient satisfaction should be assessed under the assumption that satisfaction improvements are generally achieved without affecting process quality, which may be incorrect because our analysis suggests that the most effective way to increase satisfaction is by targeting key quality improvement processes.Our study has significant limitations. Most importantly, the hospitals we analyze - for which we have Press Ganey satisfaction data - do not represent the US hospital population. Our estimates are therefore most directly applicable to large, non-profit, teaching hospitals in competitive markets. Additionally, our estimates of satisfaction's effects on loyalty may represent an unobserved third factor that is positively correlated with both, leading the estimates to overstate satisfaction's true causal impact. Although we correct for the possibility that hospitals with long-stay patients may have greater satisfaction and mechanically higher loyalty, omitted variable bias from another source is possible. Future research might seek to overcome these limitations and investigate the extent to which satisfaction data can improve healthcare quality.\n",
"role": "user"
},
{
"content": "Purpose: - This paper aims to investigate how patient satisfaction affects propensity to return, i.e. loyalty.\nDesign/methodology/approach: - Data from 678 hospitals were matched using three sources. Patient satisfaction data were obtained from Press Ganey Associates, a leading survey firm; process-based quality measures and hospital characteristics (such as ownership and teaching status) and geographic areas were obtained from the Centers for Medicare and Medicaid Services. The frequency with which end-of-life patients return to seek treatment at the same hospital was obtained from the Dartmouth Atlas. The study uses regression analysis to estimate satisfaction's effects on patient loyalty, while holding process-based quality measures and hospital and market characteristics constant.\nFindings: - There is a statistically significant link between satisfaction and loyalty. Although satisfaction's effect overall is relatively small, contentment with certain hospitalization experience may be important. The link between satisfaction and loyalty is weaker for high-satisfaction hospitals, consistent with other studies in the marketing literature.\nResearch limitation/implications: - The US hospitals analyzed are not a random sample; the results are most applicable to large, non-profit teaching hospitals in competitive markets.\nPractical implications: - Satisfaction ratings have business implications for healthcare providers and may be useful as a management tool for private and public purchasers.\nOriginality/value: - The paper is the first to show that patient satisfaction affects actual hospital choices in a large sample. Because patient satisfaction ratings are also correlated with other quality measures, the findings suggest a pathway through which individuals naturally gravitate toward higher-quality care.\n",
"role": "assistant"
}
] |
Healthcare managers' construction of the manager role in relation to the medical profession
|
[
"Discourse analysis",
"Medical profession",
"Healthcare management",
"Manager role",
"Organizational role theory",
"Sickness certification practice"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nBackground: To have a manager responsibility in healthcare organisations was long regarded as a mere administrative task in relation to the medical profession (von Knorring, 2012; Grey, 1997), with the main function to provide the necessary facilities and resources for physicians to \"get on with their work\" (Harrison and Pollitt, 1994, p. 36). During the last 30 years, management of healthcare organisations however, has undergone major changes. Improvements in medical technology and treatment methods have made it necessary to administer the increasingly complex healthcare processes more effectively, to prioritise in the use of limited resources and, from a societal perspective, to safeguard aspects of quality and equity in the care offered to an aging population (von Knorring, 2012; Kuhlmann, 2006; Kuhlmann and Saks, 2008; Kaissi, 2005; Salter, 2007). Within the scope of this development, the manager position in healthcare organisations in many Western countries has formally been strengthened through a number of political reforms resulting in a tighter control of physicians' work (Mark, 1991; Kuhlmann et al., 2011; Numerato et al., 2011). The Griffith's report, presented in the UK in 1983, can be regarded as the start of this new era of management of healthcare organisations. It concluded that general management was needed in healthcare to ensure effective leadership and clear accountability for decision making (Griffiths, 1983). The same year as the Griffith's report was presented in UK, the Swedish government passed the first law that opened the way for general management and for other professionals than physicians to become managers in healthcare organisations (1982). Today, the position of clinical department manager (CDM), i.e., to be the direct manager over physicians' clinical work, is open to any profession, with the only exception that legal medical responsibility must rest with a physician (1982).\nTheoretical starting point: To understand how managers construct the manager role in relation to the medical profession, we used theories on discourse (Johnstone, 2008) and managerial identity work (Svenningsson and Alvesson, 2003; Carroll and Levy, 2008) together with a theoretically developed role concept (Reed, 2001; Reed and Bazalgette, 2006; Sandahl et al., 2010) to inform our thinking. In the following section, we will describe the rationale for this approach.\nMethods: Following the theoretical starting point described above, the analysis of the empirical findings in the paper is based on the assumption that how we categorise and define the roles of others and ourselves is continuously created and recreated in social interaction (Johnstone, 2008; von Knorring, 2012). We also assume that this continuing categorisation of self and others forms rather stable elements (Foucault, 1980; Danermark, 2001), \"temporary answers\" (Svenningsson and Alvesson, 2003), or discourses (Johnstone, 2008) that can be studied. How managers categorise and define the manager role at a specific time in a specific organisational context can be regarded as such a rather stable element, which can be studied.\nResults: Our analyses of the data from the 38 interviews showed that the managers used two different discourses, a management-based and a profession-based discourse, to construct the manager role in their organisations. Below, we describe the characteristics of those discourses and how they were used in the managers' statements.\nDiscussion: Few of the managers in this study used a management-based discourse to construct the manager role in relation to the medical profession. Instead, a profession-based discourse dominated and managers frequently used the attributes \"physician\" or \"non-physician\" to categorise themselves and other managers in their manager roles. The two discourses clearly differed with regard to the power and scope of action available to managers in relation to physicians. The management-based discourse framed managers as powerful subjects with relevant knowledge, entitled to manage physicians. In contrast, the profession-based discourse, which clearly dominated among the managers, seemed based on an informal hierarchical system of professions (Abbott, 1988) in which physicians and medical knowledge are at the top and nurses and other healthcare professions are situated further down, entailing very little acceptance for the manager role, regardless of the manager's profession.\nConclusions: Despite the increase in managerial influence in healthcare organisations, as defined by political reforms in many Western countries, this study indicates that healthcare managers are not very strong representatives of the ideology of managerialism in relation to the medical profession - at least not on the micro level in their organisations. Our findings show that a profession-based discourse dominates in how managers construct the manager role in relation to the medical profession. Many of the interviewed managers, including the top-level executives, regarded authority and legitimacy in the manager role as being strongly related to profession rather than to managerial position. The dominating profession-based discourse served to reproduce the power and status of physicians within the organisation, thereby rendering the manager role weaker than the medical profession for both physician and non-physician managers. The results suggest that there is a need to address the organisational conditions for managers' role taking in healthcare organisations, because ambiguity in managerial role-taking may have a negative impact on the working conditions of all employees (i.e. managers, physicians, and other healthcare professionals), as well as on the quality of the care that is provided. In order to expand our understanding of managers' role taking in healthcare organisations, further studies are also needed to explore the impact of gender, managerial level, and underlying profession on how managers construct the manager role in relation to physicians.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore how healthcare managers construct the manager role in relation to the medical profession in their organisations.\nDesign/methodology/approach: In total, 18 of Sweden's 20 healthcare chief executive officers (CEOs) and 20 clinical department managers (CDMs) were interviewed about their views on management of physicians. Interviews were performed in the context of one aspect of healthcare management; i.e., management of physicians' sickness certification practice. A discourse analysis approach was used for data analysis.\nFindings: Few managers used a management-based discourse to construct the manager role. Instead, a profession-based discourse dominated and managers frequently used the attributes \"physician\" or \"non-physician\" to categorise themselves or other managers in their managerial roles. Some managers, both CEOs and CDMs, shifted between the management- and profession-based discourses, resulting in a kind of \"yes, but [...]\" approach to management in the organisations. The dominating profession-based discourse served to reproduce the power and status of physicians within the organisation, thereby rendering the manager role weaker than the medical profession for both physician and non-physician managers.\nResearch limitations/implications: Further studies are needed to explore the impact of gender, managerial level, and basic profession on how managers construct the manager role in relation to physicians.\nPractical implications: The results suggest that there is a need to address the organisational conditions for managers' role taking in healthcare organisations.\nOriginality/value: Despite the general strengthening of the manager position in healthcare through political reforms during the last decades, this study shows that a profession-based discourse clearly dominated in how the managers constructed the manager role in relation to the medical profession on the workplace level in their organisations.\n",
"role": "assistant"
}
] |
Coping with mobile technology overload in the workplace
|
[
"Information overload",
"Job satisfaction",
"Coping model of user adaptation",
"Coping theory",
"Interruption overload",
"Mobile information and communication technologies (MICTs)"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Technology overload, including mobile application overload, information overload, communication and interruption overload, has become increasingly prevalent in the digital workplace. A recent survey conducted by RingCentral found that: more than 70 percent of employees say their communication volume is a challenge to the fulfillment of their work; 68 percent of employees toggle between mobile applications up to ten times per hour; 31 percent of employees indicate toggling causes them to lose their train of thought; and 56 percent of workers find searching for information from multiple sources disruptive (Brumberg, 2018). Workplace information and communication technologies are meant to help employees work more effectively and keep team members more focused in a collaborative way so that they can be more productive. However, employees are becoming more frustrated due to the extent to which they are overloaded by technology (Brumberg, 2018).\n2. Literature review: 2.1 Technology overload\n3. Hypothesis development: In this section, grounded on the CMUA framework and coping theory, we explain our research model and hypothesis development in detail. The proposed research model is shown in Figure 1. In the research model, we explore the relationships among the technology overload factors (i.e. information overload and interruption overload), coping strategies (i.e. information processing timeliness and job control assistant support) and job satisfaction. Table I lists the definition of each construct used in this study.\n4. Methodology: In this study, we went through several procedures to empirically develop and validate the instrument for technology overload factors and coping strategy factors. First, following the literature review, we developed the measurements for the constructs, which we pre-tested through a series of procedures including card sorting and ranking (Moore and Benbasat, 1991). We then conducted a survey to validate the instrument with 178 valid data points. We detail those procedures below.\n5. Data analysis and results: 5.1 Validating the measures\n6. Discussion, implications and future research: In this section, we summarize the key findings and potential contributions and implications of this study. The future research directions are also discussed.\n7. Conclusion: The dark side of IT has aroused increasing attention in recent years (Tarafdar et al., 2015). Some research has also been conducted to examine the increasing significant negative effects of social media (Mantymaki and Islam, 2016; Fox and Moreland, 2015). However, the negative effects resulting from the use of MICTs (including various mobile social media applications and devices) have been paid little attention by academics. Specifically, with the prevalence of massive amounts of information, humans are experiencing increasing levels of overload. With the increasingly intimate relationship between humans and MICTs, the technology overload effects may be reflected in new forms, such as techno-dependency. However, few studies have been conducted to examine this phenomenon. The context-dependent technology overload effects have also not been identified or illustrated. Responding to the call of this special issue, this study constitutes the starting point for investigations into the phenomenon of technology overload. Future research needs to consider the different dimensions of technology overload and their corresponding influences on both individual performance and organizational behavior in different industry backgrounds.\n",
"role": "user"
},
{
"content": "Purpose: The overload effects associated with the use of mobile information and communication technologies (MICTs) in the workplace have become increasingly prevalent. The purpose of this paper is to examine the overload effects of using MICTs at work on employees' job satisfaction, and explore the corresponding coping strategies.\nDesign/methodology/approach: The study is grounded on the cognitive load theory and the coping model of user adaptation. The overload antecedents and coping strategies are integrated into one model. Theoretical hypotheses are tested with survey data collected from a sample of 178 employees at work in China.\nFindings: The results indicate that information overload significantly reduces job satisfaction, while the influence of interruption overload on job satisfaction is not significant. Two coping strategies (information processing timeliness and job control assistant support) can significantly improve job satisfaction. Information processing timeliness significantly moderates the relationships between two types of overload effects and job satisfaction. Job control assistant support also significantly moderates the relationship between interruption overload and job satisfaction.\nPractical implications: This study suggests that information overload and interruption overload could constitute an important index to indicate employees' overload level when using MICTs at work. The two coping strategies provide managers with effective ways to improve employees' job satisfaction. By taking advantage of the moderation effects of coping strategies, managers could lower employees' evaluation of overload to an appropriate level.\nOriginality/value: This study provides a comprehensive model to examine how the overload resulting from using MICTs in the workplace affects employees' work status, and how to cope with it. Two types of overload are conceptualized and corresponding coping strategies are identified. The measurements of principal constructs are developed and empirically validated. The results provide theoretical and practical insights on human resource management and human-computer interaction.\n",
"role": "assistant"
}
] |
Engineering education and the development of competencies for sustainability
|
[
"Competences",
"Sustainability",
"Engineering education",
"Higher education",
"Sustainable development",
"Problem-based learning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Education plays, nowadays, an important role in promoting sustainable livelihoods. The education for sustainable development (ESD) can be defined as the education that aims at preparing individuals to deal with problems that threaten sustainability (economic, social and environmental problems). Beyond the knowledge and understanding of environmental and social issues, education for sustainability also aims to affect attitudes, change behaviors and promote competencies needed to shape a sustainable future (UNESCO, 2017).\n2. Literature review: The review of the literature can be characterized by the application of a research strategy, analysis of the results found and organization of studies that are relevant to the research topic. Aveyard (2014) points out that one of the objectives of the literature review is to bring together in a single topic, a representative part of the available literature. In this way, this type of research can have two purposes. The first one has a historical function, that is, to observe the evolution of a theme over time. The second is related to the search for the state of the art, which corresponds to identify the unique contributions of different studies so far.\n3. Analysis and discussion of results: According to the questionnaire applied to the specialists, the results found indicate that all the key competences for sustainability found in the literature are relevant. Table V sets out the descriptive statistics related to each competence considered as an important condition to be developed.\n4. Conclusions: Sustainability is a very complex theme, which involves different stakeholders and interest conflicts. Engineering students have increasingly faced problems related to this theme. Inserted in this thematic, this study sought to make explicit the main competencies to be developed by this public. As a result, this research found eight relevant attributes, which are: systemic thinking, problem-solving ability (integrated resolution), ability to work in interdisciplinary group (collaboration), critical thinking, normative competence, competence of self-knowledge, strategic competence and contextualization and vision of the future (anticipatory).\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to deal with the skills for sustainable development in engineering courses. The main objective is to identify which competences should be developed in these courses to contribute to the resolution of conflicts related to sustainability, as well as the means used by universities for their development.\nDesign/methodology/approach: The research is divided into two strands: a theoretical and an empirical one. Concerning the theoretical part, the webibliomining method was adopted, to select the articles that were relevant to the theme, and 22 works were selected. Regarding the empirical study, structured questionnaires were applied to 30 specialists in the field, aiming at validating the competencies found in the literature.\nFindings: In the literature, eight main competencies required to the engineering professional, related to sustainability, were found. The empirical study showed that some competences are considered more relevant than others by the interviewees.\nPractical implications: The competencies verified in the present study and the relative relevance between them found in the empirical study can contribute to the engineering courses that wish to develop the sustainable behavior of the engineers in their professional or daily life.\nOriginality/value: The originality of the work consists in presenting to engineering students and teachers as well as professionals of the market and the government the competencies that can contribute to the formation of an engineer capable of bringing adequate solutions to the conflicts of sustainability in the twenty-first century. This paper also presents learning teaching methodologies indicated for this purpose. No other works were found with these results.\n",
"role": "assistant"
}
] |
A closed-form equation for steady-state availability of cold standby repairable k-out-of-n: G systems
|
[
"Redundancy",
"Steady-state availability",
"K-out-of-n system",
"Cold standby"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Nomenclature\n1. Introduction and literature review: Reliability and availability are two important measures to evaluate the operation of complex systems. They play significant roles in product quality and services. While reliability is defined as the probability of a system or a component to perform its required functions for a specified time, availability is usually concerned with repairable systems and is defined as the probability of the system to work at a specific time, apart from the crashes and repairs it had before. Jolly and Singh (2014) proposed a tactical approach to deal with the problems raised due to low availability of repairable systems. Kumar et al. (2013) developed a new way to obtain reliability, availability, mean time between failure, etc., for any repairable industrial system. Besides, one of the most important indices used to maintain and repair industrial systems with repairable components is the steady-state availability (Carpitella et al., 2018). Steady-state availability can be defined as the limit of the availability when time approaches to infinity. This index, which is also called asymptotic, stationary or long-term availability, is a useful tool for reliability engineers and managers (Zaidi and Goyal, 2014; Gupta and Tewari, 2011).\n2. System description and assumptions: Wang et al. (2012) calculated the reliability of a non-repairable cold standby k-out-of-n system with k active and (n-k) inactive components. Figure 1 shows the block diagram of this system. In this configuration, the first k components (X1, ..., Xk) are in active mode and the rest (Xk + 1, ..., Xn) are in the cold standby state. Herein, a new assumption is added to the problem to include repairable components. When any of the active components in the system fails, then one of the standby components will become active and replace on the failed one; then, the failed component will be repaired. If only another active component in the system fails, then the failed component of the system will be immediately replaced by the standby component, if possible. In this case, the newer failed component will be repaired if only no previous failed components exist in the system. In addition, it is assumed that all the active and cold standby components (when they become active) work independently and their failure and repair rates are governed by the exponential probability distribution function.\n3. Markov method: One of the most well-known and widely used methods for calculating the reliability and availability of repairable and non-repairable systems is the Markov method. This method is used for complex and repairable systems when the system failure as well as the repair follows an exponential distribution (Billinton and Allan, 1992; Hoyland and Rausand, 1994). The Markov method provides an appropriate venue to analyze system reliability and to compute relevant indices when the components are repairable and the failure and repair rates are constant.\n4. Numerical examples: To show the applicability of the proposed formula (29), two sets of examples are solved in this section.\n5. Discussion: In this paper, a closed-form equation (Equation (29)) was derived to obtain the long-term availability of a repairable cold standby redundant k-out-of-n system. Referring to Figures 3 and 5, it was shown that when the repair rate increases, then the system is more reliable when the number of redundant components is the lowest of its own, i.e. just one redundant component. As such, the volume is less and the system becomes less expensive. Industrial companies that use this type of systems become more available and efficient if they improve their maintenance by increasing their repair rates.\n6. Conclusions: The purpose of this study was to derive a closed-form equation to calculate precisely the steady-state availability of a cold standby redundant system with k active and (n-k) cold standby repairable components. Assuming primary and cold standby components (when they are active) were independent and identical with exponential distributions for their failures and repairs, the Markov method was employed to this aim. Using two sets of numerical examples, it was shown that as the repair rate increases, the steady-state availability of the system increases.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to derive a closed-form expression for the steady-state availability of a cold standby repairable k-out-of-n system. This makes the availability calculation much easier and accurate.\nDesign/methodology/approach: Assuming exponential distributions for system failure and repair, the Markov method is employed to derive the formula.\nFindings: The proposed formula establishes an easier and faster venue and provides accurate steady-state availability.\nResearch limitations/implications: The formula is valid for the case when the probability density function of the component failure and the repair is exponential.\nOriginality/value: The Markov method has never been used in the literature to derive the steady-state availability of a cold standby repairable k-out-of-n: G system.\n",
"role": "assistant"
}
] |
Job insecurity and motivation among women in Nigerian consolidated banks
|
[
"Employment protection",
"Motivation (psychology)",
"Employee attitudes",
"Women workers",
"Banks",
"Nigeria"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In recent times, increased economic dependence between countries, rapidly changing consumer markets and escalated demands for flexibility have necessitated various organizations such as the banking industry to embark on radical consolidation processes for survival. These adjustments have become imperative for such organizations to remain relevant in the present unpredictable economic environment. Being profitable involves embarking on a process that can either increase their gains or decrease their costs by reducing the number of employees. This is implemented through actions such as outsourcing, privatization, mergers and acquisitions which is often combined with personnel reduction through lay offs, offers of early retirement and increased utilization of subcontracted workers. Consequently, in making merger or acquisition decisions, management are often concerned with financial advantages or product synergy (Robbins, 1998) to the detriment of workers' survival. Although differences exist in the strategies adopted by organizations, the effects on organizations and the workforce vary.The purpose of this study is to examine how the aspects of skill variety, autonomy and feedback of job characteristics and organizational justice will predict perceived job security and motivation of women. It also sought to determine if significant differences existed between women employees of merged and acquired banks perceive job security and motivation. Although mergers and acquisition is a new programme in most organizations in Nigeria, its implementation has led to mass layoffs and redundancies especially in the banking industry. During the consolidation process, 25 banks were approved to continue with their banking functions by the Central Bank of Nigeria, which is the government approved regulating body for the consolidation process. Eight were merged and ten were acquired by other banks within Nigeria. Seven other banks were neither merged nor acquired. A total of 13 had their license withdrawn. The implication is the loss of jobs for 45,000 workers in the affected banks (Okafor, 2007). There is a ratio of 35 females to 55 males in the banking industry in Nigeria. For survivors in such situations, fear of the unknown and anticipation of the worst in the face of uncertainty may prevail. As a result, willingness or motivation to carry on their duties may diminish.\nLiterature review: Research in Europe has shown that the historic surge of consolidation and combination is occurring in the face of strong evidence that mergers and acquisitions, at least over the past 35 years, have hurt more than helped companies and shareholders (Zweig, 1995; Troiano, 1999). In addition, Davy et al. (1991) reported that organizations encounter problems such as inability to change after embarking on this exercise. One common effect of these strategies on the workforce is that there is feeling of insecurity and organizational morale is diminished (Armstrong-Stassen, 1993). Could it be that consolidation processes do not yield anticipated results for both organizations and employees because issues concerning security and motivational needs of survivors are not being addressed?In particular, restructuring processes may make career development become an uphill task for women. Women tend to have different needs and orientation to work. Satisfaction with one's job in terms of condition, policies and practices is very crucial, because it has been found to be positively related to career growth (Chovwen, 2004). Furthermore, in her study of the realities of Nigerian Organizational Health Policies and Practices for female employees, Udegbe (2003) reported that some employment and placement practices in the workplace are such that women are not allowed to take jobs that may be hampered by their \"reproductive health concerns\".In addition, studies have shown that despite the growing evidence that firms with high numbers of women executives tend to outperform their industry (Adler, 2001), women continue to leave their corporate positions voluntarily and involuntarily, and more women than men are leaving corporations (Catalyst, 2001). All these have implications for the career development and retention of women in the workplace and particularly in the male dominated occupations.As revealed by Webster (2001), the processes of restructuring are not positive for employees and this may be worse for women. According to him, employees are often informed of redundancies after investors have been told; public announcement made through the media and this creates an atmosphere of exclusion and mistrust. Subsequently, reactions of the employees who remain may dictate the organizations effectiveness (Brockner, 1992). It may be appropriate to suggest here that employees' perception of job insecurity and level of motivation may go a long way in determining the achievement of the goals of consolidation.Job insecurity and motivation in merged and acquired banks\nMethod: Participants\nResults: Predictors of job insecurity\nDiscussion: As mentioned earlier, restructuring processes are relatively new in Nigeria and, as a result, this study was carried out to investigate the influence of perceived insecurity and motivation among survivors in the organizations that have embarked on downsizing through merger and acquisition. According to literature, survivors of organizational downsizing are prime candidates for insecurity, having seen their colleagues made redundant through consolidation practices and policies. In addition, studies have shown that all jobs are seen as less secure after downsizing and there is decline in trust and commitment (Armstrong-Stassen and Latack, 1992). In line with the above, the findings of the study revealed significant effects of independent variables on the dependent variables. Specifically, the results of hierarchical regression indicated a significant joint effect of job characteristics and organizational justice on the threat of job loss (a dimension of job insecurity) for survivors. Individual contributions highlighted the effect of job position, feedback and procedural justice for the threat of job loss. Further analysis showed a significant effect of organizational justice on the perceived powerlessness (a dimension of job insecurity). Specifically, procedural and distributive justices were implicated as a significant individual contributor to job insecurity. Generally, it was observed that organizational justice is a salient factor to be considered in the process of restructuring. First, these results confirmed the findings of Broackner (1992) and Ciancio (2000) who stated that job security involves an employee's perceived control over job-related issues and changes in their job characteristics. Furthermore, it lend credence to the findings of Greenberg (1987) who stated that employers need to pay close attention to how people are treated in terms of unfair procedural and distributive actions during any major transition or organizational change. Thus, it is not surprising to find out that in both merged and acquired banks, perceived job insecurity for women was significantly influenced by procedural dimension of organizational justice. It is important for organizations especially those involved in merger and acquisition processes to note that survivors' perception of fairness of procedures adopted in decision making will go a long way in determining the effectiveness of consolidation processes. Thus, there is the need for organizations to provide survivors with adequate information regarding consolidation procedures and practices. This means for women to perceive that their jobs are secure in the new work environment, they must perceive that job characteristics do not change and that there was fairness in decisions adopted in layoff procedures.For perceived motivation, while all the independent variables jointly predicted motivation, feedback and skill variety exerted significant individual influences. In addition, while all the independent jointly determine perceived motivation for women in both merged and acquired banks, results indicated significant individual contributions of job characteristics. This means that for women to be motivated, it is important that they perceive changes in skill variety and feedback as favourable. If this is otherwise perceived, they may not be motivated to follow through with these changes and this may affect the organization's productivity. These findings confirmed the work of Broadbent (2002) which revealed that employees who are faced with loss of autonomy, job expertise, and reduced opportunity to participate in the changes experience an increased feeling of isolation from management. In a study carried out by Campbell, it was found out that 73 per cent of managers perceived a decrease in morale following redundancy compared to 42 per cent who had not experienced redundancy. Similarly, 54 per cent of managers experienced a decrease in motivation following redundancy compared to 33 per cent from the no redundancy group. In situations where survivors are required to take up the duties of those laid off, they need additional skills and resources to cope with the increased variety of their jobs. Consequently, this may result in lack of motivation in the face of merger and acquisition.Finally, it was found out that there was no significant difference between employees in merged banks and those in acquired banks in terms of perceived insecurity and motivation. This result could be interpreted to mean that the merger and consolidation process has not been significantly beneficial (Zweig, 1995; Troiano, 1999). In addition, Davy et al.(1991) have reported that organizations encounter problems such as inability to change after embarking on this exercise. This finding strengthens the argument of this study which was also confirmed by the findings of Armstrong-Stassen (1993) that one common effect of these strategies of merger and acquisition is on the workforce who perceive a feeling of insecurity and diminished organizational morale. All these have revealed that changes in terms and conditions policies, pay and grading structure, job description may have significant detrimental effects especially for women.\nConclusion and implications: An interesting finding of this study is the significant joint influence of job characteristics and organizational justice on perceived motivation and insecurity. This means when women perceive difficulties arising as a result of job characteristics and unfair procedures, they may feel job security is threatened and level of motivation may be affected. These individuals could just as well be motivated to oppose change as any other group (Holmstrand and Stens, 2001). Organizations risk resistance to change from survivors who see layoff as unfair or resulting in uncertainties. When this happens, the purpose and importance of consolidation is defeated. According to Mishra and Speitzer, high levels of both trust in management and perception of justice in the way downsizing is implemented and managed will increase the constructiveness of survivors' responses. It is therefore important for managers to be aware of the predictors of insecurity and motivation especially among women as they organize into strong coalitions to push for their rights when implementing changes. They may put in place human resources policies that will help cushion the effect of changes unanticipated by women. For example, it was observed during the period of data collection that two banks anticipated negative reactions from employees and prepared measure to cushion the effect of restructuring. In these organizations, existing staff were assured that they would not lose their jobs within the next 12 months of the consolidation process. A voluntary exit plan was put in place for staff that desire to pursue their personal career. Under this plan, a generous financial package was offered and seminars were organized to prepare those that would be laid off to face future challenges. In addition, special seminar and workshops for women regarding entrepreneurial skills, financial management and leadership could be organized before the layoff process commences. All these would ensure, first, that workers and, in particular, women keep their minds on their tasks of meeting customers' needs, which was one of the goals of merger and acquisition and this helped boost morale. Second, it would prepare them to face the challenges ahead. Unless changes are managed in a structured, systematic and constructive manner, organizations will remain far from realizing the benefits of restructuring and downsizing.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to examine the predictors of perceived job insecurity and motivation among 286 female participants in merged and acquired banks in two capital cities of South West Nigeria.\nDesign/methodology/approach: - The participants in this survey were accidentally sampled for convenience of participants and ease of administration because of the tight work schedule. Data collected were analysed using regression and t-test statistics.\nFindings: - The results of the research indicated joint significant influence of job characteristics and organizational justice in women's perceived threat of job loss and motivation. Specifically, organizational justice (procedural justice) exerted a strong influence on the dependent variables for women in both acquired and merged banks.\nResearch limitations/implications: - Use of accidental sampling is a limitation. Further, studies should adopt other sampling techniques that may capture a significant sample of the population.\nPractical implications: - These findings implied that unless changes are managed in structured, systematic and constructively in ways perceived by women to be fair, organizations will remain far from realizing the benefits of restructuring and downsizing.\nOriginality/value: - The paper offers insight into job insecurity and motivation among women in the context of Nigerian consolidated banks.\n",
"role": "assistant"
}
] |
Female leadership and school effectiveness in junior high schools in Ghana
|
[
"Women",
"Leadership",
"Schools",
"Ghana"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nFemale leadership and school effectiveness in junior high schools in Ghana: Historical perspective\nMethod: The purpose of this section is to discuss the methods adopted in this study to examine female principal leadership practices in junior high schools. In doing so the research design, sample and sampling procedure, data collection procedure and data analysis were considered.Research design\nEthical consideration: General safeguards to the principal during the interview included the use of an informed consent form, a discussion of the interview agenda and timeframe, and the use of tape recorder to ensure accuracy of information. All agreements made with subjects in the study were kept by the researcher. In addition, the identity of the subjects and other informants were held in confidence. As a male researching into female leadership, the researcher tried to remain as neutral as possible.\nFindings and discussions: The findings clearly demonstrate the unique leadership practices that women exhibit as they manage their schools. One thing which is certain is the emotional intelligence that the five principals exhibited and which is critical in inspiring teachers and building relationships. Indeed, female principals are transformational leaders and are efficacious in achieving the goals and objectives of their schools. Women bring different qualities to leadership and management positions and help organizations maintain a competitive advantage.Vision and mission of schools\nConclusion: Being a principal in the twenty-first century is a challenging task. To qualify for the job, principals need to be visionaries, instructional leaders, curriculum specialists, disciplinarians, facility managers, public relation experts, legal analysts, technology experts, and counselors among others. Principals must posses the requisite skills, capacities, and commitment to lead their schools. Principals must think strategically, handle multiple ambiguous tasks at once, stay clear and steadfast on fundamental vision and values, maintain integrity and interpersonal sensitivity, and handle stress effectively.Based on my interactions with principals, it is argued that principals can and do make a difference in their organization's capacities to provide constructive experiences for children. The principal is the vital actor in the school setting who can bridge context with school, policy and program, means and ends. The five principals demonstrated their abilities to tap the wishes and needs of teachers, students, parents, district education officers, community members, and resources of their communities to achieve the goals of their institutions. The principals were able to create and sustain an image of what quality schooling might be and worked conscientiously toward its achievement. Through routine activities and effective monitoring of the principals, they were able to bring to realization their overarching visions. The principals interacted regularly and appropriately with personnel; and exhibited leadership traits such as self-confidence, dependability, emotional intelligence, flexibility, assertiveness and humor. The principals upheld ethical and moral values and developed a learning community in which satisfied personnel were committed to a student-centered and instructionally focused learning environment. The five principals worked toward the total development of the students - cognitively, physically, emotionally, socially and spiritually. Though the principals cannot succeed overnight and alone, they can be the wellspring that keeps all the participants in school settings pulling together.Contemporary educational reforms place great premium on effective leadership and management of schools. The logic of this position is that an orderly school environment which is efficient and well managed provides the preconditions for enhanced student learning. The principals were transformational leaders who work progressively to transform their schools; a transformation that resulted in high academic performance of the students.The five principals were of the view that there is no secret to the success of the effective school principal, no magic formulas and no hidden models. Rather, the school succeeds to a large degree because of alertness, consistency, resourcefulness, and sustaining energy of the principal. The direct responsibility for improving instruction and learning rests in the hands of the school principals who create a culture that accepts and encourages risk-taking, experimentation, and open dialogue that lead to norms, practices, and power relationships that are uniquely suited to their schools.In the traditional Ghanaian context, women are not perceived as leaders. It is the general belief that the best for women is homemaking. Decision-making is the prerogative of men. Even though the 1992 constitution of Ghana recognizes equality of all persons before the law and prohibits discrimination on the grounds of sex, religion and ethnicity, women continue to be discriminated against in all sectors of the economy. At times, the media perpetuates the stereotypes by representing males as superior and assertive while females are presented as fragile and emotional. This stereotyping of women permeates Ghanaian society and makes it difficult for them to aspire to leadership roles. However, these five principals have been able to excel in their leadership positions by transforming their schools into a place of excellence.Female leaders have come to symbolize new type of leadership that connotes greater effectiveness and synergy than leadership of the past. Appointments of women to leadership positions signal an organization's departure from past practices and help it to capture the symbols of innovation and progressive change. The principals' leadership set the tone of the schools, the climate for learning, the level of professionalism and morale of teachers and the degree of concern for what students may become.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this research is to examine female principal leadership practices that are considered crucial in the effectiveness and improvement of schools and school administration in Ghanaian junior high schools.\nDesign/methodology/approach: - The study was qualitative and interpretive. Five principals of junior high schools were interviewed, their schools observed over a period of three months, and schools' records examined.\nFindings: - The schools had shared visions and missions that were well articulated by the principals and other stakeholders. The principals created a work environment that encouraged creative thinking; designed and implemented new and cutting edge programs; and challenged the status quo.\nResearch limitations/implications: - As a male researching into female leadership, the researcher's gender might influence some of the findings. The sample size is not large enough for any meaningful generalization to be made beyond similar context and geographical contexts.\nPractical implications: - The female principals were transformational leaders, a leadership style demanded in organizations during the twenty-first century.\nOriginality/value: - At the time that this research was conducted, it was the first study on female principal leadership in junior high schools in Ghana.\n",
"role": "assistant"
}
] |
Do bank credit rejection and financial education affect financial self-confidence?
|
[
"Gender",
"SMEs",
"Access to finance",
"Self-confidence",
"Bank credit rejection",
"Financial education"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: It is universally acknowledged that entrepreneurship is a high-risk endeavour that has a low probability of success (Shepherd et al., 2010). Since entrepreneurship is faced with market uncertainties and macroeconomic instabilities, an entrepreneur's self-confidence can act as an important antecedent to start-up a new firm and sustain growth (Busenitz and Barney, 1997; Forbes, 2005; Parker, 2006; Koellinger et al., 2007; Hayward et al., 2010; Ucbasaran et al., 2015). Thus, self-confidence may be viewed as a positive phenomenon and entrepreneurs who exhibit such reap the rewards in terms of enhancing firm performances (Robinson, 2001; Disney and Gathergood, 2013). Self-confidence in the context of entrepreneurship is a multi-dimensional concept inclusive of emotional, cognitive, social and financial resiliencies (Zhao et al., 2005; Wilson et al., 2007; McGee et al., 2009), and it is this trait that separates the entrepreneurs from the non-entrepreneurs (Minniti et al., 2004). For example, Asoni (2011) finds that self-confidence positively influences the entry into entrepreneurship.\n2. Background and hypotheses development: Bank rejection and self-confidence in finance\n3. Data: The data for analysis in this study comes from the surveys of 2,500 UK SMEs during 2004 and 2008 (a full discussion of the data and sampling method are given in Fraser 2004 and 2008). The aims of the surveys were: to provide a benchmark SME finance database made up of the availability and types of finance for SMEs in the UK, to collect information on the relationship between SMEs and their providers of finance, and to provide a general purpose database that will support quantitative research on business finance in the UK and be comparable with other countries. Both data sets are representative of the SMEs who have up to 250 employees in the UK.\n4. Methodology: This study used ordered probit estimation techniques (see Wooldridge, 2002) to model owner manager self-confidence in finance, and examine its association with credit RR, FE and gender (GE) controlling for individual and firm characteristics (X) (see Wooldridge, 2002)[7]. Entrepreneurs answered the question, \"On a scale of 1-10 (where 1 is no confidence and 10 is complete confidence), how confident are you in your own abilities in finance?\" Based on the responses, this study constructed a variable that captured self-confidence in finance by taking values from 1 (no confidence) to 10 (complete confidence), assigning the numeric values (1-10). Since there appeared to be a logical ordering in these responses, the ordered response model was utilised for this study and is presented as[8]:(1)\n5. Findings: This study estimated the self-confidence in finance of owner managers in sole proprietor and partnership businesses only because self-confidence in finance data was not available for limited companies[9]. It first combined the bank finance rejections together and then disaggregated them to examine the differences between outright and partial bank rejections (Table II)[10]. The results in Table II show that whilst outright bank finance rejection reduced the financial self-confidence of owner managers, partial bank finance rejection increased their financial self-confidence in 2004 whereas in 2008 the combined effect was found to be negative and statistically significant. To explain the sign of the partial bank rejection coefficient, it may be viewed as a successful outcome by the applicant. These findings provide partial support for H1 suggesting that (outright) bank finance rejection reduces business owners' self-confidence in finance. Moreover, the results in Table II show that FE increases the likelihood of reporting higher financial self-confidence. Thus, our H2: FE increases business owners' self-confidence in finance is supported. The overall model finds, however, no association between gender and reported self-confidence in finance and therefore H3 is rejected[11].\n6. Discussion: The financial self-confidence of owner managers can support their entrepreneurial capabilities in starting and operating one or more businesses. Additionally, this will facilitate and spur economic development as a consequence of job creation and innovation. To this end, this study exposes the importance of FE on the entrepreneur's financial self-confidence, and consistent with previous research, we find support for the view that human capital is an advantage for entrepreneurs, society and business success (e.g. OECD, 2006; Unger et al., 2011; Block et al., 2013; Disney and Gathergood, 2013). In particular, this paper finds that FE increases entrepreneurs' financial self-confidence. It is argued that financial literacy and knowledge can help entrepreneurs in providing relevant and coherent information to banks and other lenders, which in turn improve the entrepreneurs' lending relationship with their banks and related outcomes. Hence, the government should encourage policy support to enable FE and training for potential entrepreneurs (especially those currently unemployed) towards boosting their financial self-confidence and improving their zeal for entrepreneurship. Additionally, entrepreneurs and potential entrepreneurs are encouraged to explore financial literacy with a view to increasing their financial self-confidence. However, seeking financial advice or financial counselling externally may serve as an alternative to owner manager pursuing FE, but this is subject to a strategic decision by the firm as there may be cost implications.\n7. Conclusion and future directions: This paper examined empirically the link between self-confidence in finance, bank finance RRs, gender and FE using the 2004 and 2008 SME finance surveys. The results showed that outright bank credit rejection is negatively associated with financial self-confidence, whereas partial bank credit rejection may boost self-confidence. Moreover, the results showed that FE has a strong and positive effect on self-confidence in finance. Thus, our results support the need for increasing entrepreneurial and FE in higher education.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine empirically the determinants of owner manager financial self-confidence. In particular, it estimates the effect of bank credit rejection and financial education (FE) on the financial self-confidence of business owners.\nDesign/methodology/approach: This paper uses data from 2004 and 2008 surveys of 2,500 UK small and medium sized enterprises (SMEs). An ordered probit estimation is used to measure and assess the effect of bank credit rejection and FE variables on financial self-confidence for the two periods. The authors also explore potential differences in self-confidence between males and females.\nFindings: The results show that outright bank credit rejection reduces financial self-confidence among owner managers whereas partial bank credit rejection is found to help boost confidence prior to the financial crisis. There is strong evidence that FE increases financial self-confidence. Finally, the authors find no association between gender and reported self-confidence in finance.\nResearch limitations/implications: Entrepreneurs and potential entrepreneurs are encouraged to explore financial literacy and knowledge with a view to increasing their financial self-confidence. This will help SMEs to deal with the banks or other finance providers more efficiently. In addition, better application procedures and information on lending criteria may help SMEs to minimize the probability of bank credit rejection. So the current study has implications for professional bodies as well. The study, however, is restricted to sole proprietor and partnership SMEs and in the UK context only.\nPractical implications: Financial self-confidence has a progressive effect on entrepreneurship and entrepreneurial venture growth. The financial self-confidence of owner managers can support their entrepreneurial capability in starting and operating one or more businesses. As entrepreneurs successfully start and operate their own businesses, they are contributing to economic development through job creation, employment and tax contribution.\nOriginality/value: This paper makes an original contribution in highlighting the usefulness of FE in boosting financial self-confidence among entrepreneurs and potential entrepreneurs. It is also found that the experience of bank credit rejection reduces entrepreneurs' financial self-confidence.\n",
"role": "assistant"
}
] |
Female board participation and annual report readability in firms with boardroom connections
|
[
"Gender diversity",
"Inter-firm networks",
"Readability",
"Annual report",
"Female board participation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Everywhere I have been, the most successful teams have had a breadth of experience and perspectives. Diversity helps fight that feeling of being in an echo chamber (Mary L. Schapiro Former chair of the US Securities and Exchange Commission Financial Times, September 15, 2015).\n1. Introduction: The textual complexity of annual reports has generated considerable concerns among academics, policy makers and practitioners due to the negative effects on the information environment (Dyer et al., 2016; Ernest and Young, 2016; Guay et al., 2016; Lo et al., 2017). The readability of annual reports is currently a central theme in regulatory discussions, as the following quote illustrates:\n2. Theoretical background and hypotheses development: 2.1 Gender diversity\n3. Research design: 3.1 Sample and data\n4. Results: Tables II and III presents the descriptive statistics for all the variables. The mean of the Fog index for the whole sample is 15.1, indicating that the text of the Italian listed firms' annual reports is moderately difficult to read for a large audience of users. The average board size consists of 10.61 directors. The mean of per cent FEMALE_DIR is 0.07, suggesting that our sample is characterized by limited participation of women in the corporate boards.\n5. Discussion and conclusion: Building on different theoretical frameworks, we conduct an empirical examination of whether female board participation has beneficial effects on information transparency. Our hypotheses are tested on hand-collected data comprising 435 annual reports of Italian listed firms and data for over 900 individual board members for the period 2009-2013. We use the Fog index to estimate the readability of annual reports and degree centrality to separate firms with small boardroom connections and firms with large boardroom connections. We focus on Italy, as it is a \"network-oriented market\" with a low level of gender equality and increasing policy initiatives to support progress in female board participation (European Commission, 2014; World Economic Forum, 2015; Bianco et al., 2015).\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to investigate the relationship between the female board participation and the readability of annual report.\nDesign/methodology/approach: Using hand-collected data from a \"network-oriented market\", as exists in Italy, which includes 435 annual reports, this study uses a regression analysis to test whether female board participation affects the annual report readability.\nFindings: Female board participation is found to have a positive impact on disclosure readability in firms with small boardroom connections but the opposite effect in firms with large boardroom connections.\nResearch limitations/implications: This paper responds to recent calls in the corporate governance literature by investigating whether the female board participation affects the transparency of the disclosure practices.\nPractical implications: This study has policy implications, as it helps to improve evaluations of how, and under which circumstances, female board participation may lead to higher disclosure quality and thus benefit investors.\nOriginality/value: This paper shows that female board participation has different effects on the disclosure readability at different levels of board positions in inter-firm networks.\n",
"role": "assistant"
}
] |
Healthcare capacity measurement
|
[
"National Health Service",
"United Kingdom",
"Outpatients",
"Hospitals"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper investigates capacity measurement in the outpatients department of a UK National Health Service (NHS) hospital. The background to this research is the interest within the NHS of the restricted nature of resources and the importance of the service for the \"customers\" wellbeing (Department of Health, 2000). Accurate measurement of theoretical and available capacity is of vital importance for the management of a healthcare organisation.The need for accurate and up-to-date performance information increases for NHS hospitals as Payment by Results and Patient Choice introduce dramatic changes in the way patients access Services and how services are commissioned (Department of Health, 2004). Under Payment by Results, hospitals will be paid according to their activity, receiving funding for every patient that they see based on nationally agreed tariffs (Department of Health, 2002). Therefore, the ability to know at any given time the past, current and predicted activity is directly linked with the financial well being of the hospital and the survival of the services it provides. At the same time, in order to attract patients and funding, hospital trusts need to demonstrate competence in meeting the objectives and targets set by their commissioners and the UK government. Hospital trusts and departments have to prove:* the clinical quality; and* the cost effectiveness of the services they provide (Curtright et al., 2000).Theoretically, departments or even whole hospitals that do not provide cost effective and quality services will have to close (Carvel, 2006). These two aspects, quality and cost, are also emphasised by McDermott and Stock (2007, p. 1025), in their paper on hospital operations they state that \"healthcare costs have become an important issue, so the use of cost improvement or cost containment as a performance measure has been very common in healthcare research ... Quality has also been very widely used as a performance outcome in healthcare research ... measured by many different means, including self-reported quality measures, ... medical errors, and mortality rates. Organizational performance in a hospital, particularly from a strategic perspective, can also be viewed as a construct that combines multiple dimensions, such as clinical outcomes, financial performance, productivity, and operational measures\". Performance measurement takes capacity into account and to research capacity is to study one dimension of performance.This paper examines the measurement of capacity and the systems used to measure and report activity at the outpatient clinics of an NHS hospital in the North West of England. The authors believe that a focus on capacity on a micro (departmental) level is wholly appropriate for healthcare operations and can potentially make a direct contribution to the ability of both the department and the organisation to achieve and maintain its quality and cost objectives. This is the contribution that this paper makes. To this end, from the review of the literature three research questions were developed: Question 1: How is capacity defined and measured at the hospital outpatient department? Question 2: What are the main issues with measuring and reporting operational performance at the outpatient department? Question 3: How can the measurement and reporting of performance at the outpatient department be improved for the benefit of all stakeholders (patients, providers, payers)?Starting with the measurement of capacity at the outpatient medical clinics of the hospital and moving on to assess the gaps in the measurement and reporting of performance, this research identified the need of the outpatient department for a performance measurement system.\nThe literature: Capacity measurement\nResearch methodology: Research focus\nFindings and Discussion: To provide a clear structure for presentation of the findings and discussion this section has been arranged around the research questions.How is capacity defined and measured at the hospital outpatient department? (Question 1)\nConclusions: The methodology applied for collection of the research data was wholly appropriate and consistent with the perceived outcomes required. It generated ample data, which facilitated discussion and the drawing of specific conclusions. This research has also provided a foundation for future work, especially with regard to the capacity and performance measurement and management of healthcare operations. The main contribution of this research is that it has added to the body of knowledge on healthcare capacity measurement, specifically through an exploration of the implementation of the AMF.Starting with the measurement of capacity at the outpatient medical clinics of the hospital and moving on to assess the gaps in the measurement and reporting of performance, this research identified the need of the outpatient department for a performance measurement system. A simple to use, accurate system that could provide information in the requested format with a minimum of manual input.Using ideas from the literature and specifically from the Balanced Scorecard and the Integrated Dynamic Performance Measurement System, a new framework was conceived and implemented. Although providing a rather one-dimensional view of operational activity, the potential for development is significant as the hospital moves towards the development of a (strategic) Balanced Scorecard. The AMF provides the basis for connecting strategic performance with operational baselines. The next steps, which the hospital has already began working on are:* further development of metrics, especially financial performance information;* direct integration with the organisational performance measurement systems; and* assessment of the AMF effect on capacity and performance measurement and management.It will be interesting to see how these actions influence the development and refinement of the AMF.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to present findings from an action research intervention in the outpatient department of a National Health Service (NHS) Hospital. It investigates the perceived and actual problems of measuring capacity at a UK hospital.\nDesign/methodology/approach: - An action research methodology was adopted: relevant literature on capacity measurement and performance was reviewed; the motivation behind its usage explored. Systems requiring improvement were identified and the adoption and implementation of new working methods initiated and explored. In addition to considering capacity measurement and performance issues, the authors examined the effects this may have on the long-term potential of the organisation.\nFindings: - The research identified gaps in the capacity and activity measures used and in the dissemination of performance information. To address these problems and meet the changing needs of the department a new performance measurement and reporting tool was implemented.\nResearch limitations/implications: - The methodology applied was appropriate, generating data to facilitate discussion and draw specific conclusions from. A perceived limitation is the single case approach, however, Remenyi argues this can be enough to add to the body of knowledge.\nPractical implications: - The paper gives details about the introduction of new performance measures and the development of a performance measurement tool for the Outpatient Department of the Hospital. The conclusions drawn add substantially to previous commentaries, and develop interesting questions for future research.\nOriginality/value: - Specific recommendations for the implementation of a new performance tool for healthcare organisations are made.\n",
"role": "assistant"
}
] |
Exploring the interaction between R&D and production in their globalisation
|
[
"Research and development",
"Interaction",
"Case studies",
"Globalization",
"Global production",
"Global R&D"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Today, globalisation is not a new phenomenon anymore. Over the last decades, both global trade and foreign direct investment (FDI) have increased explosively, which have further changed the world permanently, for better or for worse. The pace of globalisation is accordingly quickening and will continue to have a growing impact on business organisation and practice (Friedman, 2006).\n2. Literature review: 2.1. Globalisation/internationalisation and related theories\n3. Methodology: The research methodology is precipitated by the research question. In order to answer the \"how\" questions formulated above, the case study was selected as the primary research method and mainly used for an explorative purpose (Handfield and Melnyk, 1998; Yin, 2003). Unconstrained by the limitations of other methods, such as questionnaires and models, the case study can lead to new and creative insights and have higher validity for practitioners (Voss et al., 2002).\n4. Empirical cases: In this section, we briefly present the case report of each selected company. At the end of each case description, the key information of that case about the evolution of its production/R&D facilities and networks is mapped and summarised in terms of the aspects listed in Table III. These aspects are merely useful to describe the \"snapshots\" at a particular point in time. However, the current value of each variable is normally the outcome of managerial decisions made in a particular area of the facilities and networks over a long period of time (Miltenburg, 2009). In order to trace their evolution and understand the interaction over time amongst individual facilities and between individual facilities and the networks to which they belong, comparisons amongst snapshots captured at different points in time are needed.\n5. Empirical findings: As shown in Tables IV-VII, the evolution of production/R&D facilities and networks in case companies is traced by capturing their snapshots at different points in time. The comparison and analysis of these snapshots allow us to examine how the evolution of one facility affected other production or R&D facilities and led to changes in both the production and R&D network as a whole and furthermore, to identify the approaches that case companies followed to globalise their production and R&D.\n6. Discussions and tentative propositions: In this section, we further discuss the case studies and the empirical findings, based on which we offer some tentative propositions.\n7. Conclusions and implications: 7.1. Theoretical contributions\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to extend the discussions on globalisation from production to R&D. It investigates how R&D and production interact with each other in their globalisation processes.\nDesign/methodology/approach: The research aim is developed by identifying the gaps in the reviewed literature. This paper is based on four case studies undertaken in one Chinese manufacturing, one Danish pharmaceutical and two Danish manufacturing companies. The cases provide a sound basis for developing an understanding of the interaction between the globalisation of R&D and production.\nFindings: This paper identifies three approaches the case companies followed to globalise their production and R&D: interactive globalisation, separated globalisation and a possible combination. The paper indicates that research and development might have to be treated separately with regard to their globalisation, and proposes industry and country characteristics as the key factors for globalisation approach selection, and site capability and strategic decision as the impacting factors for globalisation evolution.\nOriginality/value: This paper emphasises the dispersion of R&D activities, which is seldom addressed by existing internationalisation theories. Its investigation provides a foundation for the further extension of current internationalisation theories to consider global R&D. Moreover, the theoretical gap in the existing literature between global R&D and production is noted. This paper bridges this gap by clarifying the interaction between R&D and production in their globalisation, conceptualising three globalisation approaches, and proposing tentative factors that have impacts on approach selection and management.\n",
"role": "assistant"
}
] |
Manufacturer's characteristics that determine the choice of producing store brands
|
[
"Generics",
"Brands",
"Retailers",
"Manufacturers",
"Spain"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Academic and managerial interest in store brands has been increasing in recent years (Baltas, 2003). Despite proven significance and wide acceptance of store brands, most prior studies have been performed in the USA (Jin and Suh, 2005) and the bulk of existing studies are based on North American data, which may reflect economic, social and marketing conditions different from Europe (Baltas, 2003). However, the number of studies based on European data is growing, especially in those countries where these brands are in a maturity stage (for instance, see Le Borgne, 1996; Gallizzi et al., 1997; Burt, 2000; Baltas, 1997, 2003; Bontems et al., 1999; Gomez, 2005; Gonzalez et al., 2006).Retailers consider store brands a powerful competitive tool which allows them to improve store image and obtain greater margins, leading to greater profits (Simmons and Meredith, 1983; Nandan and Dickinson, 1994, Narasimhan and Wilcox, 1998; Recio and Roman, 1999). Store brands also enable them to establish prices and promotions in a more flexible manner, as well as offering the consumer information and a guarantee on their products at a lower cost than manufacturer brands (Nandan and Dickinson, 1994).Nowadays, the strong growth of large distribution chains and their high level of concentration in most countries in the European Union give retailers an advantageous position in negotiations with manufacturers (Hoch, 1996; Dhar and Hoch, 1997; Cruz, 1999; Burt, 2000; Casares et al., 2000). For this reason, when negotiating with retailers manufacturers emphasize the conditions regarding distribution services as evidenced on the shelf or in the outlet, such as the payment for shelf end units, point of sale promotions, product replacement on the shelf, or product range (Fernandez and Gomez, 2005).Some authors have attempted to classify the different research areas on store brands (Baltas, 1997). In our view, the typology offered by Mendez et al. (2000) is the most appropriate one for identifying research opportunities. It places the emphasis on the main actors of the demand-supply chain (retailers, consumers and manufacturers). According to this classification, we have reviewed the literature on store brands published in academic journals since 1990. There are 77 references. Of those, 38 per cent (30 articles) focus on consumer perspective, 34 per cent (26 articles) carry out their analysis from the perspective of the retailer, and 28 per cent (21 articles) focus on manufacturer strategies regarding store brands.From a manufacturer's perspective, to our knowledge, there is no previous study in the public domain that empirically assesses the rationale for the production of store brands; and identifies and analyses groups of manufacturers most willing to produce these brands based on different descriptive variables. Therefore, the intended contribution of this article is two-fold: first, from a theoretical perspective, we summarise previous work that has examined the production of store brands; second, we aim to investigate empirically that there are differences between manufacturers and non-manufacturers of store brands according to their characteristics, such as manufacturers' leadership, size, advertising investment, sector concentration and type of manufactured product.A study on rationality in store brand production and characterization of manufacturing and non-manufacturing groups of store brands has important practical implications. One of the implications is regarding the coexistence of store brands and distributor brands in large consumer markets. Another implication involves recommended practices in portfolio management of manufacturing groups for which the distributor brand production is found to be more recommendable.Any examination of the manufacturer-level factors that explains the production of store brands should start with a framework to analyse this behaviour. Hence, the next section reviews the existing literature on the production of store brands. Then we go on to describe the research methodology and the main results. The final section of the paper presents some conclusions, managerial implications and the limitations of this study.\nTheoretical background on the production of store brands: The existing literature on store brands based on manufacturers' decisions has mainly studied two areas:1. The differences and competition between manufacturer and store brands (De Chernatory, 1989; Makoto, 1995; Aggarwal and Cha, 1998; Narasihman and Wilcox, 1998; Putsis and Cotterill, 1999; Bontems et al., 1999; Cotterill et al., 2000; Burt, 2000 and Horowitz, 2000).2. The competitive manufacturer strategies against store brands (Parker and Kim, 1997; Ashley, 1998; Apelbaum et al., 2003; Ailawadi et al., 2003, Mills, 1995, Verhoef et al. 2002, Hoch, 1996; Quelch and Harding, 1996).Nevertheless, literature on the production of store brands is rather scarce (Glemet and Mira, 1993a, b; Quelch and Harding, 1996; Dunne and Narasimhan, 1999; Verhoef et al., 2002). This issue is important, however, as store brands are strategic negotiating tools for retailers (Simmons and Meredith, 1983; Narasimhan and Wilcox, 1998; Corstjens and Lal, 2000) and the manufacturers are conscious that their production can affect the nature of manufacturer-retailer relationship (Narasimhan and Wilcox, 1998; Le Borgne, 1996; Bell et al., 1997; Cotterill et al., 2000).Store brands have been developed from competition based on imitating a manufacturer brand and selling it at a lower price (copycat brand - Kapferer, 1997) to try to create a distinct identity in the market based on value for money. Nowadays, many European retailers have improved their store brand positioning, relegating the role of low quality product to white-label generic or to a distinct second-tier store brand. These brands are called \"price leader brands\". Their name comes from their price, since these brands are the cheapest ones on the shelves. Sometimes retailers do not want to be identified with the label of their price leader brands and the chain logo does not appear on the packaging. On the other hand, some retailers have developed products with superior quality for some of their store brands with packages not based on replication but on actual innovation (Fernandez and Gomez, 2005).Manufacturers have begun to consider the store brand as another brand; that is, as important competition that may threaten consumer loyalty towards their own brands (Rubio, 2004). Retailers are increasing their marketing efforts for their own brands with the aim of increasing consumer loyalty towards their store (Corstjens and Lal, 2000).The analysis of the rationale for the manufacture of store brands includes a study of the benefits and costs associated with the decision for producing these brands that are not the same for all manufacturers. There are certain characteristics associated with the producers that intensify the benefits and reduce the costs of this decision and, therefore, promote the existence of potential manufacturer segments in the production of these brands. Next we analyse the rationalization for producing store brands, in both a general and a specific way, according to the characteristics of the manufacturer.Rationale for the manufacture of store brands\nMethodology: From the previous literature review we have identified some main aspects that constitute our research questions:RQ1. What are the main manufacturers' reasons which favour the production of store brands?RQ2. What is the rationale for the manufacture of store brands according to the manufacturers' competitive position?RQ3. Is it possible to identify different store brand manufacturer segments in relation to their descriptive variables?In order to answer these questions, we undertook a survey directed at business units of consumer packaged goods companies, in particular food and drink, personal hygiene and home cleaning sectors. The questionnaires were sent by post to 801 companies, with a turnover higher than 4.8 million euros. This threshold allowed us to include all of the Spanish provinces in the study. The companies with various business units were requested to complete a separate questionnaire for each one. The key informants of the companies were the business unit general or marketing managers.The questionnaire was developed in close collaboration with various representatives of companies in the sector, who co-operated in the design and pre-test phases. A business database was used to find the necessary information - the Spanish National Industry Classification (CNAE) and sales volume - in order to select the sample of companies. Moreover, the Spanish Association of Brand Manufacturers (PROMARCA) cooperated in encouraging the response of member manufacturers of their association (manufacturers with a high turnover in the consumer packaged goods markets). Questionnaires were sent from December, 2000 to April, 2001. The response rate was 22 per cent. The total number of valid questionnaires was 161. The data-gathering stage was divided into two phases. First, 72 questionnaires were received, 69 of which were valid; following-up on the non-responses a further 102 questionnaires were received, 92 of which were valid. Only five companies answered more than one questionnaire, one for each of their business units. One of them answered for five different business units, two companies answered three questionnaires each and two questionnaires were received from each of the remaining two companies.In their answers, 117 companies identified themselves. For this group of companies and the other 684 which were included in our original sample, we checked to see if there were significant differences in the selection variables used (Spanish National Industry Classification and sales volume). We found only significant differences in the latter variable on behalf of bigger companies (X2: 32.774, significant at p<=0.01). Of the large companies with a turnover above 120.20 million EUR, 47.2 per cent (17) identified themselves, compared to 12.1 per cent (50) and 14.4 per cent (50) of the small and medium-size companies respectively. However, no significant differences were found between manufacturing companies in the general food and drink sector (14.9 per cent (103)) compared to the personal hygiene and home cleaning sector (13 per cent (14)).We use descriptive analysis to determine the principal reasons for the manufacture or non-manufacture of store brands and bivariate techniques (cross tabulation) to identify the most significant descriptive variables leading to the production of store brands. Then, we use a multivariate technique of hierarchical segmentation in order to identify different segments of manufacturers and non-manufacturers of store brands in relation to their descriptive characteristics. In this case, due to the scale type of the criterion variable, we used the C&RT algorithm, a technique that uses trees of classification and regression developed by Breiman et al. (1984). Finally, we apply correspondence analysis to assess the rationale for the manufacture of store brands according to the manufacturers' competitive position.The statistical analysis of the information was carried out with the statistics package SPSS v.13.0. In the application of C&RT, we were aware that the small size of the sample (161 companies) may influence the results. We will refer to this limitation in the conclusions.\nResults: Table III shows the results of all the valid questionnaires received regarding the distribution of the strategic business units (SBU) by sector, the descriptive characteristics - type of manufactured product, market share of the three leading competitors, geographical scope of reference, turnover of the SBU, average market share of the SBU, innovation and advertising expenditure - and the percentage of manufacturers that produce store brands and price leader brands.Main reasons for the manufacture of store brands\nConclusions, limitations and further research: Store brands have been gaining an increasing share of the market all over Europe, which represents the region with the most development in private labels with an aggregated share in value sales of 23 per cent. The share of private label in North America is currently at 16 per cent. Private label sales grew in Europe by 4 per cent in 2005, while aggregated sales of manufacturer brands remained flat. This growth is illustrative of the fact that sales of private labels have not peaked, even within the largest private label market (ACNielsen, 2005).In this research, we have shown the main reasons for not producing store brands, which are non-satisfactory profitability, cannibalization and low guarantee of distributor loyalty, and the reasons for doing so, such as lower unit production costs, non-production of store brands by competitors and cooperation with the retailer. For store brand manufacturers, several retailers' demands have been found, the most important being periodic revisions of the production contract, the internationalization and the differentiation of store brands. In addition, we provide empirical evidence in favour of the rationale for the production of store brands by non-leading manufacturers and against this practice by leading manufacturers.Until now, from a manufacturers' perspective, no research has focused on the manufacturers' group identification regarding their decision to produce store brands according to their descriptive characteristics. With this study we have sought to fill this research void, focusing on the production of store brands and determining the most significant variables in the identification of manufacturer segments according to their descriptive characteristics. We have shown that there are statistically significant differences between manufacturers and non-manufacturers of store brands in the descriptive variables of product type, competitive position and production of price leader brands. Store brand manufacturers produce these brands mostly for the food sector rather than for the personal hygiene and house cleaning sector; they are mainly non-leading manufacturers, and they are more prone to produce price leader brands.Moreover, through hierarchical segmentation, several segments with different percentage of store brand manufacturers are identified according to their descriptive characteristics. On the one hand, we find differences in relation to the size of manufacturers that decide to produce store brands. Our findings show that SBUs with the lowest turnover present a high percentage of store brand manufacturers, compared to SBUs with the highest turnover. On the other hand, regarding mid-sized manufacturers, it must be noted that there is a high percentage of store brand manufacturers in the food and drink sector, compared to the low percentage in the personal hygiene and home cleaning sector. Moreover, we must highlight the influence of the advertising expenditure in the food and drink sector in the production of store brands. Generally, mid-sized manufacturers with a low advertising expenditure are the main group that chooses to produce these brands.Managerial implications\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to study the rationale for the production of store brands and the possible existence of different store brand manufacturer segments according to their competitive position, size, advertising investment, sector concentration and type of product.\nDesign/methodology/approach: - The method to obtain information is a survey developed in Spain and answered by manufacturers of consumer packaged goods. The methods applied are descriptive (Crosstabs and Anova) and multivariate techniques (correspondence analysis, chi-square automatic interaction detector (CHAID), and cluster).\nFindings: - The article provides empirical evidence in favour of the rationale for the production of store brands by non-leading medium to small-sized firms and against this practice by leading manufacturers. It also shows the existence of different manufacturer segments.\nResearch limitations/implications: - The main disadvantages of this research are based on their exploratory nature and on the method to obtain information (postal survey): self-selection, small sample size and measuring of the variables (perceptual bias). Additionally, the goodness of fit of the CHAID applications is not very high.\nPractical implications: - The insights presented here can be used by manufacturers from other countries with a less developed store brands market, such as Eastern European or Latin American. Leading manufacturers or mid-size manufacturers with high advertising investment can still compete head to head with store brands. The other manufacturers must adapt their strategies to collaborate with retailers in order to produce store brands.\nOriginality/value: - There are no previous studies that empirically assess the rationale for the production of store brands. Moreover, no empirical research has focused on the manufacturers' group identification regarding their decision to produce store brands. This study determines the most significant variables in the identification of manufacturer groups.\n",
"role": "assistant"
}
] |
The corporation and private politics
|
[
"Reputation",
"Activism",
"Citizens united",
"Private politics",
"Religious freedom",
"Same-sex marriage"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: \"The public be damned [...] I answer to my stockholders\" (Dow, 2006). William Henry Vanderbilt's response to a reporter's question has a somewhat murky attribution but it still serves as a trenchant example of the traditional corporate response to questions about a corporation's duty to stakeholders other than investors. Whether Vanderbilt was misquoted or not, he did reiterate the substance of this view in a later interview with the Chicago Tribune: \"Railroads are not run for the public benefit but to pay. Incidentally we may benefit humanity but the aim is to earn dividends\" (Bernays, 1928).\nDynamic sentiment mapping: As search analytics and conventional opinion polling demonstrate, consumer sentiment on certain issues is changing as rapidly as ever before. This is especially true of an issue such as same-sex marriage in the USA where there has been a profound shift in as little as the past ten years. Rising support and dropping opposition began to accelerate in 2004 and support bested opposition in 2010. These trends have continued to today with the gap, now positive beginning to increase (Loeb, 2014; Silver, 2013). Clearly, in the context of rapidly changing opinion the best defense to being caught in the shifting sands of private politics is a good offense, a continuous and dynamic mapping process to determine the speed of change to help organizations determine when their policy or perspective falls out of favor. In terms of reputation, there appears to be an optimal window for shifting one's position on an issue as a corporation - not so early as to be regarded as a zealot, with the risk of offending those other customers opposed, not so late that the company can be painted as a Neanderthal hold out. The optimal window in this scenario is to be just slightly ahead of the curve, just enough to be given credit for leadership, but not so much that opponents can still make life difficult. American drugstores and retailers are currently faced with this calculus, trying to decide whether to emulate pharmacy chain CVS Caremark in banning the sale of all tobacco products (Strom, 2014).\nDeveloping points of view: Most companies have well-established points of view on the subjects of most direct interest to them or on issues on which activists have challenged them in the past. Companies like John Deere and Caterpillar that make heavy machinery, for example, are veterans of dealing with public sentiment on politically and socially sensitive issues such as deforestation in the Amazon or the building of Jewish settlements on the West Bank, in which the use of their machinery has sometimes been attacked (Nichols, 2012; Hale, 2012). However, every company needs to explore their own views and positions on subjects that have very little or no connection to their activities. These can be wide ranging indeed, from citizenship for the children of undocumented aliens to the tension between food and fuel uses of soybeans. In most instances, this preparation will be purely prophylactic, but having a point of view when called upon creates a strong impression of corporate leadership and corporate citizenship. It is also consistent with the increasing perception that private commercial entities such as multinational organization and public institutions such as governments and government departments should be acting together to solve the world's most challenging problems.\nContinuous consumer conversations: Engagement with customers and consumers is never ideal if it only takes place during the heat of legislative battles or other activist campaigns. The best time to have a dialog with consumers on this issue is early enough in the process so that the corporation can position itself as interested in sentiment on the issue rather than defending itself against activist attacks. An example of this might be the issue of the use of animals in advertising. Animal rights groups are vehemently opposed to the use of animals in commercials because they believe that the training required to enable animals to \"perform\" on camera is abusive by its very nature. However, it is not clear that this is a subject of enormous concern to the general public. Using social media polls and asking questions of consumers about a range of issues before they become enflamed accomplishes two aims: first, it helps inform a potential decision about changing company policies to reflect consumer concerns, and second, it also signals a willingness to engage on social topics and correct such policies if necessary.\nDecision clarity: Armed with insights from both of these processes help create the most important element when companies become involved in private politics: decision clarity. Decision clarity in this context means being clear on what the company's policy is and whether and how/if it might change. Knowing when and how the corporation might shift prevents the slide into well-meaning vagueness. The reason that vagueness is debilitating and harms corporate reputations is that it encourages activist groups to continue to apply public pressure in the hope that the company's position might change. If you do pharmaceutical research, for example, and have no intention of ending testing on primates, then it is best to be clear on this issue. It may produce a more aggressive response in the short term, but will prevent a long-running and damaging game of cat and mouse in which opponents try to shift corporate policy inch by inch.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this article is to examine the increased public pressure on corporations to express a point of view on social politics such as same-sex marriage that have no intrinsic connection to their line of business.\nDesign/methodology/approach: By analyzing recent cases in which corporations have suffered from negative public sentiment regarding their policies on social issues, the authors adduce some guidelines for handling the intersection of corporate reputation and private politics.\nFindings: This analysis led the authors to propose four strategies for coping with reputation assaults relating to private politics: dynamic sentiment mapping, developing points of view, continuous consumer conversations and decision clarity.\nOriginality/value: There is existing literature on this subject, but few recommendations for how to handle the issue.\n",
"role": "assistant"
}
] |
Has the link between inflation uncertainty and interest rates changed after inflation targeting?
|
[
"Kalman filter",
"Inflation uncertainty",
"Interest rates",
"Inflation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: During the early 1990s several countries adopted explicit inflation targeting (IT) as a tool for monetary policy under the operational independence of the central bank. They recognized the benefits of price stability and consequently adopted it as the principle goal of monetary policy. The effect of inflation on economic performance is an important but complex topic, because it may influence growth negatively[1]. Other than actual inflation, the inflation uncertainty may also effect economic growth either way. Therefore it is also important to study the effect of inflation uncertainty on inflation[2]. Friedman (1977) in his Nobel address claims that inflation and nominal uncertainty are positively correlated and causation runs from inflation to uncertainty about future inflation. An increase in inflation may induce an erratic policy response by the monetary authority that creates more uncertainty about future rates of inflation. Demetriadas (1988) shows the presence of asymmetric information but did not mention the direction of causality. Increased inflation uncertainty may increase unanticipated inflation. In turn, costs will be associated with the anticipated and nominal uncertainty effects on interest rates (see Huizinga, 1993). Wilson (2006) studied the link between inflation, inflation uncertainty and output growth using a EGARCH-M model and found that increased inflation uncertainty increased inflation and lowered economic growth for Japan. Wilson and Culver (1999), Grier and Perry (2000), Hayford (2000), Fountas et al. (2002) and Grier et al. (2004) found a similar negative relationship between inflation uncertainty and output growth, while, Levin and Renelt (1992), Levin and Zervos (1993) and Clark (1997) failed to find any such relationship.This study explores the relationship between inflation uncertainty and interest rates. Interest rates and price are important variables in the macroeconomy and are often monitored by the policymakers, especially for inflation targeting (IT) purposes. The relationship between these variables has been subject to substantial research. Wilcox (1983), Uribe (2002), Berument and Jelassi (2002), Fahmy and Kandil (2002) and Kandil (2005) focused on the relationship between prices and interest rates. In general, inflation uncertainty affects the economy by increasing long-term interest rates. Ball and Tchaidze (2002, p. 108) note that:A large literature argues that monetary policy under Alan Greenspan is well explained by a simple reaction function. Interest rates rise when inflation rises and fall when there is a greater economic risk.The relationship between interest rates and inflation risk is first mentioned by Irving Fisher. Fisher's hypothesis suggests that interest rates and inflation (or expected inflation) are positively related to each other. There are different opinions regarding this type of relationship. Boudoukh and Richardson (1993) argue that this positive relationship exists generally. Mishkin (1992) found a long run relationship for the US in the long run but failed to establish the Fisher relationship in the short run. Yuhn (1996) found stronger Fisher effect over long horizons for Germany. Berument et al. (2005) studied the relationship between three different types of inflation uncertainty with the interest rates of the UK before and after the inflation targeting period, and supports the notion of inflation targeting regimes. In this paper we extended our research from the UK to other inflation targeting countries to examine the affects of inflation uncertainty on interest rates before and after the inflation targeting regime.1.1 A note on inflation targeting countries\n2. Data and methodology: Monthly data from February 1973 to January 2007 has been used for this study. All interest rates and inflation data (change in natural log of consumer price index) were collected from the International Financial Statistics (International Monetary Fund) and the industrial production data were collected from Data Stream.2.1 Monthly interest rates calculation\n3. Empirical evidence: Table I shows the mean and standard deviation of economic growth, interest rates, inflation and expected inflation for five countries under study for the full period, before and after the inflation targeting period. It is clear from the table that after the inflation target period, the growth rate for all the countries increased, inflation has reduced to within the targeted range expect for Spain. The average inflation for Spain is 3.17 per cent after the inflation targeting - a shortfall of 0.17 per cent less than the actual target. Interest rates have also decreased for all the countries after the inflation targeting period. Moreover for all variables the volatility is decreased after the inflation targeting period. Therefore, from the visual description it looks like that the inflation targeting is successful.The impulse uncertainty for the inflation dynamic is captured by the parameters a0, a1 and b1 (Table II). For all the series most of these parameters are statistically significant. This component of inflation uncertainty represents the shocks that hit the economy. In a GARCH specification (a1+b1) denotes the persistence of the shocks. In this sense, the persistence of such shocks is highest in Spain and lowest for Canada. For the other three countries this is very similar. In general, the persistence of impulse uncertainty is high for all the series.The structural parameters of the inflation dynamic represented by the parameters ph1, and ph2 capture the changes in association with past inflation to the present realization, and ph0 is indicative of the present level of the inflation. Since these are all time varying the uncertainty introduced by this time variation is the structural component of inflation uncertainty. In other words the time varying parameters show how the shocks hitting the economy propagate through the system. The structural time series model of inflation implemented in this study allows us to separate these two components easily and examine any differing behaviour subsequently. In the case of Finland, the variance of the noise term of one of the autoregressive parameters is significant. However, the level component captured by ph0 has the variance of the noise term significant for all the series. The overall indication from this result is that the structural component of inflation uncertainty is important for all the samples examined in this study, but the behaviour differs from country to country. It is also interesting to visualize the patterns of structural inflation uncertainty from Figure 1. This plots the sum of the two auto-regressive parameters of the inflation equation (1), (ph1+ph2). When the estimated sum is negative it would indicate some error correction mechanism at work. In this sense, Canada, Finland, Spain and Sweden exhibit this feature over much of the sample period examined here. In other words, as inflation tends to increase the monetary authorities take corrective action. When this sum is positive it tends to indicate high persistence as in the case of the UK. This may indeed lower the structural uncertainty of inflation.Tables III-V demonstrate the estimates of the coefficients of equation (6) using ordinary least square method for full period, pre and post inflation targeting period. Table III shows the estimated coefficients and the t-statistics for equation 6 for the full period. We have considered different types of interest rates as dependent variables. Almost for all the countries the coefficient of the output gap is positive except for the money market rates (MMR) for Spain and long-term interest rates (LTR) for Sweden but none of the coefficients are significant, which is expected and parallel to the findings of Berument et al. (2005). Expected inflation shows a positive and significant effect on interest rates except for UK as expected. Therefore we can say that central banks increased interest rates when the expected inflation increased. The positive risk premium is also an important result and concordant with other research results. The estimated coefficients for Impulse and Structural uncertainty are insignificant for most of the cases except for Spain and for LTR for the UK. Some of the results are different from Berument et al. (2005), which may be due to the fact that we have considered industrial production in lieu of real GDP for the calculation of the GAP.The expected inflation and the output gap coefficients for the pre-inflation targeting period are insignificant for most of the countries expect for Canada. Impulse and structural inflation uncertainty has no effect on interest rates for this period.For the post-inflation targeting period the coefficients are positive and significant for Finland. Other results are similar. If we compare the effect of expected inflation, structural uncertainty, impulse uncertainty and output gap we cannot see any significant difference for the three said periods, however it is clear that for Finland and Sweden the central bank acted on expected inflation much quicker after the inflation targeting period. The first lag of the interest rate coefficient on the interest rates is significant and for most of the countries it is more than one.\n4. Conclusion: The results are rather conflicting regarding the effect of inflation uncertainty on interest rates. But one can reach some interesting conclusion from this study if the observed long run effect of inflation on interest rates is considered. We note that the estimated value of (1-d1-d2-d3)-1g1 is less than one for all the countries under study for the post-inflation targeting period and more than one for Finland (lending rates), Sweden (Bank rates) and for the UK (Interbank rates) for the pre inflation targeting period. The estimated long run coefficient of 1.47 per cent for Sweden suggests that for a 1 per cent increase of the expected inflation, the central bank increased the interest rates by more than one percent, or the real interest rates increased by 0.47 per cent. Therefore, after the inflation targeting period, the real interest rates have decreased for all the countries under study and the respective central banks are less aggressive with respect to interest rate increase. It is also clear from Table I that the volatility of expected inflation was much lower during the post-inflation targeting period and therefore, there is no need to react as promptly as before. Moreover the inflation is well under control after the inflation targeting period. These results are also important from a policy perspective. In general the purpose of the monetary authority is to eliminate the uncertainty arising from higher inflation. It is clear that the respective monetary authority has successfully eliminated these uncertainties and consequently controlled the level of inflation as well as the uncertainties through a credible inflation targeting strategy.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to establish a link between inflation uncertainty and interest rates for five inflation-targeting countries.\nDesign/methodology/approach: - The approach takes the form of a time-varying parameter model with a Generalized Autoregressive Conditional Heteroskedasticity (GARCH) specification, used to derive impulse uncertainty and structural uncertainty.\nFindings: - This study attempts to establish a link between inflation uncertainty and interest rates for five inflation-targeting countries, i.e. Canada, Finland, Spain, Sweden, and the UK. Decomposing inflation uncertainty into two components - impulse and structural, a positive association was found between the expected inflation and interest rates. Structural uncertainty has a positive and significant effect on interest rates for some countries. It has also been found that the long-run effects of inflation on interest rates are less than unity for the post-inflation targeting period, which implies that in some respect the Central Bank has been successful in targeting inflation. This has allowed the Central Bank to employ a less restrictive monetary policy in an environment of a credible inflation-targeting strategy.\nResearch limitations/implications: - Exponential Generalized Autoregressive Conditional Heteroskedasticity (EGARCH) can be used instead of GARCH modelling.\nOriginality/value: - This is the first study that has tried to establish the link between different types of inflation uncertainty and interest rates for the inflation-targeting countries to see the effect of inflation targeting.\n",
"role": "assistant"
}
] |
Getting the team together: the role of emotional intelligence and self-efficacy in team performance
|
[
"Team performance",
"Self-efficacy",
"Emotional intelligence",
"Teams",
"Team cohesion",
"Collective emotional intelligence"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nWhat makes a good team?: As organizations respond to an increasingly complex business environment, the role teams play in organizational performance is more important than ever. Workplace environments have become progressively collaborative as organizations shift from an emphasis on individual achievement to team performance. Team performance is integral to achieving organizational success.\nTeam cohesion: Previous research has indicated that team cohesion is the most important factor in achieving peak team performance. Team cohesion refers to the strength of interpersonal relationships between team members, driving the team members motivation to want to work with each other and to stay part of the team. Teams with strong cohesion have smooth and efficient internal team processes and tend to reach a faster, stronger agreement in collaborative decision making. They provide psychological safety for members enabling potential risks or challenges to be raised early, and conflict resolved quickly. Team members are motivated, engaged and feel increased satisfaction from being part of the team. Cohesive teams are more cooperative and effective in achieving the goals they set for themselves. Although in most cases team cohesiveness contributes to the achievement of organizational goals, strong team cohesiveness can operate to counter organizational performance when the team goals and organizational goals are not aligned.\nThe role of emotional intelligence: Emotional intelligence is key to embodying the social skills necessary for teamwork. It enables team members to manage relationships, build networks, communicate effectively and recognize how others are feeling. Emotional intelligence allows people to navigate and leverage conflict to achieve greater team results.\nThe link between self-efficacy and emotional intelligence: Previous research has indicated that self-efficacy positively correlates with individual performance. Self-efficacy is the belief a person has in their ability to complete a certain task and overcome challenges to achieve their goals and create results. Self-efficacy could be a precursory factor that supports favorable conditions for team cohesion. Self-confident team members who have belief in their competence and ability to overcome challenges and resolve problems are more likely to be positive and proactive in team interactions. Teams containing members with high self-efficacy have more effective conflict management, proactive members, share knowledge and improve performance all of which support the development of strong team cohesion.\nThe study: Study participants were business college seniors at a US university. The participants all trained and competed online in a business simulation involving running a start-up company in a technology-oriented industry. A total of 146 students in 35 teams participated in the study, with the simulation accounting for 20% of their final mark. Teams were structured to be as diverse as possible consisting of four to five students of diversity by ethnic background, age, gender and academic major.\nHypotheses and results: All hypotheses were supported.\nKey findings: The study found the teams with higher average emotional intelligence had greater levels of team cohesion and achieved a better financial performance. Team members with high emotional intelligence also had high self-efficacy, which was found to mediate the relationship between emotional intelligence and team cohesion.\nImplications for practice: Teams are a critical resource for organizations.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to review the latest management developments across the globe and pinpoint practical implications from cutting-edge research and case studies.\nDesign/methodology/approach: This briefing is prepared by an independent writer who adds their own impartial comments and places the articles in context.\nFindings: Teams with higher collective emotional intelligence have stronger levels of team cohesion. In a business simulation, these teams achieved a better financial performance. High emotional intelligence was linked to high self-efficacy, which was found to mediate the relationship between emotional intelligence and team cohesion.\nOriginality/value: The briefing saves busy executives and researchers hours of reading time by selecting only the very best, most pertinent information and presenting it in a condensed and easy-to-digest format.\n",
"role": "assistant"
}
] |
Categorizing corporate social responsibility (CSR) initiatives in B2B markets: the why, when and how
|
[
"Business ethics",
"Green marketing",
"Sustainability initiatives",
"Corporate social responsibility (CSR)",
"Sustainability"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Today the impetus for firms to practice sustainability initiatives comes from many directions, such as consumers increasingly demanding products and services produced in an environmentally friendly manner (Cone Communications, 2013), regulatory bodies aware of the long-term environmental effects of climate change and the ensuing government regulations, and investors' increasing awareness of the relationship between a firm's sustainability initiatives with its long-term profitability and viability. While corporate social responsibility (CSR) is considered an embryonic and contestable concept (Windsor, 2006), Porter and Reinhardt (2007) suggest that going forward, demonstrably practicing serious sustainability initiatives will be a strategic imperative rather than solely a CSR issue.\nCorporate social responsibility - a broad view: CSR has been debated in the business literature for more than five decades. Some of the earliest conceptualizations of CSR saw it as a continuum from low-to-high responsibility (Eels, 1960; Walton, 1967). For Davies (1973), CSR begins where the law ends: \"It's a firm's acceptance of a social obligation beyond the requirement of the law ...\" For us, corporate socially responsible behavior begins with the law. Firms may, at their peril, choose to behave outside regulations that establish expectations mandated by government. Further, government regulations are generally mandated to address social expectations that businesses might otherwise ignore should they simply choose to pursue economic goals. Some businesses do: While many recyclers of electronic waste behave in a socially responsible manner, an increasing number export such waste to developing countries where regulations seldom exist, or, if they do, are seldom enforced. The risk of getting caught and the small penalties if caught outweigh the economic advantages that result from avoiding costly recycling standards (McClearn, 2013). Edmunds (1977) ascribes the emergence of interest in CSR to the debates concerning business versus government responsibility that were popular in the 1960s. For us, then, socially responsible behavior mandated by legislation is at the lowest end of the continuum. At the opposite end of the continuum, we would place socially responsible behavior that is motivated purely by altruism; that is, behavior that goes beyond the requirements of the law, but that also does not consider financial goals. It is simply doing what is perceived as the right thing to do. Behavior in the middle of the continuum is profit-focused; short-term profit that results from lower costs or higher prices, or long-term profit that results from increasing customer loyalty as more customers make their market decisions based on an increasing awareness of and importance for socially responsible behavior. This range of behaviors is focused on achieving competitive advantage. Fleming and Jones (2013) argue that CSR has become a key marketing exercise for most medium- and larger-sized firms.\nWhat is socially responsible corporate behavior?: As early as 2006, Dahlsrud identified 37 definitions of CSR, and even then, his methodology for identifying them understates the true number (Carroll and Shabana, 2010). A definition we find particularly useful is an early one. Carroll (1991), building on an earlier definition (Carroll, 1979), states that the social responsibility of business has four dimensions: legal, economic, ethical and discretionary (philanthropic). This definition is congruent with the motivations we identified for performing socially responsible corporate behavior. Government regulation defines the legal dimension. Firms seeking competitive advantage may do so by seeking increased profits (economic dimension), or by fostering stakeholder support and loyalty (ethics dimension). The discretionary dimension covers those things that a business does simply because they are the right things to do. This sometimes involves philanthropic donations to causes not related to the business, but it may also include any socially responsible behavior not motivated by competitive advantage or by government regulation. We see this as altruism, but of course, it is seldom possible to determine exactly the motivations behind seemingly altruistic behavior. Barnett (2007) addresses this issue and brings forth the business case of the path-dependent nature of firm-stakeholder relations in developing a construct of stakeholder influence capacity to link sustainability and firm financial performance. The literature suggests that there are firms that believe improving their environmental performance can also lead to a better bottom line (Polonsky et al., 1998; Jay, 1990; Langrehr et al., 1992).\nSocially responsible behaviors and corporate motivations: Appendix 1 provides examples of B2B companies selected by the authors from a collection of 179 student research papers that focused on firm environmental sustainability initiatives and were a required component in two B2B marketing courses, one undergraduate and one graduate. Students were constrained to choose either Canadian companies or multinational companies with Canadian operations. The companies they chose cover a broad cross-section of industries and include firms at all value/supply chain levels. For example, Du Pont Inc. produces chemicals and raw materials that it uses to manufacture its own products, and that it also sells to other manufacturers for inputs into the manufacture of their products. Bombardier, a global manufacturer of aircraft and trains, purchases parts from suppliers, as well as manufactures some in-house. Walmart is a retailer that purchases its merchandise from thousands of suppliers, and that it then sells to its customers. Starbucks is a retailer that buys inputs from various suppliers and then creates the products it sells to its customers.\nBeing socially responsible from cradle to grave - practicing sustainability initiatives encompassing the entire value/supply chain: Businesses that seek to be socially responsible can focus on their manufacturing process through to customer product use and final product disposal. That is, they can consider the entire physical life cycle of their product. Appendix 2 provides examples of firms identified from the same collection of 179 student research papers and using the same method that we used to construct Appendix 1. In Appendix 2, our examples illustrate socially responsible corporate behavior based on a product's physical life cycle - the when and how of such behavior.\nExploring when and how business-to-business firms can be socially responsible: In Appendix 2, the horizontal dimension includes the three principal stages in a product's life cycle:\nDiscussion of Appendices 1 and 2: What we have attempted to accomplish with our paper is to bring together a framework that considers a continuum of motivations (meeting legal requirements, seeking competitive advantage and behaving altruistically) for socially responsible behavior (Appendix 1), and when in a product's physical life cycle (manufacturing, consumption and disposal) these behaviors can occur (Appendix 2). While there are many behaviors that may be socially responsible, we have focused on areas that are popular among businesses today: pollution - and how to reduce it or increase recycling; energy - how to improve its efficiency or change to cleaner, more sustainable sources; and materials - how companies can use fewer, safer and more sustainable ones.\nHow do sustainability initiatives add value to the firm?: Current academic thought in addressing the question of how sustainability initiatives result in economic value to the firm suggests that it is through facilitating better relationships with the firm's stakeholders (Peloza and Shang, 2011). As just what \"stakeholders\" means is far from unanimous in the business literature (Miles, 2012), we define stakeholders as \"those firms, organizations, groups, or individuals who directly or indirectly have an interest in or are affected by the subject firm's activities\". With the growing cynicism and distrust of large corporations (Verschoor, 2008), firms that are engaging in meaningful sustainability initiatives must communicate this to their various stakeholders to try to counter any negative perception. A logical extension of this is that marketing and communication tools are an essential part of firms' value creation (Delmas and Toffel, 2008). In the next section, we address this communications issue, which ties in with our paper's third objective of providing guidelines to B2B firms throughout the value/supply chain on how to strategically communicate relevant and successful sustainability initiatives.\nCommunicating sustainability initiatives to the firm's stakeholders: How effectively a firm communicates its sustainability initiatives to its stakeholders can affect the organization's reputation and brand equity (Fraj et al., 2013; Jahdi and Acikdilli, 2009), and marketing - most particularly through public relations - is the logical firm function to spearhead communication with the various stakeholders (Kurtz et al., 2016). However, marketing has an image problem; it is often viewed by those outside the firm (including external stakeholders) with a lot of cynicism and distrust when the firm's marketers report on its sustainability initiatives (Nantel and Weeks, 1996). Compounding this image problem, some marketers are reluctant to communicate their sustainability initiatives to avoid appearing self-aggrandizing to their stakeholders (Peloza et al., 2012). Pomering and Johnson (2009) address this with their research propositions linking socially responsible corporate image and diagnostic dimensions of the CSR image advertising claims. This means that the firm must have credibility in its efforts to get its sustainability initiatives message across to its stakeholders, and the way this is done can have a positive or negative effect on its organizational reputation and brand equity. The communications messages and methods used, however, must be differentiated from those used by firms that are less serious about their sustainability initiatives, and that are simply giving lip-service to environmental sustainability. Dangelico and Pontrandolfo (2010) put forth their Green Option Matrix as a marking tool to analyze competitors' green product offerings and practices for firms to better communicate to their own stakeholders their firm's superiority in the eco-environment compared to the competition. As one of this paper's objectives is to provide guidelines for communications strategies that can assist firms with serious sustainability initiatives, we suggest the following:\nManagerial implications: In the future, firms will find that engaging in environmental sustainability will be a strategic imperative if they wish to increase sales, reduce costs and improve both their short- and long-term profitability. For companies that have not yet or are in the earlier stages of implementing sustainability initiatives, our paper, through categorizing such initiatives, provides a discussion of why, when and how they should do so. At the same time, within our two frameworks, we provide examples of firms that currently are involved in various facets of sustainability, which could be utilized as role models for firms serious about enhancing their sustainability initiatives.\nSummary and conclusions: What we have attempted to accomplish with our paper is to bring together two frameworks that consider a continuum of motivations (meeting legal requirements, seeking competitive advantage and behaving altruistically) for socially responsible behavior, and when in a product's physical life cycle (manufacturing, consumption and disposal), these behaviors can occur. While there are many behaviors that may be socially responsible, we have focused on areas that have been popular among businesses today: pollution - how to reduce it or increase recycling; energy - how to improve its efficiency or change to cleaner, more sustainable sources; and materials - how companies can use fewer, safer and more sustainable ones.\nLimitations and suggestions for future research: There are limitations inherent in all research, and certainly, we are aware that there are limitations here. We have given students the right to select any organization of their choice, with the constraint that it must be a Canadian business, or a multinational organization with Canadian operations. When preparing their reports, students were constrained to written materials concerning the sustainability initiatives of their chosen organization. There were no interviews or surveys of key informants. To take what is available in print and to then accurately ascribe motivation for an organization's behavior is a task that necessitated some personal opinions on our part, and most certainly resulted in some choices that may be questioned. We do, however, believe that we have provided good justification for the examples we have chosen, and that we have accomplished our intent to demonstrate the veracity of our two conceptual frameworks - Appendices 1 and 2 - which together provide a model for empirical testing for those who wish to further explore this line of research.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to address the questions of why, when and how business-to-business (B2B) firms engage in sustainability initiatives. The authors believe that this is the first attempt to address all three questions in a single paper, and one of the earliest to focus on these in B2B markets.\nDesign/methodology/approach: The sustainability initiatives of B2B firms throughout the value/supply chain were examined. Input data came from external sources and the firms themselves. Two conceptual frameworks were developed, illustrating why firms partake in sustainability initiatives and when and how they may do so.\nFindings: This paper provides two conceptual frameworks that address why, when and how firms get involved in sustainability initiatives, and how they can better communicate their involvement to stakeholders.\nResearch limitations/implications: To obtain a broader perspective of B2B firms' involvement in sustainability initiatives, a variety of third-party sources were used, augmented with data from firm websites. Examples of firms the authors selected were constrained by the collection of firms described in student research papers.\nPractical implications: This paper suggests useful guidelines for firms considering starting or expanding sustainability initiatives by providing frameworks that address why, when and how firms do so, with examples of firms illustrating engagement in each area. It also provides communication guidelines, necessary for enhancing stakeholder relations.\nSocial implications: Integrating environmental sustainability within a firm's strategy can improve corporate image and increase efficiency, while contributing to a better world environment.\nOriginality/value: A review of the corporate social responsibility (CSR) literature indicated that most research has focused on business-to-consumer markets. This paper addresses CSR in B2B markets, examining players at all levels of the value/supply chain: manufacturers, channel intermediaries and end-users.\n",
"role": "assistant"
}
] |
Enhancing resilience against floods in the Lower Motowoh community, Limbe, Southwest Cameroon
|
[
"Politics",
"Local/non local knowledge sharing",
"Receptiveness/reticence",
"Collaboration and influence",
"Power or powerlessness"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nBackground: Cameroon is a West-Central Africa country, with a surface area of 475,440 km2 (Figure 1). It is a middle/lower income food-deficient country (2013 GDP per capita of $2,400) with a population (2016 estimates) of ~24.4m, giving a population density of 51 persons per km2. People of 24 years and below constitute about 62 percent of the population. In total, 60 percent of the population - projected to reach 40m in the next 30 years - live in cities. This means that the need for largely under-developed services like electricity, water and housing leads to growth of shanty towns in which people are vulnerable to disasters. Cameroon can be divided into four zones based on climate, geography, vegetation and socio-economy. The disaster risk profile of the country reflects these divisions. Common intensive risks are earthquakes, volcanic eruptions and gas explosions from lakes. Extensive risks are floods, landslides, epidemics, urban fires, conflicts, industrial/transport accidents and social pressures resulting from poverty (Aka et al., 2016). Increasing climate change and variability exacerbate the situation. Temperatures have generally risen throughout Cameroon, with an average annual temperature increase of 0.7degC from 1960 to 2007, representing an average rate of 0.15degC/decade. Predictions indicate that temperatures will be up to +4.8degC and+3.6degC in 2100, respectively in the northern and southern parts of the country. Projections of extreme climate events indicate a further increase in the frequency and magnitude of droughts, erosion and floods. Mass movements (landslides, mudslides, rock-falls, earth collapses, etc.) will also increase in intensity, particularly in mountainous areas. It is projected that sea level will rise by 9 to 38 cm by the year 2050, and by up to 80 cm by 2100. Consequences of this include disappearance of trees, which will speed up erosion and cause degradation of mangrove ecosystems. Floods, splashes, storms, sedimentation, soil erosion and sediment load in rivers will increase (Cameroon National Adaptation Plan to Climate Change (CNAPCC)).\nThe case study: One important outcome of the VFL program carried out by GNDR and her CSO partners including GEADIRR was that in order to strengthen disaster resilience and policy making, local knowledge is indispensable. However, views of local people are often not taken into consideration enough, so GEADIRR participated in the GNDR initiated-Frontline process in Cameroon with the aim to gather local risk knowledge and priorities, support local learning and action, contribute to the national frameworks for policy making and effective implementation, and participate in the global baselines and DRR monitoring mechanisms.\nDedication: We dedicate this paper to Chief Enow John and Pa Vando Pius, respectively Quarter Heads for Lower Motowoh and Unity Quarters who gave us tremendous support since 2009 in our DRR work with local communities in Limbe. Chief Enow John passed away in March 2016 while Pa Vando Pius passed away in Jan 2017.\n",
"role": "user"
},
{
"content": "Purpose: Floods and landslide threats were addressed under the Frontline process in the city of Limbe, Southwestern Cameroon. The purpose of this paper is to present actions undertaken through building local community resilience to floods which are a major threat in the city, with impacts on the local community ranging from death to complete destruction of services and livelihoods.\nDesign/methodology/approach: The actions carried out were informed by the GNDR-supported Frontline survey conducted in 2015 in which the Lower Motowoh community rated floods as an important threat. A series of reflection and learning sessions with the community members was carried out to better understand the problem. Scoping studies on the causes and extent of floods along river Njengele were undertaken by GEADIRR and the community team.\nFindings: The findings of this paper indicated that the main problem resulted from river channel blockage caused by indiscriminate dumping of refuse into the waterway and sediment deposition from upstream. Further reflection and action planning led to preparatory meetings between GEADIRR and ten community leaders. The unanimous action adopted was to dredge the river. Dredging was carried out in late April and early May of 2016 using a hired bulldozer.\nSocial implications: Follow-up shows that after many years of misery from floods, often associated with the loss of loved ones and property, about 500 community residents who benefited from the action did not go through this dreadful ordeal again during the rains of 2016 and 2017. People are currently rebuilding on the reclaimed land which was previously abandoned due to flooding.\nOriginality/value: Current challenges include changing the mindset of community members about the adverse effects of indiscriminate dumping of household waste into the waterway. It was also a big challenge convincing some members of the community who felt that floods are a natural phenomenon unstoppable by man.\n",
"role": "assistant"
}
] |
Research and practice: A critical reflection on approaches that underpin research into people's information behaviour
|
[
"Information behaviour",
"Phenomenology",
"Epistemology",
"Methodology",
"Social constructivism",
"Action research",
"Participative design"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper provided an opportunity to reflect on research into people's information behaviour and, in turn, to attempt to clarify the range of approaches and associated methodologies and research techniques. As Marcia Bates states (1999, p. 1043): \"We [information scientists] need to become more fully conscious of the research and practice paradigm from which we operate [...] much of the paradigm of any field lurks below the water line, largely unconscious and unarticulated, even by its practitioners [...]\".\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to provide a critical discussion on the nature of research into people's information behaviour, and in particular the contribution of the phenomenological approach for the development of information solutions.\nDesign/methodology/approach: The approach takes the form of a conceptual analysis drawing on the research literature and personal research experience.\nFindings: The paper brings to the foreground the relative value of different conceptual approaches and how these underpin and relate to the development of information solutions.\nResearch limitations/implications: The paper, due to the breadth and complexity of the subject, serves to highlight key issues and bringing together ideas. Some topics deserve further explanation. However, this was beyond the scope of this paper.\nPractical implications: A conceptual framework is provided that indicates the value of the epistemic spectrum for information behaviour studies and provides support for action research and participative design.\nSocial implications: Taking a phenomenological approach, and consequently either a first person approach and/or a highly participative approach to research, challenges the relationship between researcher and respondent. It also raises questions about why the authors conduct research and for whom it is intended.\nOriginality/value: The paper makes explicit the underlying philosophical assumptions and how these ideas influence the way the authors conduct research; it highlights the significance of Cartesian dualism and indicates the significance of these assumptions for the development of information solutions. It supports the view that researchers and developers should be open to respondents leading the exploration of their needs.\n",
"role": "assistant"
}
] |
An acquisition or a greenfield subsidiary? The impact of knowledge on sequential establishments in a host country
|
[
"Subsidiary",
"Acquisition",
"Greenfield",
"Experience",
"Phase",
"Sequential"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: A multinational firm that expands in a host country by establishing one wholly owned subsidiary after another generally builds a knowledge advantage (Ogasavara and Hoshino, 2009). Furthermore, sole ownership signals high commitment and facilitates firm's marketing activities in the local market. Such sequential establishments, thus, reduce costs through sharing marketing-related knowledge across the subsidiaries (Gao et al., 2008; Jiang et al., 2014; Song, 2002).In accordance with knowledge-based theory (Arend et al., 2014; Barney, 1991; Grant, 2002), the firm needs to align the mode of establishment with relevant knowledge in order to exploit the potential advantage in a host country. Any misalignment will be problematic because the firm will encounter difficulties to adjust the high commitment mode. For example, once an acquisition is finalized it will be troublesome for the firm to adjust the terms. It will also be difficult to withdraw if the firm has set up local activities on its own regarding, for example, R&D and production. Such local activities represent value adding beyond sales and reflect firm's efforts to create more value for customers and to offer customized products and production.However, international marketing literature does not pay enough attention to the way experience affects sequential choices between foreign establishment modes where the firm is the sole owner. Therefore, this article explores relationships between a firm's marketing-related experiences and modes of sequential establishments of wholly owned subsidiaries in the same host country. Regarding full ownership, two establishment modes are relevant (Brouthers and Hennart, 2007), namely, an acquisition in which the firm enjoys fast market penetration and a greenfield investment in which a new organization is built from scratch. Both alternatives mean extensive commitment making it difficult to make adjustments (Dow and Larimo, 2011; Slangen and Dikova, 2014).The review by Dikova and Brouthers (2016) shows that general explanations of firm's single choice of mode include factors related to knowledge. Some studies highlight international experience (e.g. Dow and Larimo, 2011; Harzing, 2002; Mudambi and Mudambi, 2002), but they found no consistent relationship with establishment mode. Other studies pay attention to mode experience (Schwens et al., 2018; Slangen and Hennart, 2008) and show that firms sometimes tend to reduce risks by repeatedly using the same mode or host country experience (e.g. Barkema and Vermeulen, 1998; Chang, 1995; Chen, 2008; Dow and Larimo, 2011; Drogendijk and Slangen, 2006; Schwens et al., 2018). However, the latter studies display mixed relationships with mode.Because previous studies exhibit equivocal results, it is important to go deeper and recognize phase of establishment in a host country. Two knowledge-based studies (Chang and Rosenzweig, 2001; Zhang et al., 2007) capture the phase and the former shows that explanations of an initial mode differ from explanations of subsequent modes. Thus, initially a greenfield investment seems to be effective and there is a strong relationship with the firm's international experience. For subsequent establishments in the same product area the firm tends to use the same mode as before and, thus, relies on mode experience. However, Zhang et al. (2007) display conflicting results concerning impact of knowledge types. This study found that the extent of a firm's international experience and host country experience is crucial both initially and in postinitial phases of establishments in an emerging economy. Initially the firm prefers a joint venture and later an acquisition, or a greenfield investment. Thus, in light of conflicting results research needs to explore further relationships between knowledge types and modes of sequential establishments in the same host country.Other knowledge-based studies of relevant sequences focus on: (1) factors that are associated with subsidiary performance (Barkema and Drogendijk, 2007; Jiang et al., 2014; Ogasavara and Hoshino, 2009; Gao et al., 2008), (2) the impact of experience on establishment of a second subsidiary (Pehrsson, 2016) and (3) relationships between establishment speed and mode (Gao and Pan, 2010). However, all studies suffer from a lack of focus on the choice between wholly owned subsidiaries established via acquisitions or via greenfield investments. While most studies do not recognize the difference, Chang and Rosenzweig (2001) and Zhang et al. (2007) include acquisitions, greenfield investments and joint ventures. However, including joint ventures and shared ownership introduces complexity (Brouthers and Hennart, 2007). For example, two or more parent firms mean that the parents may not view the venture's legitimacy equally (Kostova and Zaheer, 1999), and parents' knowledge affecting choice of mode may differ making way for disagreements.The review above underscores two major gaps. First, more research is needed about the impact of knowledge types on sequential establishments of wholly owned subsidiaries in a particular host country. Second, further studies need to focus on wholly owned subsidiaries before including shared ownership and knowledge of more than one parent firm. Hence, it is necessary to answer this question: how does firm's knowledge related to marketing affect the choice between an acquisition and a greenfield investment in regard to sequential establishments of wholly owned subsidiaries in the same host country?Drawing on knowledge-based theory, the purpose of this study is to extend the current understanding of firm's modes of sequential establishments of wholly owned subsidiaries in a host country. This study captures relationships with knowledge relevant to marketing since the knowledge operates through the mode in a host country and, therefore, an acquisition and a greenfield investment needs to be aligned with knowledge. Furthermore, this study captures experiential knowledge (Schwens et al., 2018) of manufacturing firms in terms of international experience, mode experience and value-adding experience. This study takes a longitudinal approach and analyzes 288 establishments of Swedish firms in Germany, the United Kingdom and the United States.The results of this study contribute to the understanding of the sequential choices of a firm in several ways. First, the results contribute to knowledge-based theory because this study demonstrates the relevance of experiential knowledge when it comes to a firm's expansion in a host country. Second, this study shows that the importance of different experiences varies between phases of establishing wholly owned subsidiaries through acquisitions or greenfield investments. Third, this study reveals the crucial impact of timing and the need to align the establishment mode with the phase of establishment in a host country. Fourth, this study introduces firm's local experience of value adding and the results support an association between the experience and a postinitial establishment through an acquisition. Experience connected to local activities such as R&D and production facilitates smooth and flexible responsiveness to needs of customers in the host country. Acquisition of knowledge is aligned with a constant updating of a responsiveness advantage and limited loss of time. Fifth, surprisingly the results do not support any association between experience and a greenfield investment in the postinitial phase.\nTheory base and hypotheses: Application of the knowledge-based theory\nMethods: Sample and data collection\nResults: Descriptive statistics and correlations\nDiscussion of the results: Theoretical contributions\nLimitations and future research: This study suffers from limitations. First, a reliance on other sources of experience might have been relevant in order to examine the association between experiential knowledge and a postinitial greenfield investment. Second, further research is needed on firms' strategic intentions to extensively invest in a particular host country. For example, what is the importance of a firm's product scope? Does a broadening of the scope require an additional subsidiary in the target market? Such questions call for an evaluation of relatedness among involved business units and possibilities to enjoy synergies.Third, this study did not control for effects of the size of a subsidiary, which may affect issues related to international expansion (Beamish and Inkpen, 1998). Instead, firm size was used since this was assumed to represent the experience that influenced the firm's choice of mode at the time of subsidiary establishment in a particular country. However, for postinitial entries in the same country, the size of previous subsidiaries indicates market experience that may have affected subsequent entries. Such size data were not accessible; thus, size effects need to be examined further.Fourth, biases could have occurred due to the host countries of the sample firms. It would be interesting to incorporate the choices of modes of establishments in foreign markets with varying cultures. For example, Dow and Larimo (2009) found that experience regarding countries similar to the target country are associated with the establishment of a wholly owned subsidiary, and experience concerning countries with different national cultures strengthens that link.Fifth, further research is needed on the impact of an industry's phase of development. It may be the case that a mature industry consists of experienced firms that encounter less uncertainty in the environment. For example, a firm operating in a mature industry will be familiar with the behaviors of customers and competitors, and with potential acquisition candidates. Therefore, the likelihood of an acquisition may increase.Sixth, a focus on the establishment of types of investments other than wholly owned subsidiaries might generate different findings. Future research may, therefore, draw on theory bases other than the knowledge-based theory, and may analyze a broader set of establishment modes including, for example, joint ventures. Provided joint ventures are highlighted, it would be interesting to scrutinize the relative impact of experiences from the previous establishment paths and corporate strategies of each partner firm.\nConclusion: This study extends the current understanding of a firm's sequential choices between acquisitions and greenfield investments as a way of establishing wholly owned subsidiaries in a particular host country by revealing associations with experiential knowledge of the firm. With the knowledge-based theory as its starting point, this study performs an examination that contrasts the effects of broad international experience on mode of establishments with the effects of locally bound mode experience and value-adding experience. It was found that associations with experience vary between greenfields and acquisitions, and from one phase to another. The results close two major research gaps. First, conflicting results of previous studies regarding the impact of experiences are sorted out by recognizing the role of phase of establishment. The results also close a gap of sequential studies regarding firm's alternatives since the results focus on wholly owned subsidiaries established through acquisitions, or greenfield investment.\n",
"role": "user"
},
{
"content": "Purpose: Establishment of wholly owned subsidiaries in a foreign market is central to international marketing because sole ownership and high commitment facilitate firm's marketing in the local market. Drawing on knowledge-based theory, this study extends the current understanding of firm's sequential establishments of wholly owned subsidiaries in a host country.\nDesign/methodology/approach: Swedish firms' establishments of wholly owned subsidiaries in Germany, the United Kingdom and the United States were analyzed using a longitudinal approach.\nFindings: A firm's broad international experience is associated with an acquisition in any phase, while mode experience and value-adding experience are associated with postinitial acquisitions. There is no association between mode experience and greenfield investments.\nResearch limitations/implications: Knowledge-based theory explains a firm's choice of establishment mode when establishing in the same host country. Effects of marketing experiences are due to the establishment mode and different experiences explain choices for initial and postinitial establishments.\nPractical implications: In choosing between a wholly owned subsidiary in terms of an acquisition or a greenfield investment, for a foreign establishment the firm is advised to consider the impact of marketing experiences and establishment phase.\nOriginality/value: Research is needed on how experiences affect choices between foreign establishment modes where the firm is the sole owner. This study is the first to focus on the choice between wholly owned subsidiaries in terms of acquisitions and greenfield investments, and the impact of experience and phase of establishment in a particular host country.\n",
"role": "assistant"
}
] |
In "the zone": The role of evolving skill and transitional workload on motivation and realized performance in operational tasks
|
[
"Human resource management",
"Motivation (psychology)",
"Skills",
"Employee productivity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: How workers respond to the variety of operating conditions they are exposed to throughout the course of a day, let alone the full length of their employment in a particular position at a firm, have significant consequences for the nature (both quantity and quality) of the work they contribute to. This in turn has direct implications for the profitability of firms that rely critically on the effectiveness of these workers. It therefore comes as no surprise that the field of operations management has become increasingly interested in factors and dynamics that drive productive activity at this most fundamental level of the individual worker (New, 1998; Bendoly, 2006; Wingaard and de Vries, 2006). A testament to this increased interest is demonstrated by the recent appearance of highly ranked journal special issues and departments devoted to behavioral dynamics in operations (e.g. as per JOM, MSOM and POMS), as well as the spawning of academic interest groups and colleges in well-established OM academic societies.Rather than continuing to model individuals as a set of homogeneous automatons, or even as a collective of agents that follow simple and fairly predictable work patterns, empirical research in this area has begun to attempt to tap into the full range of behavioral dynamics observable among real humans in various tasks emblematic of OM practice (Schultz et al., 1999; Croson and Donohue, 2006; Fransoo and Wiers, 2006). These investigations help refine, occasionally support and, often refute past modeling assumptions and hence present essential contributions that truly can help close gaps between operations research and effective management practice. Yet there clearly remains a vast terrain of behavioral dynamics in operations that remains under-investigated and misunderstood.A particularly nebulous, though often discussed, dynamic is the relationship between the pressure placed on workers and their productivity. In part, this lack of clarity is due to the various means by which, and variety of settings within which, concepts such as pressure and productivity can be operationalized. The concept of \"pressure\" in particular poses a challenge in that most arguments around its effects are based on how individuals react to their \"perceptions of pressure\" - an issue that cannot be objectively observed and can only be assessed through proxies provided either by observable work conditions thought to be conducive to pressure, or by subjective assessments provided by those supposedly experiencing it.In the operations literature, Schultz et al. (1998, 1999), for example, use inventory level as a measure of workload and hence a proxy for situational pressure. In contrast, in their analytical modeling of learning in kanban systems Suri and De Treville (1986) assert that kanban-driven disruptions (forced idleness) imply uncomfortable work conditions that, at least at a moderate level, workers may feel can be avoided through more effective work. They imply that the frequency of these events are hence suitable proxies for work pressure. Still others consider the existence of visible regulations on work expediency, cost, quality and the use of monitoring as pressure inducing, and that the presence of such systems themselves can be proxies for more persistent levels of work pressure (Cooper and Slagmulder, 1997; Everaert and Bruggeman, 2002; Fransoo and Wiers, 2006; Bendoly and Hur, 2007).In our present study, we most closely follow Schultz et al.'s (1998, 2003) lead in our operationalization of pressure in part as a function of transitional levels of workload. However, rather than assume that objectively observable levels of workload have a simple linear impact on pressure and subsequent effort, we incorporate the non-linear notions posed in related research as well (Suri and De Treville, 1986; Bendoly and Hur, 2007). Specifically, through use of a behavioral experiment, we empirically test for the existence of an inverted-U effect between objectively observable workload and pressure-driven motivation to exert effort towards high performance in an operations management task.At the same time, we also go beyond these relatively static models of non-monotonicity to suggest that the workload level conducive to peak effort and performance is itself a function of skill. Following again the suggestion of Suri and De Treville (1986) that learning is also likely to be an artifact of pressure-driven effort, we empirically demonstrate a feedback system in which the ideal workload level (i.e. \"the zone\") associated with peak effort should be expected to shift as learning takes place. Accordingly, learning eventually may give rise to lower levels of effort if workload is not modified commensurately to maintain motivating levels of work pressure. The implications for such a dynamic mechanism on both work modeling and work assignment strategies by higher level managers are discussed.\n2 Theoretical background: The operations research community has long allowed the assumption of \"stasis\" in average levels of individual performance (e.g. average work rate, m), and hence the assumption that such static characterizations are acceptable in modeling the performance of larger operating systems (Bendoly et al., 2006). At best, performance has been described as a potential function of skill and associated learning (Carrillo and Gaimon, 2004) thus reflecting, at least in part, the dynamics of real-world activity. Unfortunately, such recognition alone still falls appreciably short of reality, and the acceptance of such limited views of behavioral dynamics in the workplace continues to limit the ties between modeling and practice (Bendoly and Speier, 2008). However, there are signs that this is now changing in the literature.Across all management disciplines, including OM, it is now well established that individual performance is only in part a function of individual skill. The other most common element brought up in discussions of antecedents of workplace performance is motivation. Motivation and skill both clearly interact to give rise to observable levels of performance (Locke et al., 1978; Niepce and Molleman, 1996; MacCarthy and Atthirawong, 2003; Yeo and Neal, 2004). Yet the true nature of their interaction, particularly over time, remains obscure. In part this obscurity is due to the paucity of research to date that has attempted to empirically examine multi-period and potentially non-linear dynamics associated with these and other critical mediating factors such as work pressure (Kanfer and Ackerman, 1989; Moorman and Matulich, 1993).2.1 Interactions and non-monotonicities in antecedents to performance\n3 Experimental design: Following the success of recent laboratory experimental methods as a re-emerging paradigm in operations management (Bendoly and Swink, 2007; Croson and Donohue, 2006; Sterman, 1989), we use a controlled experiment to test for the impact of both skill and workload, both initially and as they may change over time, on motivation and associated levels of task performance. The use of a controlled experiment in particular is a convenient mechanism for capturing objective data related to alternating human behavior over a multi-period framework (Podsakoff et al., 2003). Such methods avoid biases which are often crippling to single-source single-period survey methods which often rely predominantly on retrospective and subjective accounts (Bendoly et al., 2006). The context of the laboratory task in this case was that of a fairly standard traveling salesman (vehicle routing) problem. The task itself was instantiated through the use of a computerized graphical user interface, intended to assist in developing feasible routes.The interface provided a means by which to develop solutions to a set of routing problems, as well as feedback regarding workload levels (both in terms of the rate of new work arrival, one of our treatments, and level of jobs in queue). It also provided a running total of points earned (based on both quality and completion rate for each solution) and a 7-point Likert-type scale for providing self-assessments of motivation levels (required for each solution submission). Along with three treatment levels on workload, we also employ a two level training treatment (to help distinguish initial on-task skill levels among participating subjects). Since accounting for intrinsic individual differences can be critical in ensuring the effectiveness of treatments, we describe in the following sections the steps taken to ensure our treatments were calibrated to avoid spurious confounds in our analysis.3.1 Calibration phase\n4 Data characteristics: Our pool of subjects included 414 managers enrolled in evening executive education programs at a top ranked US business school. The choice of this pool helped guarantee work experience and practical interpretability relevant to the nature of the task, while simultaneously providing an opportunity to approach these individuals for recruitment purposes. Of the 414 potential subjects in our pool, 156 were willing to fully and voluntarily participate in the experiment and subsequent retrospective questionnaire (37.7 percent response rate). The average number of years of management experience across the respondents was 2.32 (standard deviation 1.91). Industry representation varied greatly though the primary discernable sectors represented were logistics (27 percent), information technology (18 percent), marketing (14 percent) and financial services (16 percent). Sixty-five percent of the sample population was male.4.1 Base-rate estimation\n5 Model analysis and results: 5.1 Generalized treatment effects\n6 Discussion: 6.1 Contribution to OM literature\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to examine how training specific to a given operational task, and subsequent experiential learning, can heighten skill and hence shift the level of workload at which individuals are most productively motivated.\nDesign/methodology/approach: - To analyze these effects, a laboratory experiment was used involving a vehicle routing application and 156 managers exposed to a 2 x 3 complete treatment design. Both multi-period objective in-task data and subjective self reports are collected to tap into skill levels, actions and behavioral variables of interest.\nFindings: - In the absence of additional workload challenges, the paper finds that increases in skill may in fact significantly limit and in some cases actually degrade overall motivation, as well as objective performance.\nResearch limitations/implications: - Limitations potentially stem from specific operationalizations of the factors studied as well as selectivity of the subject pool and the context (vehicle routing task).\nPractical implications: - The implications of the skill-challenge-motivation dynamics observed have direct repercussions for existing management models in which training and experience are viewed as having strictly monotonic benefits to performance. The implications also go far to promote more informed models of worker behavior in operations modeling that otherwise view performance as static or monotonically increasing based on experience.\nOriginality/value: - This is believed to be the first study that has explicitly studied the inverted-U dynamics stemming from the interplay of both skill and workload on motivation and performance, over a multi-period framework of analysis.\n",
"role": "assistant"
}
] |
Re-storying an entrepreneurial identity: education, experience and self-narrative
|
[
"Entrepreneurship education",
"Narratives",
"Identity",
"Higher education",
"Individual psychology",
"Entrepreneurialism"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Entrepreneurship education is unique in that it directly connects the individual, with his or her particular interests, knowledge, experience and social networks with the marketplace in which he or she seeks to gain acceptance, implement plans, perform commercial transactions, interact with stakeholders and develop a project, business or organization. This fortuitous confluence would seem to make the entrepreneurial classroom a potentially powerful identity workspace, namely a place where individuals construct, revise and reconstruct their narrative identities (Ibarra and Barbulescu, 2010; Ibarra, 2007, 1999; Petriglieri and Petriglieri, 2010).This paper explores how the entrepreneurial classroom might serve as an identity workspace for participants. It proceeds as follows. First, I present a literature review on entrepreneurship education, laying out what the field has uncovered in terms of context, content and process and explaining how this study fills a gap in the existing literature. Next I explore the theoretical foundations, with particular emphasis on Dewey's theory of experience and education and recent literature on identity workspaces. I conclude with a discussion of the implications for both theory and practice.\nEntrepreneurship education in the literature: In this first section, I highlight the literature addressing the context, content and process of entrepreneurship education to date, following a clear pattern of development in the field around these three areas of analysis. In the earlier days, when entrepreneurship was just beginning to be offered in universities, typically as a \"sideshow\" in the overall management curriculum (rarely gaining the status of a stand-alone department), the focus of much of the entrepreneurship education literature was on documenting the increasing prevalence of entrepreneurship programs in universities (Vesper and McMullan, 1988, Solomon and Fernald, 1991), in non-university contexts in the developing world (Brockhaus, 1991) or on rather general suggestions for how to improve the offerings (Hills, 1988). These suggestions were largely based on why and how entrepreneurship education is different than, and should not be placed within, the functional framework of, general management education. (McMullan and Long, 1987). As a result of these earlier efforts, the full prevalence and legitimacy of entrepreneurship education has recently been recognized in the entrepreneurship literature, as described by Katz (2008, p. 550):The major consequence of entrepreneurship's full maturity is identified as the growing centrality of the business-school based discipline of entrepreneurship in relation to the emerging entrepreneurship efforts across campuses.Over time, due to both increasing \"enlightenment\" on the part of university faculty and administration and increased market demand (Solomon and Fernald, 1991), the focus shifted to more nuanced analyses of the actual content and process of entrepreneurship education programs (Gartner and Vesper, 1994, Vesper and Gartner, 1997) and controlled experiments to compare entrepreneurial attributes of entrepreneurship students vs. those same attributes in other management students (Kolvereid, 1996). The trend to analyze in greater depth both the content and role of entrepreneurship education continued with a number of studies looking at the overall vision for entrepreneurship education (a particularly interesting one, Jones-Evans et al. (2000), suggested creating an \"entrepreneurial analogue to the teaching hospital\"), the role it might play outside of the university setting (Gray and Allan, 2002) and a more recent paper that suggests a shift in the measures of success for entrepreneurship education away from the number and quality of start-ups toward a change in the societal and intellectual attitudes toward entrepreneurship in the long run. (Galloway and Brown, 2002).A number of large-scale quantitative studies have come out more recently that attempt to measure the effect of entrepreneurship education on intentions, attitudes and perceptions (Galloway et al. (2005); DeTienne and Chandler (2004); Peterman and Kennedy (2003); Galloway and Brown (2002); and Hindle and Cutting (2002) are examples). Particularly relevant to the aim of this paper, one of these (Galloway et al., 2005) explicitly states the serious limitations of quantitative studies in evaluating attitudes toward entrepreneurship and the economic environment, arguing that while a study of that sort may be able to suggest a generally positive impact from entrepreneurship education, and therefore implies that investment in entrepreneurship education is likely warranted, it does not allow for a more fine-grained examination of exactly what is having an impact on students, why and how. Finally, a recent study published in AMLE (Edelman et al., 2008) compares start-up activities of nascent entrepreneurs in the Panel Study of Entrepreneurial Dynamics (PSED) dataset to data collected from a sample of entrepreneurship textbooks and finds that while there is some overlap in start-up activities practiced by nascent entrepreneurs and those covered in textbooks, there are many differences to be found.Perhaps in response to the dearth of meaningful analysis of the actual content of entrepreneurship education programs, two theoretical pieces (Honig, 2004; Shepherd, 2004) attempted to take a more rigorous look at course content and to tie specific elements of entrepreneurship courses much more closely to the larger body of existing theoretical literature in entrepreneurship. For example, Honig (2004) writes:In particular, practical programs that provide real-world experience seem particularly useful in enhancing intentionality through increased perceived desirability and feasibility (Peterman and Kennedy, 2003). Unfortunately, the literature attempting to systematically connect much entrepreneurial formal or traditional education to entrepreneurial activity or performance is virtually nonexistent (Autio et al., 1997).Honig also uses Sarasvathy's (2001) notion of entrepreneurial effectuation and a wealth of literature on business plans in entrepreneurship to argue that business plans are ubiquitous not because of their usefulness, but rather because, essentially, that is just the way things have been done. In academic terms, he describes it as a \"normative isomorphism influencing nascent entrepreneurs to produce business plans\" (Honig, 2004, p. 260). He concludes that while a business plan is \"meant to be a first step toward a specific process widely known as entrepreneurship [...] unlike the activity of entrepreneurship, it focuses primarily on ideas as opposed to actions\" (Honig, 2004, p. 260). This conclusion is based on the work of Sarasvathy (2001) and others who propose that entrepreneurship is not the result of positive deductive rationality which is reflected in typical business plan approaches, but rather consists of an \"inductive process by which various products, services, and ideas are examined, attempted, modified and delivered\" (Honig, 2004, p. 260).Shepherd, following up on an earlier piece on the emotions surrounding entrepreneurial failure (Shepherd, 2003), looks at the role of emotions and failure in the entrepreneurial classroom (Shepherd, 2004). He suggests specific - although untested - pedagogical approaches for introducing the emotions of failure into the classroom.Both of these articles make serious attempts at merging theory, practice and actual observation of what entrepreneurs do, and how they behave. They represent a good start at a more rigorous approach to bringing theory and practice together in the entrepreneurial classroom and offer potentially valuable pieces of information on the content of entrepreneurship education. However, as Sardana and Scott-Kemmis (2010, p. 441) very recently stated:Despite a proliferation of research in the field of entrepreneurship, our understanding of entrepreneurial learning remains limited. We do not have systematic answers to many key questions.Those authors were looking at learning in biotechnology ventures and concluded that both the context and the content of entrepreneurial learning need much deeper examination. Likewise, Leitch and Harrison (1999, p. 83) point to a:[...] significant reawakening of interest in the applicability of action learning [...] against a background of a wider re-examination of the theory and practice of management education and development. Fundamentally, this re-examination has been stimulated by criticism of the static, content-oriented approach as inappropriate to the changing environment [...] the implications are twofold: first, the range of teaching techniques must be extended to include process-oriented approaches; second, and more fundamentally, there is a need for an expanded definition of the learning arena itself.To that end, as far as context, there have been very few examinations of the effect of entrepreneurship education in non-traditional (e.g. non-business school) settings. Nafukho (1998) did a survey of Entrepreneurial Skills Development Programs (ESOPS) for unemployed youth in a number of African countries, concluding that while these programs have some effective elements, they will fail to have a meaningful effect until African governments must provide an appropriate business environment which favors entrepreneurship and small business development, and Peterman and Kennedy (2003) did a pre- and post- course evaluation of changes in the perceptions of a sample of secondary school students enrolled in the Young Achievement Australia (YAA) enterprise program, reporting significantly higher perceptions of both desirability and feasibility of starting businesses as a result of participation in the course.With regard to content, there are, as Honig (2004) also pointed out, virtually no studies that examine the lived experiences of students who have actually gone through an entrepreneurship education program in order to better understand the process of transformation (if any) that they undergo. This study, which combines both the non-traditional setting and a deeper examination of the potentially transformative effect of entrepreneurship education in participants' identity, clearly fills a gap in the literature in terms of both context and content.Supported by Dewey's work on experiential learning, and specifically, his idea of intervening terms between remote ends and given means, I will show how entrepreneurship education works to transform the participant's ability to operate in remote (entrepreneurial) worlds, and how holistic entrepreneurial action reinforces this transformation. And supported by work from Ibarra (multiple cites) and others on narrative identity and identity workspaces, I will show how the entrepreneurial classroom can serve as a powerful identity workspace, leading participants to self-identify as entrepreneurs and transforming their view of themselves and their futures.Nelson Goodman's \"World Versions\"\nDiscussion: entrepreneurship education as identity workspace: A burgeoning body of work in management scholarship has evolved around the idea of identity work, meaning the activities that individuals undertake to develop, maintain and exhibit both personal and social identities (Petriglieri and Petriglieri, 2010; Snow and Anderson, 1987; Sveningsson and Alvesson, 2003). This work, \"which is both intrapsychic and social in nature\" involves individuals \"crafting, protecting, and modifying their views of themselves, as well as gaining social validation for those views\" (Petriglieri and Petriglieri, 2010). Identity work is understood to be an ongoing process, but one that intensifies during particular life junctures and transitions (Ibarra, 1999; Ibarra and Petriglieri, 2007; Pratt et al., 2006). While there is quite a substantial body of literature on identity work, a recent paper by Petriglieri and Petriglieri (2010) introduced the concept of the identity workspace, which they define as \"an institution that provides a holding environment for identity work\" (Petriglieri and Petriglieri, 2010, p. 44).Interestingly, the authors go on to explore an educational institution, namely business schools, as ideal identity workspaces and suggest the importance of \"developing management education that goes beyond influencing what managers know and do, and supports them in understanding and shaping who they are\" (Petriglieri and Petriglieri, 2010: 44, emphasis added).Within the context of business schools and other non-traditional educational settings, entrepreneurship education can serve as an ideal identity workspace, a place where participants can try on different identities and learn more about who they are. In fact, entrepreneurship education specifically would seem to be a particularly promising identity workspace with its unique ability to connect the individual, with his or her particular interests, knowledge, experience and social networks with the marketplace in which he or she seeks to gain acceptance, implement plans, perform commercial transactions, interact with stakeholders and develop a project, business or organization.If indeed one \"buys into\" the notion that the entrepreneurial classroom is a place where important identity work can and should occur, this study would suggest a number of ways to encourage that work, some of which involve changing the focus of existing entrepreneurship education programs. Until very recently, as Honig (2004) pointed out, entrepreneurship education has largely revolved around the business plan. If the model presented in this paper has merit, and if it truly reflects the potential for entrepreneurship education to help the participant perform vital identity work, then the focus on business plans has been misguided. This model would instead suggest exposing students to narratives of how entrepreneurial actors have created new artifacts - New Ends, and then giving students the kinds of tools (New Means) - language, social skills, behavioral models, etc. - to move toward a place in the other worlds they find intriguing. Finally, the model entails allowing students the possibility of having holistic entrepreneurial experiences - performing market transactions, presenting their plans to professional audiences, working alongside real entrepreneurs and their day to day businesses - such that they begin to see themselves as entrepreneurial actors in their own right - creating new \"Re-storied\" selves, new entrepreneurial identities.Second, the process of entrepreneurial re-storying as described in this paper is an evolving one and understanding it helps us in an important way to understand in general the ways in which self-narratives form. Here, the process has to do with exposure to new worlds, access to the tools for operating in those worlds, action/doing, and an emerging self-narrative of entrepreneurial identity. If one important component of taking on the role of an entrepreneur, with all that entails, is the actual narrative process of becoming an entrepreneur, of assuming an entrepreneurial identity, then surely more work should be done on the potential effect of entrepreneurship education on identity transformation.Finally, and perhaps most importantly, this paper illuminates the powerful potential of entrepreneurship education not only to provide participants with new knowledge and new tools, but also to help them have new experiences and develop new identities - to discover not only who they are, but also who they can become. This may be the greatest impact of an exploration of entrepreneurship education as identity workspace - it may bring urgency and relevance, and a desire to learn, to people and places where those attributes have until now been in short supply.\n",
"role": "user"
},
{
"content": "Purpose: - This paper aims to explore the ways in which entrepreneurship education may serve as an identity workspace.\nDesign/methodology/approach: - This is a conceptual/theoretical paper based on previously completed empirical work.\nFindings: - The paper makes the connection between worldmaking, experience, action and identity.\nPractical implications: - The paper furthers understanding of entrepreneurship education and its potential effect on the identity of participants. It stresses the importance of offering entrepreneurship education participants the opportunity to take entrepreneurial action. It has implications for the existing state of entrepreneurship education, e.g. the focus on business plans in the absence of an exploration of the identity of participants.\nOriginality/value: - The paper is an original exploration of the linkage between entrepreneurship education and identity and has implications for both pedagogy and practice.\n",
"role": "assistant"
}
] |
Servitization of manufacture: Exploring the deployment and skills of people critical to the delivery of advanced services
|
[
"Service systems",
"Operations strategy",
"Manufacturing",
"Successful manufacturers",
"Product-service systems",
"Servitization",
"Operations management",
"Manufacturing systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: The evolution of manufacturing industry has seen significant changes in the organisation and expected skill-sets of workers. During the industrial revolution of the 1700s, for example, the introduction of the assembly line meant that relatively unskilled men and women (who were trained to carry out small repetitive tasks) could significantly increased their output over artisans in the conventional craft based systems. Ford (1922) is well known for using such systems to revolutionise automotive manufacture. Today, the process of servitization (where manufacturers seek to build their revenues through service provision) is similarly enabled by workers who are organised and skilled in a particular way. This paper explores this topic.\n2 Context and questions: Servitization is the process of transforming manufacturers to compete through product-service systems (PSSs) rather than products alone (Baines et al., 2007, 2009a, 2010). The commercial and environmental benefits of PSSs are compelling and well documented (Rolls-Royce earning over 50 per cent of their revenue from services is cited to exhaustion). The opportunities are immense (three quarters of wealth world-wide is now created through carrying out services) and so politically PSSs are seen as key to industrial success in the twenty-first century.\n3 Research design and execution: In this paper we deal specifically with those people engaged in the delivery of advanced services, their behaviours, and particularly how this is fostered through their organisation and skill-sets. As mentioned above this has been a strand of a macro-study, and at the onset we posed four specific questions, namely:\n4 Behaviours and skill-sets of people delivering advanced services: Across our cases we sought to identify that behaviour that each organisation values highly amongst people that deliver advanced services. Here, we were mindful that these would vary for workers in differing roles (front-line versus support staff), that there are basic hygiene skills that are required of all workers (e.g. an ability to work safely), and that only a few staff across all cases would fulfil all these expectations.\n5 Organisation and deployment of people delivering advanced services: Across the cases, the policy is to co-locate (most) people who are responsible for the delivery of services in a front-office with its own facilities, processes, and a large extent of autonomy. The term \"front-office\" refers to a company's departments that come into contact with customers and typically includes the marketing, sales, and other customer facing staff. The \"back-office\" is the part of the business dedicated to running the company itself and typically includes people who deal with design, development, production, and other activities that are rarely seen by customers. Here, it is important to highlight that the front-office/back-office distinction should not be confused with the physical location of facilities.\n6 Relationships between people practices and the delivery of advanced services: The six behaviours and skill-sets summarised in Table II are considered by our case companies to be key to the successful delivery of advanced services. In this section we set out to present this rationale by explaining how the people practices (both organisation and skills) support the attainment of key performance measures for advanced services contracts. Typically, these are measures of asset performance, availability and reliability, which are needed to be delivered at the lowest cost. Figure 1 shows this general relationship as an influence diagram, showing how the practice in people organisation and skill-sets translates into successful delivery of the key performance measures for an advanced services contract.\n7 Concluding remarks and future research: This short paper has set out to present our preliminary findings about the organisation and skill-sets of people who are in the front-line of delivering advanced services. To achieve this we have introduced our research programme, summarising how we have found people to be organised, outlined their desired behaviour and skill-sets, and then why these practices are consistent with success. To conclude this preliminary report on our research, we have summarised our findings about people in the following hypothesis:\n",
"role": "user"
},
{
"content": "Purpose: This \"research note\" sets out to fuel the debate around the practices and technologies within operations that are critical to success with servitization. It presents a study of four companies which are delivering advanced services and reports on the organisation and skill-sets of people within these.\nDesign/methodology/approach: This has been case-based research at four manufacturers leading in their delivery of services.\nFindings: It describes the desirable behaviour of people in the front-line of service delivery, identifies the supporting skill-sets, how these people are organised, and explains why all these factors are so important.\nOriginality/value: This paper contributes to the understanding of the servitization process and, in particular, the implications to broader operations of the firm.\n",
"role": "assistant"
}
] |
Senior consumers' dining-out behaviors: the roles of physical, psychological and economic health
|
[
"Aging",
"Restaurants",
"Restaurant decision-making",
"Self-perceived health",
"Senior consumer",
"Restaurant behaviour",
"Physical health",
"Psychological health",
"Economic health",
"Perceived health"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The world's population is aging. Between now and 2030, the USA is expected to experience a considerable growth in its senior population as well. By 2030, the number of people of 60 years or more is projected to account for 25-29 per cent of the US population. (Engel, 2014). Longer life expectancies and aging baby boomers are responsible for seniors constituting a large part of the adult population and, therefore, consumer markets (Yoon et al., 2009). More recently, this phenomenon of an aging population has guided a good deal of service marketers. Consequently, researchers have identified and developed products and services that appeal to the senior segment (Yoon, 1997). Many business sectors have treated the senior market as a homogeneous segment of the population and stereotyped aging consumers as psychologically and physically limited, as well as financially poor (Chen and Shoemaker, 2014). In other words, because of such treatment, stereotypes senior consumers have been largely ignored in the marketplace. However, today's senior market is the fastest-growing age group with an increasingly high spending power (Cummings, 2006; Gonzalez et al., 2008). In particular, a growing demand for food services can open up opportunities for service marketers. A consumer study reported that senior consumers are making more visits to restaurants than prior to the recent 2008-2009 recession (The NPD group, 2013). Contrary to past information claiming that senior consumers eat out less frequently than those in younger age groups, the proportion of senior consumers' dining-out has grown steadily over the past five years, while the number of younger adults dining-out has declined (Jang et al., 2007).\nLiterature review and hypotheses: Age-related changes and dining-out behaviors\nMethodology: Sample\nResults: Effects of aging and physical health\nConclusions: Given that aging is a global phenomenon, senior consumers can be an attractive target market for the service sector because of their relative wealth, discretionary income and available free time. Therefore, understanding restaurant behaviors among senior consumers provides more business opportunities to service marketers. In particular, this study focused on how perceptions of three main age-related changes - physical, psychological and economic healthiness - may influence restaurant behaviors in different ways. This study suggested that perceptions of physical, psychological and economic health influence the abilities and resources that consumers bring to bear on restaurant consumption decisions.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the impact of three types of perceived healthiness (physical, psychological and economic health) on restaurant consumption activities among senior consumers.\nDesign/methodology/approach: Data were collected from 325 restaurant customers in the USA through a Web-based survey conducted by an online marketing research firm. Following previous studies' categorization of senior consumers, this study distinguished between \"senior consumers\" (age 60 or older) and \"younger consumers\" (age 39 or younger) to better compare different age-related behaviors. A series of two-way ANOVA was conducted for dining-out frequency and the number of choice sets.\nFindings: The results revealed that senior diners' perceptions of healthiness are critical in determining senior consumers' restaurant behaviors, such as information processing and purchasing behaviors. The results provide evidence that seniors with a positive perception of their physical and psychological status seek a greater number of alternative restaurant choices, which is actively related to purchasing frequency.\nPractical implications: The managerial implications indicate that restaurant marketers should avoid stereotypes and instead rely on more recent and accurate information regarding today's senior consumers.\nOriginality/value: The position taken in this study recognizes the need to enhance the understanding of senior consumers' patterns regarding their perceived physical, psychological and economic health. To the best of authors' knowledge, this is the first attempt to investigate the impact of three types of health on seniors' dining behaviors.\n",
"role": "assistant"
}
] |
On portfolio optimization: Forecasting asset covariances and variances based on multi-scale risk models
|
[
"Factor models",
"Minimum-variance portfolio selection",
"Wavelet decomposition",
"C10",
"C32",
"G11",
"G15"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The main aim of our paper is to demonstrate how investors can incorporate the multi-scale nature of asset and common risk factor returns into their portfolio decisions[1]. As it presents a difficult task to forecast mean returns, portfolio decisions are commonly based on risk optimization in practice, while mean-variance optimization is barely used. Portfolio risk optimization does usually only rely on the forecast of the variance-covariance matrix (Zakamulin, 2015). In our application, we do therefore follow Chan et al. (1999) and focus on portfolio risk optimization or, more specifically, minimum variance portfolio selection. To account for the multi-scale nature of asset and common risk factor returns, we use various (multi-scale) factor models to forecast the required variance-covariance matrix. The construction of multi-scale factor models is put into execution by the use of wavelet filter techniques.\n2. Methodology: 2.1 Wavelet analysis\n3. Data: The data comprise all stocks that are listed in the Dow Jones Industrial Average (DJIA). Due to its late listing, we excluded Visa from our analysis and focus on daily return series of 29 stocks from 2000 to 2015. As this sample covers both calm and turmoil market times, we are able to separate this period into different sub-samples to explicitly assess pre- and post-crisis periods.\n4. Empirical results: To evaluate the out-of-sample performance of portfolio decisions that account for the multi-scale nature of asset and factor returns, we assess the performance of minimum variance portfolios determined by the variance-covariance matrices that are derived by different factor models and different scales (as described in Section 2.2). The variance-covariance matrices are determined on a daily basis in a rolling window regression analysis. Along the lines of Patton and Timmermann (2010), we set the rolling window size to 250. The daily variance-covariance matrices of the different models and different scales at time t are used to forecast variances and covariances at time t + 1 and to determine the portfolio weights of the global minimum variance portfolio[2]. The respective portfolio weights for each level of decomposition (wi,t(t)) are then applied to the original return series to determine the out-of-sample portfolio return at time t + 1. Consequently, we assess nine different portfolio strategies: minimum variance portfolio weights based on the covariance structure of the undecomposed asset returns and the eight decompositions, respectively. Moreover, we apply four different factor models to estimate the variance-covariance matrices.\n5. Conclusion: Motivated by the work of Trimech et al. (2009), we demonstrate the use and evaluate the performance of portfolio strategies which take the multi-scale nature of asset and common risk factor returns into account. We find that optimized portfolios based on variance-covariance matrices stemming from multi-scale factor models outperform their counterparts based on undecomposed variance-covariance matrices and, thus, provide a promising alternative for portfolio management. We find that a multi-scale four-factor model as proposed by Carhart (1997) is sufficient for minimum variance portfolio selection. Moreover, as we show that the explanatory power of risk factor models increases if we consider time scales that describe the long-run trend, our findings confirm the results of Gencay et al. (2003, 2005) and Trimech et al. (2009).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to show how investors can incorporate the multi-scale nature of asset and factor returns into their portfolio decisions and to evaluate the out-of-sample performance of such strategies.\nDesign/methodology/approach: The authors decompose daily return series of common risk factors and of all stocks listed in the Dow Jones Industrial Index (DJI) from 2000 to 2015 into different time scales to separate short-term noise from long-run trends. Then, the authors apply various (multi-scale) factor models to determine variance-covariance matrices which are used for minimum variance portfolio selection. Finally, the portfolios are evaluated by their out-of-sample performance.\nFindings: The authors find that portfolios which are constructed on variance-covariance matrices stemming from multi-scale factor models outperform portfolio allocations which do not take the multi-scale nature of asset and factor returns into account.\nPractical implications: The results of this paper provide evidence that accounting for the multi-scale nature of return distributions in portfolio decisions might be a promising approach from a portfolio performance perspective.\nOriginality/value: The authors demonstrate how investors can incorporate the multi-scale nature of returns into their portfolio decisions by applying wavelet filter techniques.\n",
"role": "assistant"
}
] |
Service quality framework for clinical laboratories
|
[
"Mauritius",
"Service quality dimensions",
"Clinical laboratory",
"LabSERV"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Health service organisation staff play a crucial role in economic growth especially in developing countries (Zarei et al., 2012). In Mauritius, Government health services are free and accessible to everyone (Kassean and Juwaheer, 2010). Over the years, the health sector has witnessed fast growth and, through a Governmental policy, the country has embarked on an ambitious project of becoming a leading medical hub for the African region (Devi, 2008). In this context, much emphasis is laid on the quality of diagnosis and treatment for patients attending public health premises. The medical laboratory services (MLS), a department in the public health sector, plays a vital role in the screening, diagnosis and monitoring diseases to assist clinicians in providing a healthcare service of high standard to patients. The MLS is a medical laboratory network located in regional, specialised and district hospitals. As the public health service is free, MLS staff face much pressure from its customers for service quality and timeliness (Ramseook-Munhurrun et al., 2010). This prompts senior managers in the health sector to identify indicators for assessing clinical laboratory quality levels (Chawla et al., 2010). Moreover, healthcare system MLS and medical practitioners are being required to operate in-tandem for synchronised actions leading to more outcome-oriented, patient-focused and evidence-based approaches to service quality (Shaw and Mass, 2012). The major concern is how well the MLS is performing from the medical practitioners' point of view (Allen and Harris, 1992). Staff in the MLS also realise that the services that they offer must be in line with the needs and expectations of their customers (Wisniewski, 2001). This provided the basis for the current research which aims at developing a service quality framework specific to clinical laboratories.\nMethodology: This study is a survey based research whereby a modified SERVQUAL questionnaire has been designed and administered to medical practitioners, the main customers of the MLS. In public hospitals, it is only the medical practitioners who are allowed to make official requests for analyses to the laboratory. The questionnaire is primarily based on the work of Parasuraman et al. (1985). The questions were, however, adapted to enable relevant data capture, which were eventually compiled, verified and statistically analysed in line with established research questions and hypotheses.\nHypotheses: The following hypotheses were brought forward and tested to elaborate the service quality conceptual framework within clinical laboratories:\nDiscussion: The eight SQDs addressed in the framework have been adapted to suit clinical laboratories. They interact to enhance service quality, which impacts on customer satisfaction. The framework also allows flexibility about the service quality attributes that can be changed, modified or re-structured upon changing circumstances. Also, any improvement in the gap score of any SQD will lead to an improvement in the overall service quality. The framework is moreover useful for implementing change for achieving continual improvement within the ISO15189:2012 standard. The overall service quality gap score generated through LabSERV for instance provides an insight of customer needs and satisfaction levels. Gap determination across the different service quality dimensions is vital for identifying shortcomings in service delivery that can be prioritised, quantified and addressed. Data from LabSERV can also be broken down into its specific service quality dimensions for certain targeted corrective and preventive actions. The findings and analysis allow users to prioritise SQDs where there is room for improvement. For example in this study, improvement in turnaround time requires further investigations about the time intervals such as from request for analysis to sample collection; the time necessary to get the sample to the laboratory; the laboratory pre-analytical procedures; time for sample analysis, results validation and reporting; and communicating these results to the medical practitioners (Stotler and Kratz, 2012). So, all these intermediates can be helpful in designing appropriate service quality attributes for the SQD turnaround time in the specific context of clinical laboratories. Turnaround time is considered among the most noticeable laboratory service sign and is used by many clinicians to judge the laboratory service (Hawkins, 2007). Additionally, the intermediates involved in turnaround time can be individually monitored to spot out where the main problem lies and tailor-made solutions can be applied to address particular problems. In doing so, the time taken by each of the intermediate stages can be improved to ultimately impact positively on the laboratory turnaround time.\nConclusion: In evidence-based medicine, medical practitioners rely to a much greater extent on clinical laboratories for timely decisions. It is thus imperative for the laboratory staff to collect information about customer needs, expectations and perceptions to provide the required service at expected level. The laboratories must be properly equipped to be able to gauge customer satisfaction level such that appropriate measures can be taken to improve the service. It is at this stage that the service quality framework is beneficial. By using LabSERV, the different gap scores associated to each SQD can be determined and in-depth analysis can be undertaken based on captured data. After implementing the required corrective actions, a situational assessment can subsequently be performed by re-running the LabSERV to verify whether set targets have been achieved. Furthermore, the service quality framework opens further avenues for clinical laboratories to participate in meeting the challenges of the healthcare sector and continuously improving the service offered.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to illustrate a service quality framework that enhances service delivery in clinical laboratories by gauging medical practitioner satisfaction and by providing avenues for continuous improvement.\nDesign/methodology/approach: The case study method has been used for conducting the exploratory study, with focus on the Mauritian public clinical laboratory. A structured questionnaire based on the SERVQUAL service quality model was used for data collection, analysis and for the development of the service quality framework.\nFindings: The study confirms the pertinence of the following service quality dimensions within the context of clinical laboratories: tangibility, reliability, responsiveness, turnaround time, technology, test reports, communication and laboratory staff attitude and behaviour.\nPractical implications: The service quality framework developed, termed LabSERV, is vital for clinical laboratories in the search for improving service delivery to medical practitioners.\nOriginality/value: This is a pioneering work carried out in the clinical laboratory sector in Mauritius. Medical practitioner expectations and perceptions have been simultaneously considered to generate a novel service quality framework for clinical laboratories.\n",
"role": "assistant"
}
] |
Are you a protean talent? The influence of protean career attitude, learning-goal orientation and perceived internal and external employability
|
[
"Learning-goal orientation",
"Perceived external employability",
"Perceived internal employability",
"Protean career attitude",
"Protean talent"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Due to an increasingly volatile economy, today's organizations face numerous challenges such as organizational mergers and acquisitions, downsizing, restructuring, flattening, outsourcing, and layoffs, and have been forced to reduce the promise of long-term employment relationships (Bremner et al., 1991; Greenhaus et al., 2008; Arthur and Rousseau, 1996; Sullivan, 1999). Organizations are now looking for new managerial concepts and ideas that will enable the use of talent management as a strategic tool for attracting protean talents and retaining valuable ones who can contribute to the competitive advantage of the organization (Silzer and Dowell, 2010; Lepak and Snell, 1999).\nLiterature reviews and hypotheses development: Connecting protean career attitude with its dimensions of value-driven career orientation and self-directed career management to perceived internal and external employability\nMethods: Sample and procedure\nResults: Confirmatory factor analyses (CFA)\nDiscussion: Two of the hypotheses results represent the most important conclusions of this study. The first interesting result revealed that protean career attitude (value-driven career orientation and self-directed career management) was a significant predictor of perceived internal and external employability. It empirically showed that protean career attitude physically influences protean talents to move across organizational boundaries to follow their personal value-driven career orientation and self-directed career management, which is consistent with other studies such as Hall (1996) and Hall and Moss (1998). Many previous studies also support this result that protean talents are independent agents who seek opportunities and manage their own careers beyond the organizational boundaries rather than inside of the organization (Forrier et al., 2009; Direnzo and Greenhaus, 2011; Murphy and Ensher, 2001). This increases the importance of career development as it is gained within the context of a boundaryless career.\nLimitations and directions for future research: This study had several limitations. Overall, the limitations of this study are consistent with the limitations of prior empirical studies conducted in this field. Because of the study limitations, any generalizations are subject to a number of caveats.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the relationship between protean career attitude and perceived internal and external employability, along with the mediating effect of learning-goal orientation.\nDesign/methodology/approach: Data were collected by distributing paper-based questionnaires to 527 workers in private banking sectors in Taiwan. Hierarchical regression analysis was used to examine the results of the relationships.\nFindings: The results supported the idea that protean career attitude is a significant antecedent of perceived internal and external employability. Protean talents with a higher degree of protean attitude toward value-driven career orientation and self-directed career management have an external employability that is greater than their internal employability. Learning-goal orientation fully mediated the effect of protean career attitude on perceived internal employability, but only partially mediated external employability.\nPractical implications: The findings can help human resource managers gain a better understanding of the use of an appropriate strategy to influence an employee's perceived internal and external employability, which can increase the motivation and improve employer-employee relationships that contribute to organizational success and performance. Employees should recognize the increased importance of continuous learning with goal-setting in order to deal with changes at work.\nOriginality/value: This paper empirically establishes the association between protean career attitude and perceived internal and external employability. The protean career concept may provide organizations with a valuable perspective in the evolution of careers. Valuable and protean talents place an emphasis on individuals' core values, and while learning goals are meant to suit employer organizations, they may also establish opportunities that could cross-organizational boundaries.\n",
"role": "assistant"
}
] |
Synthesizing technical skill building framework for extended enterprises in emerging economies
|
[
"Theoretical framework",
"Collaborative enterprise",
"Extended organization",
"Technical skill building"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The current globalized era characterized by dimensions such as cut-throat competition among enterprises, continuously improving technologies, dynamic consumer needs and increasingly shorter product life cycles poses significant challenges to the enterprises operating in the emerging countries. These challenges are further amplified by the fact that firms operating out of emerging countries are at time handicapped by their own organizational inefficiencies. Some of these inefficiencies pertain to: a silo-based work culture; deficit of desired level of collaboration among internal agencies; and lack of trust in the value chain comprising of vendors, suppliers and so forth.\n2. Methodology: To examine some of the key ingredients and building blocks of products, processes and people, we rely on the review of the extant research literature. Search for relevant papers was conducted employing structured keyword search in Google scholar and major databases such as Emerald, Elsevier, JSTOR, Informs, Sage, Springer, Taylor & Francis and Wiley. Further, primarily the articles published in last ten years, i.e. 2007 onwards, have been included as reference for our study. However, some of the seminal work (such as that of Shewhart, 1931) belonging to pre-2000 era has also been included in this study. Since we seek to evolve a pragmatic and effective technical skill building framework in context of a manufacturing firm in an emerging economy, it is imperative to examine pertinent technical, managerial, human, organizational and economy-related dimensions. These dimensions are essentially the subject matter-related criterions based on which relevant articles have been examined. Further, we also relied upon the research papers that were relatively qualitative and conceptual in nature as opposed to being purely quantitative. As far as technical and managerial dimensions are considered, we rely on technical dimensions-oriented journals such as International Journal of Production Research, Total Quality Management and Business Excellence and International Journal of Project Management. On the other side, as far human, economy, and organization-related aspects are considered, we review the literature from the journals as International Journal of Human Resource Management, Entrepreneurship Theory Management and Practice and Leadership and Organizational Development.\n3. Literature review: Bell and Figueireod (2012) reviewed the research literature on learning processes as a source of innovation in context of firms belonging to developing/emerging countries in last 25 years. One of the key findings of this work was that these firms (also termed as latecomers in this research) tend to remain locked into forms of \"follower innovation\" rather than \"leader innovation.\" A key implication of this work was in terms of the proposition that the organization as a collective entity need to collaborate horizontally if the process of innovation is to be accelerated. Das et al. investigated the synergies in advanced manufacturing systems and their impact on manufacturing performance. This work integrated the socio-technical theory with subsystem-level manufacturing design. A noteworthy contribution of this paper was that it promulgated high-involvement work practices as a key component of manufacturing systems competitiveness. Allwood et al. advocated job rotation as a key enabler for enhancing the workforce's competitiveness and organization's collective problem-solving skills. In this research, the learning curve is modified to describe problem solution times, incorporating forgetting effects and treating both general skills and specific skills to solve the particular problem. Ates and Bititici (2011) demonstrated that change management process capability is fundamental to creating resilience in small and medium enterprises (SMEs). This research further illustrated that sustainability and resilience in SMEs will be enhanced by ability to embrace organizational and people dimensions as well as operational aspects of change management, and paying attention to long-term planning and external communication to drive change proactively. Shankar and Hanson (2013) provided a conceptual framework comprising the drivers and consequences of innovation architecture across emerging and developed markets. This study also highlighted emerging market innovation characteristics using detailed examples. Further, this research inferred that most of the future growth in the global economy will come from emerging markets. Successful global firms will have to rethink and develop their innovation architecture by leveraging innovations developed for emerging markets. A key proposition of this study was that technical competency would be key to furthering the organizational mission and vision. Wu (2015) formulated a model in which a firm simultaneously chooses its organizational structure and product position. The firm's production is knowledge intensive, requiring employees to solve problems. A vertical hierarchy, in which workers refer unsolved problems to managers, facilitates the acquisition and leveraging of managers' superior knowledge. Mehrzi and Singh (2016) examined research conducted in the area of employee engagement and proposed a conceptual framework that can be used by practitioners to engage employees and motivate them toward organizational growth and sustainability. In this work, framework for employee engagement was developed, illustrating the linkage between leader, team, perceived organizational support and organizational culture that is being mediated by employee. Harris et al. (2014), in their research focused on trainer style and learner orientation, found that directiveness and trainee learning goal orientations each additively predicted training satisfaction and transfer over and above one another and study controls. Chen et al. (2005) researched the moderating effect of work environment and managerial skill utilization. Specifically, it focused on five related dimensions i.e. technical skills, human skills, conceptual skills, environment for managerial development and training performance-based promotion policies. A relevant contribution of this work pertained to evolving deeper understanding of the dynamics of management training and well as the insights into the design of managerial training programs and skill-utilization assessments. McSherry and Taylor (2006), based on their empirical study, inferred that supervisors' use of skills and terminology from the training program; creating opportunities for trainees to make decisions based on newly learned skills; reinforcing trainees' use of trained skills; creating opportunities to practice new skills; and providing feedback on skill use. This study demonstrated the effectiveness of technical skill building within organizations in context of supervisory support.\n4. Establishing the SIPOC framework: In context of an auto OEM as considered in our research, there are a number of key project areas that need to be addressed from a technical skill building standpoint. In Table I, we enlist the areas as well as establish their applicability with respect to both internal and external stakeholders.\n5. Evolving DMEDI-based technical skill building process map: The DMEDI framework is commonly used as a problem-solving model in wide variety of business-related problems. According to this framework, any effective framework can be formulated employing the steps: define (D), measure (M), explore (E), develop (D) and implement (I). As opposed to improvements in existing frameworks DMEDI's emphasis is on design and formulation of new frameworks (Chakrabarty et al., 2007) (Figure 3).\n6. Discussions: In our research, illustrating a case example of an automotive company's value chain, we first attempt to map the suitability of respective project area with respect to both internal organizational stakeholders and external stakeholders. This mapping provides us the foundation for evolving the SIPOC-based structure and DMEDI-driven technical skill building map.\n7. Conclusions and future research: In this research, we aim to evolve extended enterprise-based (modeled along the features of collaborative enterprise) structure and process map for technical skill building within firms belonging to emerging economies. Initially, by establishing a high-level theoretical framework containing stakeholders belonging to the extended enterprise, i.e. manufacturer, customer, dealers and supplier and key dimensions related to product, process and people, we institute technical skill building as a source of competitive advantage for the growing enterprises. Thereafter, employing the SIPOC-based framework, we identify the suppliers, inputs, processes, outputs and customers in the value chain of technical skill building as far as the extended enterprise is concerned. An explicit benefit of SIPOC framework in our underlying research is that it establishes tangible elements, flows, requirements and so forth such that interactions of individual dimensions can be visualized. Finally, by leveraging the features of the DMEDI approach, we model the skill building process map in terms of a continually improvable user-friendly framework.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to formulate extended enterprise-oriented structure and process map for technical skill building for firms belonging to emerging economies.\nDesign/methodology/approach: The authors identify technical skill building (lies at the intersection of product, process and people) as a key enabler for furthering organizational competitiveness. The authors employ define measure explore develop implement- and supplier, input, process, output, and customer (SIPOC-) based framework for arriving at a technical skill building framework.\nFindings: By utilizing the SIPOC-based framework, the authors identify the supplier(s), input(s), process(s), output(s) and customer(s) in the value chain of technical skill building as far as the extended enterprise is concerned.\nOriginality/value: The paper uses pragmatic and user-friendly conceptualization for technical skill building framework.\n",
"role": "assistant"
}
] |
"My bad": investigating service failure effects in self-service and full-service settings
|
[
"Service satisfaction",
"Service failure",
"Internal attribution",
"Self-services",
"Service reusage intention"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Imagine that two rail travelers, Mr Meedo and Mr Yoodo, have missed their connecting train, and thus were delayed 2 h in reaching their final destination. Further assume that Mr Meedo used the railway company's website to schedule and book the journey on his own, and he did not allow the necessary time to make the connection. In contrast, Mr Yoodo went to the company's service counter, where a frontline employee performed the scheduling and booking task. Although both travelers found themselves in an equivalent situation - both missed their connection and arrived later than desired - they may react differently to the service failure. For instance, would Mr Yoodo be more upset about the delay than Mr Meedo? In either situation, who would be held responsible for the faulty booking process? Would either traveler use the same service delivery mode to book the next journey?\nRelevant background and hypotheses development: Consumers' attributions of service failures\nStudy 1. Consumer response to service failures in self-service and full-service settings: In Study 1, we test our assumption that, although the negative effect of service failures on service satisfaction is similar in full-service and self-service contexts, the impact on service reusage intentions diverges (H1). In addition, we explore how the causal relationships underlying this phenomenon differ across both service delivery modes.\nStudy 2. The role of internal attributions of service outcomes: Building on the results of Study 1, which shows that service failures entail unfavorable consequences to a different extent between the two service delivery modes, the purpose of Study 2 is twofold. First, Study 2 aims to replicate the findings of Study 1 in a different context. The second objective is to test our assumption that different degrees of internal attributions of service outcomes associated with full-services and self-services, respectively, can be held responsible for this phenomenon (H2).\nConclusions: Owing to the ubiquity of self-services as an additional service delivery option in various industries (e.g. financial services, travel and retailing), customer roles in service encounters have changed radically over the past years. Our research offers an initial examination of the diverging effects of service failures as compared between self-service and conventional full-service contexts. The findings of our investigation suggest that customers who are actively involved in the service delivery process react less sensitively to unsatisfying service outcomes than do customers who experience unsatisfying full-service outcomes. More precisely, we demonstrate that although service failures have an equivalent negative effect on service satisfaction across full-services and self-services, the intention to use the same service delivery mode again is significantly higher in the self-service context than in the full-service context. Whereas consumer response in terms of future service usage intentions strongly depends on service satisfaction in the full-service context, the service satisfaction-repurchase intentions link in the self-service context is diminished, such that reusage intentions remain high even if service failures occur. These effects can occur in both technological service settings (Study 1) and in non-technological service settings (Study 2). In addition, our studies confirm that this phenomenon is a function of different levels of internal attributions of service outcomes associated with different types of service delivery that moderate the relationship between service failure and service satisfaction, as well as between service satisfaction and reusage intention.\n",
"role": "user"
},
{
"content": "Purpose: Companies in diverse branches offer a variety of service alternatives that typically differ in terms of the degree to which customers are actively involved in service delivery processes. The purpose of this paper is to explore potential differences in consumers' reactions to service failures across services provided by a service employee (i.e. full-services) and services that require customers' active involvement (i.e. self-services).\nDesign/methodology/approach: Two 2 (full-service vs self-service) x 2 (no service failure vs service failure) scenario-based experiments in technological and non-technological contexts (i.e. ticket purchase and furniture assembly) were conducted.\nFindings: Study results reveal that although service failures have a similar negative impact on satisfaction across both full-services and self-services, in the self-service context, the negative effect on the willingness to use the same service delivery mode again is attenuated.\nResearch limitations/implications: By emphasizing the role of customers' active involvement in the service delivery process, the study extends previous knowledge regarding customer response to service failures in different service settings.\nPractical implications: By highlighting that self-service customers' future behavioral intentions are less severely affected by service failures, the authors present an additional feature of customer involvement in service delivery processes that goes beyond the previously recognized advantages.\nOriginality/value: Despite the abundance of research on the effects of failure attributions, previous studies have predominantly examined main effects of attributions on customer responses, such that insights into potential moderating effects of failure attributions on established relationships - as investigated in this study - are still scarce.\n",
"role": "assistant"
}
] |
Key success factors of strategic sourcing: An empirical study of the Hong Kong toy industry
|
[
"Sourcing",
"Leadership",
"Supplier relations",
"Continuous improvement",
"Toys",
"Hong Kong"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Sourcing is a complex process that contributes tremendously to the competitive advantage of an organisation (Novack and Simco, 1991). AS much as 70 per cent of an organisation's sales revenues or total manufacturing costs is spent on purchasing raw materials, components, finished goods or services (Presutti, 2003; Lo and Yueng, 2004; Tayles and Drury, 2001). If sourcing costs can be reduced, this can improve returns on investment by increasing both profit margins and asset turnover rate (Dobler and Burt, 1996; Leenders and Fearon, 1997). During the past two decades, the purchasing function has changed from playing a supporting role to becoming a strategic activity, and now makes a significant contribution to the competitive advantage of an organisation (Quayle, 2002; Carr and Smeltzer, 1997). Strategic sourcing - also labelled purchasing, procurement or outsourcing by Banfield (1999), Greaver (1999), Carr and Smeltzer (1997) and Talluri and Narasimhan (2004) - is critical to the success of a company through its supply chain by using strategic tools to manage suppliers and achieve the company's long-term goals. Strategic sourcing is a systematic and comprehensive process that adds value to a company, which in turn helps to achieve the company's long-term objectives (Smeltzer et al., 2003; Anderson and Katz, 1998). Carr and Smeltzer (1997) specifically point out that strategic sourcing consists of the strategic processes of planning, evaluating, implementing and controlling all sourcing activities by a company to achieve its long-term goals. Smeltzer et al. (2003) find that all strategic sourcing strategies emphasise the integration of business practices such as early supplier involvement, supplier development, supplier assessment, supplier certification and measurement.Hong Kong toy companies are famous for producing a full range of quality toy products. According to the Hong Kong Trade Development Council (2007a, b), the export values of Hong Kong toy products were US$9.616 billion in 2003, US$9.849 billion in 2004, US$10,198 billion in 2005 and US$9.944 in 2006. The differences in export values from the past four years were minor; one of the reasons is that Hong Kong toy manufacturers are facing the rapid growth and development of indigenous local Chinese enterprises in mainland China and other developing countries, such as Thailand and Vietnam, which put a great deal of pressure on the Hong Kong toy companies to compete with them at low cost. Surveys conducted by the Hong Kong Trade Development Council (2000, 2003) indicate that Hong Kong as a sourcing centre for Chinese-made products is under challenge from Shanghai and Guangzhou in terms of cost competition. This implies that Hong Kong companies have to tighten their sourcing performance to deliver goods to customers in a cost-effective manner. Quelin and Duhamel (2003) indicate that the most important criteria of sourcing are to lower operational costs, focus on core activities, reduce capital investment, gain assess to external competencies and improve quality. As many Hong Kong toy manufacturing activities are sourced to second tier suppliers, if they can find good and reliable suppliers in mainland China they will be able to control the new product development schedule, production planning and shipment schedule, thus maintaining an acceptable quality standard within an acceptable budget. This, in turn, will result in improved competitive advantage. Based on the aforementioned background information and with reference to the relevant literature, it is worthwhile investigating strategic sourcing management issues in the Hong Kong toy industry and identifying their key success factors (KSFs). The findings will provide insights into the assessment and improvement of the industry's strategic sourcing performance.The research conducted by the authors had two aims. The first aim was to investigate the KSFs of strategic sourcing so that companies can then focus on these KSFs to improve their sourcing performance. The second aim was to prioritise the relative importance of these KSFs so that companies can evaluate their current practices and re-allocate reasonable resources to improve their sourcing performance.The remainder of this paper is organised as follows. First, KSFs are identified and discussed. Second, those KSFs are validated in the Hong Kong toy industry via the results of a questionnaire survey. Third, the relative importance of the KSFs is determined via the analytical hierarchy process (AHP) approach. Fourth, the practical implications of the research findings are discussed with reference to the current state of the Hong Kong toy industry.\nIdentification of strategic sourcing categories (SSCs) and key success factors (KSFs) in strategic sourcing: A full review and evaluation of the strategic elements in the literature was conducted; 14 KSFs in strategic sourcing were identified. After reviewing the nature of the 14 factors, they were grouped into three SSCs: \"Visionary leadership in strategic sourcing\" \"Supplier management system\" and \"Continuous improvement\". Above all, SSCs and KSFs must be supported by senior management's commitment to and involvement in strategic sourcing. Table I shows the relationships between the SSCs and KSFs from the literature.Visionary leadership in strategic sourcing\nExamination of the SSCs and KSFs in strategic sourcing: Questionnaire survey\nPrioritisation of relative importance of SSCs and KSFs: The findings provide good evidence that the Hong Kong toy industry can improve its sourcing performance by implementing the three SSCs and 14 KSFs. However, allocating company resources to all SSCs and KSFs at one time is not feasible due to limited resources, and it also requires detailed implementation planning and time to achieve a positive result. As the product life-cycle of toys is very short, usually less than one year, all sourcing decisions and functions need to be made very quickly, and this includes making changes without interrupting the existing system. To effectively and efficiently implement or improve strategic sourcing management and obtain immediate benefits, it is practical to implement the most important factors which, in turn, provide positive results to allow companies to further evaluate the feasibility of implementation of other KSFs by allocating more resources to them in the future. To identify the priorities of these three factor categories and 14 KSFs, experts in the Hong Kong toy industry were interviewed. The AHP approach was used in this study.The AHP is designed to decompose a complex, multi-criteria problem into multiple levels of hierarchy with the top level as the goal or objective, while the intermediate levels are the criteria and sub-criteria and the lowest level offers alternatives. Experts are interviewed and pairwise comparison judgments are applied to pairs of homogeneous criteria, eventually generating the overall priorities for ranking the alternatives (Saaty, 1980). The AHP is extensively applied in different areas with different applications. Vaidya and Kumar (2006) review 150 articles on the AHP that cover ten different applications within nine different areas. Priority and ranking is one of the applications widely used in different areas, such as manufacturing and engineering (Chin et al., 2002, 2006), and it is applied in this study. As the AHP uses experts to assign pairwise comparison judgments it is a subjective methodology and a large number of expert evaluators is not necessary (Cheng and Li, 2001). Many researchers have used a small number of expert evaluators for their research studies. Cheng and Li (2001) invited four, Tam and Tummala (2001) invited five, Muralidharan and Anantharaman (2001) invited ten, Chin et al. (2002) invited 13 and Lam and Chin (2005) invited 14. In this study, 20 toy industry experts were invited as evaluators to assign pairwise comparison judgments to determine the relative weights with respect to SSCs and KSFs shown in the AHP model (Figure 1).All 20 experts had been working in the Hong Kong toy industry and involved in sourcing activities, particularly at strategic levels, for between ten and 32 years (the profile of the evaluators is given in Table IX). Their positions range from Managing Director to managers in engineering, operations and purchasing. They had sufficient knowledge and understanding of the general souring practices in the Hong Kong toy industry. This implies that this group of evaluators well represented the viewpoints and opinions of the Hong Kong toy industry, and was thus able to evaluate the criteria and sub-criteria and assign relative importance to the criteria and sub-criteria in the AHP model.By using \"Expert Choice\" software, the pairwise comparison judgment matrices were translated into the corresponding largest eigenvalue problem and solved to determine the normalised and unique priority weights for each SSC and KSF in the AHP model. The priority weights were divided into \"local weight\" the priority weight with respect to the preceding hierarchical level, and \"global weight\" the priority weight with respect to the highest hierarchical level - the goal.Results and analysis\nDiscussion: Through the literature review, three SSCs and 14 KSFs were identified, and then a questionnaire survey was conducted to validate the factors. The survey results indicate that the three SSCs and 14 KSFs are important. In addition, there is a positive relationship between the extent of the implementation of the 14 KSFs and sourcing performance. Unfortunately, the implementation of the three SSCs and 14 KSFs lags behind their importance, and the Hong Kong toy industry should address this problem. We further identify the priority rankings of the three SSCs and 14 KSFs lags for the Hong Kong toy industry. Understanding their relative importance, Hong Kong toy companies can allocate resources and efforts to these factors to improve their Chinese-made products in quality, price and delivery which, in turn, will satisfy the requirements of overseas buyers. Detailed recommendations for the Hong Kong toy industry are given as follows.Three strategic sourcing categories (SSCs)\nManagerial implications: The identified KSFs are grouped into three SSCs; a review of this research provides a number of valuable findings and implications for strategic sourcing for individual practitioners.First, visionary leadership in strategic sourcing requires senior management to set and communicate clear sourcing objectives and performance expectations to all levels of relevant employees. Encouraging cultural change to motivate and reward employees in organisations facilitates collaboration across intra-organisational boundaries, thus helping to improve sourcing performance. It is necessary for senior management to include sourcing as part of corporate strategy and to provide adequate resources to support sourcing functions, thus encouraging the development of long-term buyer and supplier relationships. Further, it is necessary to utilise departmental expertise to form cross-functional teams to conduct competitive analysis. By interpreting the analysed information, action plans can be developed to enhance the strengths and weaknesses of sourcing.Second, the management of suppliers is crucial. Suppliers should be required to pass evaluation based on multiple selection criteria, with the supplier with the highest score selected for a long-term relationship. However, it is also necessary to continue to monitor supplier performance using a performance assessment programme; suppliers can use the assessment results to take corrective actions for improvement when necessary. Moreover, poorly performing suppliers can be pruned from the supply base. Suppliers with average performance can be developed by the transfer of best practices among suppliers through education and training. As a result, companies can improve their competitive advantage through improving supplier capabilities. Collaboration with suppliers by providing more information, thus enabling early supplier involvement in new product development, can improve development lead-time and cost, thus can improving sourcing performance.Finally, CI in strategic sourcing is also crucial. Management should set CI tasks in sourcing, allocating resources, providing an organisation-wide culture, training employees and motivating them by reward and recognition. Systematic approaches should be used to solve problems and improve the sourcing system. CI tools and techniques should be used to improve sourcing performance by eliminating the causes of problems. Companies should be developed into learning organisations by reinforcing the sharing of tacit knowledge through interpersonal communications, by transferring existing and new knowledge within the company and by integrating this knowledge into processes, products or services through lessons learned from previous failures.\nConclusion: Strategic sourcing has been increasingly recognised as an integral part of business strategies and practices. As companies cannot allocate equal amounts of effort or resources to each individual area of the strategic sourcing process due to limitations of human and monetary resources, the findings presented in this paper are very important in providing insights to manufacturers to better deal with strategic sourcing management decisions. First, the 14 identified KSFs examined in the literature review and questionnaire survey and the results of the empirical findings indicate that sourcing performance is positively related to the implementation of the 14 KSFs. This indicates that manufacturers could improve their sourcing performance by focusing on these key factors.Second, the priority of the KSFs was ranked by experts from the Hong Kong toy industry. These people understand the business well and have a strategic view of sourcing. Hong Kong toy companies should focus on the factors that that these experts gave the highest priority in the initial stage of implementing strategic sourcing. After the implementation of the selected factors, companies can review and evaluate the results to consider whether it is feasible to apply the remaining factors.The KSFs identified cover a wide range of business processes and activities from leadership and supplier management to CI. The implementation of strategic sourcing requires that the entire company become involved: it is not only the responsibility of senior management, but also requires the involvement of all employees. Management's role should be to establish an environment to facilitate the processes and support of employees in implementing it. In general, employees work under guidelines and directions from management. A good plan or system can help employees to improve their efficiency and performance. The implementation of a new strategic sourcing management system requires many changes that people may be reluctant about. To effectively implement strategic sourcing, companies should focus on the most important factors first and address the less important factors at a later stage. By doing so they would gradually reduce employee resistance to change, especially if positive results were generated.\nLimitations and future research: Although this research makes significant contributions to the literature on strategic sourcing, it has several limitations. First, the research was conducted through a questionnaire survey by mail and through interviews with experts from the Hong Kong toy industry. Although the results are positive and promising, future research into strategic sourcing should be conducted in the Hong Kong manufacturing industry to ascertain whether there are differences between industries.Second, the study focuses on toy companies in Hong Kong; future research should be conducted in cities in mainland China, such as Shanghai and Guangzhou, to identify any differences in strategic sourcing, which could provide the Hong Kong industry with ideas for improvement.Finally, the identification and prioritisation of SSCs and KSFs is only the first step in providing the Hong Kong toy industry with an insight into how strategic sourcing can be improved. Future research must develop a self-assessment system so the industry can evaluate its current performance. This will help to identify strengths and weaknesses, and will provide information to develop appropriate strategies for making improvements.\n",
"role": "user"
},
{
"content": "Purpose: - This research aims to identify and examine the importance of the key success factors (KSFs) of strategic sourcing, and investigate the relationship between KSFs and sourcing performance in the Hong Kong toy industry. The KSFs are ranked, and the results show those factors that should be given the highest priority in implementing strategic sourcing.\nDesign/methodology/approach: - A literature review, a questionnaire survey, interviews with industry experts and the analytical hierarchy process (AHP) approach were employed in the study. First, the KSFs were identified after a full review and evaluation of the strategic sourcing literature. A questionnaire survey of 205 companies was then conducted in the Hong Kong toy industry to verify the importance of these KSFs. Finally, 20 sourcing experts from the toy industry were interviewed and assigned pairwise comparison judgements on a proposed AHP model, which prioritised the relative importance of these KSFs for the Hong Kong toy industry.\nFindings: - About 14 KSFs are identified through the literature review. The survey results indicate that these 14 KSFs are important and they are positively related to sourcing performance. Moreover, the results show that there is only marginal implementation of these 14 KSFs amongst Hong Kong toy companies. With the prioritisation of the identified KSFs by the AHP approach, it is strongly suggested that the toy manufacturers assess and improve their strategic sourcing performance.\nOriginality/value: - The paper identifies the relative importance of the 14 KSFs, which will allow manufacturers to use their limited resources to focus on the most important factors in improving their strategic sourcing practice and performance.\n",
"role": "assistant"
}
] |
Investment in marketing: the allocation conundrum
|
[
"Sales budgets",
"Investments"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Ever-rising marketing budgets are becoming an explosive issue. On advertising alone, companies spend fortunes: Nestle, $11 billion; Unilever, $8 billion; General Motors, $4.7 billion; Procter & Gamble, $3.8 billion; Sony, $3.4 billion; and Coca-Cola, $1.7 billion1. With no end in sight to escalating marketing outlays, many CEOs, CFOs and CMOs are asking two questions: \"Is our company spending the right amount on marketing?\" and \"Are we spending it in the right places and on the right activities?\"\nLimitations to how marketing budgets are traditionally set: Many companies determine their marketing budgets and allocations by benchmarking themselves against competitors. This can provide some useful insights if, in fact, their strategies are similar to those of competitors. Yet this is often not the case, even within an industry. Consider Target Corp. and Wal-Mart: Wal-Mart spends relatively little on marketing (0.3 percent of revenue) while Target spends 2.5 percent, eight times more as a percentage of sales2. Why? Target's business strategy is radically different from Wal-Mart's. While Wal-Mart is known for low prices across its merchandise lines, Target positions itself as more upscale and fashionable, with a number of exclusive-to-Target items. To do this, Target must advertise extensively to promote its brand, its unique products and their benefits. Wal-Mart's marketing is more institutional (its whole business is geared to everyday low prices), which enables it to reduce investments in advertising.A second way companies set marketing budgets is through ROMI and similar techniques. Marketers gauge the sales impact of various activities such as advertising, direct marketing and public relations and then let the profit impact of those activities drive their investment priorities. Programs meeting a hurdle rate of, say, a 15 percent return are permitted; those below the hurdle rate are cut or scaled back. The main advantage of such approaches is that they indicate which investments have low paybacks. By eliminating those with lower returns, executives can substantially boost their marketing ROI.However, ROMI approaches suffer from several shortcomings. One is a bias toward short-term results. The return on a campaign is typically measured in a short period of time after the campaign (usually less than a year). That doesn't account for the longer-term impact of marketing activities, such as brand building. Second, ROMI techniques tend to favor activities that improve profitability (return on sales) over those that drive revenue growth. Our analysis of many leading companies has shown that both are necessary for superior shareholder returns3. Third, new budgets are based on historical data. This data is irrelevant because if consumer responsiveness to the firm's marketing wanes, the marketing programs of the past will not have the same effectiveness in the future. Lastly, ROMI approaches don't factor in changes to business unit strategies. If the business changes tack - to emphasize, for example, premium products - marketing budgets set by ROMI will continue to allocate spending according to price-based promotions as before. In other words, the new marketing budgets and allocations will be clearly incompatible with the future direction of the business.But the biggest shortcoming of ROMI is time and effort. Collecting sales information and tying it directly to every marketing campaign in every product unit, geographic territory and division can take months, even years4. Frustrated by the sheer effort of ROMI approaches, many companies give up and revert to \"last year plus\" or competitor benchmarking.\nFocus on how to slice the pie, not how to size it: An alternative approach generates a much faster answer to the marketing budget dilemma. It's not a replacement for ROMI, but rather a much simpler tool that yields many of the same benefits. Within just a few months, it can provide the CEO, CFO and CMO with insights that greatly increase the return on marketing investment. The reason is that the approach focuses on the allocation issue - how much business units, products, brands or country units of a large company should spend on marketing in relation to each other.In many companies, marketing allocations are out of proportion, with some business units, product lines or geographic units consuming more than their fair share and others starved of funds to grow.Our experience based on working with many Global 500 multinationals is that reasonably sophisticated marketers are not likely to grossly over- or under-spend on marketing as a whole. Moreover, as long as spending is close to the optimum level, little extra value can be created from optimizing the level of marketing investment (see Figure 1). By contrast, decisions on where marketing investment goes can have quite a significant impact on value creation. So, focusing on the allocation question should be the highest priority.We have found five factors that should drive how companies apportion their marketing budgets (see Figure 2):1. The profitability that new revenue from a business unit or product would generate. For example, in a mature market with heavy competition, using marketing to increase revenue and gain market share can generate unprofitable or low-profit revenue. In such industries, there's little chance for marketing to create value - i.e. to yield profits over and above the company's cost of capital.2. The responsiveness of customers to marketing. In some markets, merely making the product more available to customers can be much more important than running ads and promotions to convince them to buy it. To sell soft drinks, for example, distribution is at least as important as marketing (after demand is generated).3. The revenue growth potential of a market. A company with a small share of a booming market has great growth potential in that market or \"headroom,\" as we refer to it. Conversely, a firm with a large share of a flat or receding market has little headroom. In general, companies with business units or products in markets with significant headroom must spend more on marketing than they would in markets with little headroom. A prime example is the market in China and other developing countries. In these regions, marketing must introduce a company's offerings to a great unwashed audience. That means reaching many with multiple messages.4. The marketing responses of competitors. Increases in marketing can often spark an arms race with no winners. Competitors raise their marketing spending and/or cut prices, which can reduce the sales impact of marketing and raise its cost. Cola, ice cream and cookie companies are well aware of the toll marketing wars can take on profits.5. The degree to which a particular type of investment builds longer-term growth in addition to near-term impact. For example, promotions can have an immediate and significant impact on revenues for commodity products like gasoline, toilet paper and milk, but little longer-term impact. For the same impact on near-term sales, we would clearly prefer investments that also had lasting impact - through brand building, for example.\nHow the process works: the case of a large consumer products business: To illustrate the process, we draw on our recent experience with the consumer business unit of a large, global company. After a series of discussions, top management decided to tackle the allocation question and put aside the total spending issue. Marketing and product managers in each business unit identified the three most important drivers of marketing effectiveness in the organizations: the potential for profitable new revenue (or value creation), headroom for growth and consumer responsiveness to marketing. After analyzing this data, the managers then determined how to allocate marketing dollars across the business unit's five brands (see Figure 3).Two brands (\"Brand 1\" and \"Brand 2\") had greater potential to generate profitable revenue over the next three years. The company forecasted its average margin for Brand 2 to be 14 percent and Brand 1's to be 13 percent. Brand 5, in contrast, was forecast to generate only a 6 percent margin over those three years.Furthermore, the company's managers found that Brand 5 had far less headroom for growth than the other four brands. It had 23 percent share of a market that was only forecast to grow 5 percent a year (CAGR) over the following five years. In contrast, Brand 4 had only 8 percent of its market and a 9 percent CAGR forecast. A further indictment against Brand 5's receiving a significant share of the marketing budget was that consumer responsiveness to its marketing was very low. Only 4 percent of consumers in its market chose a product on the basis of marketing.The biggest misalignment in marketing spending turned out to be between Brands 3 and 4. Margin and growth for both were similar, and the majority of their consumers selected their brands on the basis of marketing. Yet Brand 4, which had higher headroom for growth, was initially earmarked to receive less than half of Brand 3's marketing investment (as a percentage of sales).Following the portfolio review, managers reallocated spending among the brands to better leverage these insights. Management then used the same process to realign its marketing allocations across countries, product categories and channels. The impact: an additional $80 million in revenue, $35 million in profit and $130 million in intrinsic value - over and above what the previous plan had predicted. This happened despite the fact that the new allocations were based on a 15 percent reduction in the total marketing budget versus the prior budget.But the numbers alone didn't determine the new marketing allocations. An important influencing factor on these decisions was management's assessment of the capabilities the company had to manage additional investment in each market.For example, in one market, the modeling work suggested doubling the marketing budget, but management's marketing skills were relatively weak. Management recognized this and decided therefore to increase spending only marginally in the market and instead focus on building capabilities for the longer term. In this way, management judgment played a significant role in the final allocation decisions.\nBenefits: financial and beyond: The approach outlined here has four main benefits:1. It can help a company with multiple brands and products across multiple countries generate a greater and faster payback from marketing. By focusing on the factors that drive effective budget allocation decisions, the annual marketing budget \"tug-of-war\" can operate on a more objective, factual basis.2. It improves the quality of the often contentious discussions among the CEO, CFO, CMO, business unit heads and their marketing and product directors on who should get what resources. For example, it helps senior managers ask the right questions, such as, \"If this market is our fastest-growing, why aren't we investing more in marketing there?\" or \"If this market is our least profitable, why are we investing so much in marketing?\"3. It can be extended up and down the organization. For example, it can help a BU head to allocate marketing dollars across products or countries, or a country unit head to spread marketing resources across brands. As a result, managers at all levels improve their discussions and, ultimately, their decisions on marketing investments.4. It can identify the new marketing capabilities a company needs. For instance, a company that must dramatically increase marketing investments in a developing market will quickly find it is not practical to ramp up activities unless the BU has sufficient capabilities in marketing management and knowledge of the local market (the marketing channels, consumer responsiveness, competition, etc.).All in all, the approach we are proposing here helps big companies shift their marketing budget debate from \"how many dollars to spend altogether?\" to \"where to spend those dollars?\" While companies should continue their quest to answer the first question with rigor, until they get there, answering the second question can create substantial value.\n",
"role": "user"
},
{
"content": "Purpose: - Ever-rising marketing budgets are becoming an explosive issue. On advertising alone, companies spend fortunes: Nestle, $11 billion; Unilever, $8 billion; General Motors, $4.7 billion; Procter & Gamble, $3.8 billion; Sony, $3.4 billion; and Coca-Cola, $1.7 billion. With no end in sight to escalating marketing outlays, many executives are asking two questions: \"Is our company spending the right amount on marketing?\" and \"Are we spending it in the right places and on the right activities?\" To get more bang for your buck, the authors suggest setting aside the first question and focusing on the second.\nDesign/methodology/approach: - In this article, the authors explore the limitations of current approaches to setting marketing budgets and then outline five factors that should drive how companies allocate their marketing budgets, namely: the profitability that new revenue from a business unit or product would generate; the responsiveness of customers to marketing; the revenue growth potential of a market; the marketing responses of competitors; and the degree to which a particular type of investment builds longer-term growth in addition to near-term impact.\nFindings: - The new approach described in this article produces four main benefits: It provides a more objective, factual basis for determining annual marketing budgets; it improves the quality of discussions and decisions around marketing allocations; it can be applied at all levels throughout the organization; and it can identify the new marketing capabilities a company needs.\nOriginality/value: - The article offers a new, simpler approach to thinking about marketing allocation that helps big companies shift their marketing budget debate from \"how many dollars to spend altogether?\" to \"where to spend those dollars?\" While companies should continue their quest to answer the first question with rigor, until they get there, answering the second question can create substantial value.\n",
"role": "assistant"
}
] |
Feasibility study of digital document supply services in Iranian research institutes and academic libraries
|
[
"Digital services",
"Academic libraries",
"Document delivery",
"Research institutes",
"Academic libraries",
"Iran"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The decline in library budgets and purchasing power, rapid information growth, increases in client expectations for better and faster access to updated information, and the role of modern technologies in raising these expectations, have made resource sharing and interlibrary loan essential necessities of libraries and information centres. Access to information meets the clients' satisfaction if it is achieved in a limited time; therefore, the existence of a united centre for coordinating and expediting access to requested materials seems necessary. The internet and its related technologies have altered the nature of document supply (Zahedi and Joukar, 2007).Collaboration and networking among libraries and information centres facilitate access to more extensive collections, improve service provision, enhance resource sharing activities, reduce the need to provide redundant copies, and offer services at lower costs. Developed countries get maximum benefit from such services, while these services are less used in developing countries; the need to establish these systems is thus evident. The key components of establishing collaboration and networking among libraries and information centres in developing countries include:* promoting awareness about the potentials of these services;* developing communication infrastructures; and* maximising the use of technology.Furthermore, the reasons that justify the need for systems of collaboration and digital document supply include:* the growth of information and media;* financial limitations;* rapid growth of technology; and* the widespread and successful use of networks and collaboration systems in libraries in developed countries (Khalid, 2000).The networks of resource sharing and document supply services are currently active at local, regional, national, and international levels. In order to prepare their information resources, the universities in Isfahan City collaborate with most Iranian information centres, universities, and research institutes as well as with international institutes with document delivery services such as the Linda Hall Library (LHL) and the British Library (BL). Moreover, they use online information databases. Taking into consideration the technological changes and access to the internet, the existence of a digital environment for providing document delivery services to clients seems crucial. Besides investigating the existing conditions, we attempt in the present research to study the prerequisites for establishing such services and investigate the facilities offered by such a system for the clients in academic libraries and research institutes in Isfahan. This study also considers the hardware, software, network, human resource, and other facilities in universities and research institutes in Isfahan province, so as to suggest the necessary strategies and solutions to remove the barriers after recognizing them, with the aim of providing the conditions required for making use of these facilities in the digital era.\nGoals of the study: The goal of the research was to conduct a feasibility study on establishing digital document supply services in academic and research institute libraries in Isfahan province. The study's objectives were as follows:* determine the facilities required for digital document supply service at the local level;* determine the type of digital receiving and delivery of document requests in the service; and* determine the service level in the above-mentioned service (e.g. receiving documents from Iranian or international regional libraries).\nMethodology and statistical population: This is a descriptive-survey study. The study is a descriptive one, since the aim of the researcher is to provide an objective, real, and systematic description of the characteristics of a situation or problem (Busha and Harter, 1980). Moreover, as the main presupposition of the method is to employ accurate scientific methods to achieve some deductions about a larger group using the findings obtained from the study of a small group of participants, the study can be considered a survey (Kumar, 2002). In general, considering these items, this is a descriptive survey. The study population consists of the librarians of central libraries and faculty libraries at three main state universities of Isfahan - University of Isfahan, Isfahan University of Medical Sciences, and Isfahan University of Technology - as well as the librarians and managers in the information sector of Isfahan province research institutes in 2010. The study population with respect to the three universities and the research institutes can be observed in Table I.Since the study population includes some subgroups, a random stratified sampling method was used to ensure the contribution of different population subgroups in proportion to their percentage of the population (Sha'bani and Sa'adat, 2009). The sample size was determined to be 132 with 95 per cent confidence using Krejcie and Morgan's (1970)[1]. The data was collected using a researcher-made questionnaire which consisted of two parts - one on the demographical characteristics of the participants and the other on items related to evaluation of the facilities of universities and research institutes of the province required for establishment of digital document supply services. The items were rated on a five-point Likert scale[2]. Face validity of the questionnaire was approved by experts in the field, and its reliability was determined through a Cronbach's a coefficient of 0.9[3]. To prevent the risk of data loss because of unreturned questionnaires, 145 questionnaires were distributed and 137 were handed in. Thus, the questionnaire return ratio was 94.48 per cent. Finally, the data was analyzed using the SPSS software. The results were provided as descriptive and inferential statistics. For the descriptive analysis, frequency distribution tables, percentage of frequency, mean, and standard deviation were used; in inferential analysis, a univariate t-test was employed[4].\nResearch background: In recent years digital document supply services have been studied by many researchers from different countries. For instance, Erokina (2010) considered document supply by the Russian State Library, Ping (2010) studied the development of document supply services in the People's Republic of China from an economic perspective, Abazari and Isfandyari-Moghaddam (2010) addressed the present facilities in Islamic science centres in the Islamic Republic of Iran, Soares (2010) considered the necessity of development of IT and libraries in Portugal, Schopfel and Gillet (2007) considered customers' requirements in France, and Siddiqui (1995, 1996) evaluated the development of a network of document supply in the Persian Gulf Arab countries and Saudi Arabia. Each of these studies has presented some novel aspects of the problem.The present viewpoints in the research background show that the position of document supply services in the modern world of digital information has been retained. Furthermore, all the researchers have emphasised electronic strategies in data transaction in their novel ideas. The approach is obvious in both developing and developed countries. Researchers who have tried to justify a combination of document supply services with resources, strategic searches, and their delivery, have evolved the development of digital services in libraries and modern IT facilities in accessing the materials. Therefore, document supply services with an emphasis on digitisation have provided an effective area in today's modern libraries. Considering this, the necessity of developing such systems in the current libraries of developing countries should be taken into account.\nFindings: The first research study deals with the hardware, software, network, and human resource facilities of the study population. Table II presents the frequency and percentage of the participants' responses with regard to the facilities of establishment of digital document supply services at the local level.The results in Table II indicate that 3.6 per cent, 23.4 per cent, 42.3 per cent, 19 per cent, and 11.7 per cent of the participants consider the possibility of establishing digital document supply service to be a great lot, a lot, quite some, few, and very few, respectively. The highest mean value was related to the possibility of internet access (42.3 per cent), while the lowest mean was related to access to scanning facilities (18.2 per cent). Overall, 69 per cent reacted at least somewhat favourably to the concept; however, the mean comparison to the expected score was not as positive. As shown in Table III the observed t-value is smaller than the critical value of the table at the 0.05 error level. Therefore the possibility of establishing a digital document supply service at the local level is less than the mean level.The second research question deals with the different methods of receiving and delivery of document requests in digital document supply services. Table IV demonstrates the frequency and percentage of the participants' responses with respect to the different methods of receipt and delivery of document requests.According to Table IV, the highest and lowest mean values are related to the facilities of using e-mail for receipt and delivery of requests (3.28 per cent) and the facilities of multimedia delivery in software packages (2.08 per cent), respectively.According to Table V, considering the various methods of receipt and delivery of document requests in the digital document supply service, the observed t-value is smaller than the critical value of the table at the error level 0.05. Therefore, the possibility of receipt and delivery of documents in the digital document supply service is less than the mean level. The third question of research evaluates the different levels of services in supplying digital documents. Table VI presents the frequency and percentage of the participants' responses on the different levels of supplying digital documents.The results of Table VI indicate that the highest and lowest means were related to using the documents of Isfahan universities (3.26 per cent) and using services of international document supplying (2.15 per cent), respectively.As can be observed in Table VII, since the obtained t-value is smaller than the critical value of the table at the error level 0.05, the possibility of providing different levels of services in digital document supply service is less than the mean level.\nDiscussion and conclusion: The current study was carried out to evaluate the feasibility of establishing digital document supply services in academic and research institute libraries of Isfahan province. As it was observed, the overall feasibility score at the <=0.05 error level was determined to be less than the mean level. Therefore, the possibility of establishing digital document supply services in the libraries is less than the mean level. With respect to the facilities required for the establishment of the service at the local level, it was observed that librarians believe the state of facilities such as scanners, printers, CD writers, purchase and installation of software, and even internet speed, which is based upon bandwidth, is not desirable. This is consistent with the findings of Abazari and Isfandyari-Moghaddam (2010). However, considering hardware and software development in digital document supply services, Erokina (2010) and Ping (2010) have addressed the competitive viewpoints in continuous access to these items, and mentioned the potential of libraries in Russia and China as witnesses to the role of these items in this area.Considering the items on methods of receipt and delivery of document requests in digital document supply services, it was observed that the librarians do not have enough confidence in the facilities of display of document request delivery and receipt on web pages, creating personal profiles, employment of SMS and fax, obtaining and devising software packages, and delivering multimedia formats. In fact, the drawbacks in their view were based upon some evidence of their problems in interacting with the data. This is while Soares (2010) and Siddiqui (1995) attempted to focus on digital electronic interaction in Portugal and establishment of a network in Saudi Arabia, which are not compatible with the current and conventional frameworks.Regarding the different levels of digital document supply service, it was observed that the librarians believed using documents of Isfahan province research institutes, as well as local and international document supply services, is undesirable. This item is related to internal local and national factors, as well as external relationships between the country's political and social systems. This problem has challenged the country's scientific and research policies in different ways. In fact, this is a common problem among many developing countries with respect to its economic costs, although China seems to be in a fairly good state in this regard (Ping, 2010) and Erokina (2010) believes that the relationship of Russia with regional and international networks is fairly good and effective.Therefore, in an accurate plan, the components and facilities of digital document supply service in universities of Isfahan province should receive special attention. The problem is addressed in the following suggestions:* carrying out educational programs for librarians in the new technological environment, particularly in academic libraries, with the goal of developing digital library services;* establishing a systematic consensual policy for the fulfilment of digital document supply services at the province level;* developing a single web-based portal in Isfahan province; and* providing an updated library resource index for the province.\n",
"role": "user"
},
{
"content": "Purpose: - The paper's aim is to study the feasibility of establishing digital document supply services in libraries of research institutes and universities in Isfahan province.\nDesign/methodology/approach: - The research is of a descriptive survey type. The population under study included librarians working in academic and research institute libraries in Isfahan province. The data were collected via a researcher-made questionnaire using a five-point Likert scale.\nFindings: - The study indicates that the feasibility of establishing digital document supply services in academic and research institute libraries in Isfahan province is less than average, which is not desirable.\nOriginality/value: - The paper provides information about the state of document supply services in Iranian academic libraries and research institutes.\n",
"role": "assistant"
}
] |
Identifying the effects of perceived values of mobile banking applications on customers: Comparative study between baby boomers, generation X and generation Y
|
[
"E-satisfaction",
"E-loyalty",
"Mobile banking",
"E-trust",
"Perceived values"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In a banking environment characterized by deep changes such as globalization, tough competition, financial crisis, development of information and communication technologies (ICTs) and shift in customers' expectations, banks have to diversify the services they offer to customers, in particular by taking advantage of technologies that will help facilitate interactions with them (Berraies et al., 2015).\n2. Literature review: Despite an extensive literature, a small number of academic research studies has been conducted to better understand factors that affect e-trust, e-satisfaction and e-loyalty in the context of MB applications, particularly in developing countries. In this regard, this research focuses on mobile technology used by banks as an innovative channel to provide financial services to their customers.\n3. Empirical study: In order to identify the links between the research variables and the moderating role, an empirical study was conducted among a sample of bank customers using the MB applications.\n4. Discussion: The results of this study provide interesting insights and reveal the specificity of the MB applications' users in the Tunisian context. Within this spectrum, this context is imbued with an economic crisis which causes a trust crisis among Tunisian banks' customers. This research is in the perspective of studies that have discussed the importance of trust, in particular in the context of MB, especially because banks are in the wake of financial crisis and thus they have to gain/repair customers' trust (Avlonitis and Indounas, 2014).\n5. Conclusion: This research focuses on the customers' perceived values in the context of use of the MB applications and their impact on e-trust, e-satisfaction and e-loyalty. In response to the tendency of customers to use more and more smartphones with internet access, banks have developed mobile applications to get closer to them. Technology innovation in terms of integrating MB applications becomes necessary for banks aiming to attract customers, particularly digital natives. Furthermore, in an environment weakened by the prevailing crisis, customers' trust remains an essential factor. Banks have to react to this trust deficit by developing a more personalized relationship with their customers and being increasingly close to them, namely through the implementation of these mobile applications.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is twofold: first, the authors empirically examine the effect of the perceived values of mobile banking (MB) applications on customers' e-trust, e-satisfaction and e-loyalty and second, the authors investigate the moderating role of customers' age in the relationship between perceived value of MB and e-trust.\nDesign/methodology/approach: A survey was conducted on a sample of 361 Tunisian banks' customers using mobile applications. Structural equation modeling method was performed to test the research hypotheses.\nFindings: The results of this research revealed that quality, price and emotional perceived values' dimensions of MB applications are predictors of customers' e-trust. Within this spectrum, age acts as a moderator in these latter relationships. Findings also suggested that e-trust has positive effects on e-satisfaction and e-loyalty.\nPractical implications: This paper presented a reading grid for bank marketers allowing them to identify the levers they must focus their efforts on so as to retain the customers using MB applications. It also permits them to identify the needs and values of three generation of customers, namely baby boomers, generation X and generation Y, so that they can offer better services for each group.\nOriginality/value: This paper sheds light on an innovative and unexplored topic in the literature. The authors broaden understanding of the relationship between the perceived values of MB applications, e-loyalty, e-trust and e-satisfaction. Research devoted to investigate these different links through an integrative model does exist, yet embryonic. This study also highlights differences between three group of customer, namely baby boomers, generation X and generation Y in terms of perceived values that builds their trust in the context of MB applications use.\n",
"role": "assistant"
}
] |
Service productisation: systematising and defining an offering
|
[
"Developing, reviewing and defining service productization",
"Productization",
"Service blueprinting",
"Service modularity",
"Service offering"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Managing services is crucial, especially in the case of large or expanding service businesses, but it also benefits organisations of a smaller scale. The service offering is the focal point of business performance, and the ability to explain the mechanisms of service offerings can prove beneficial for service management (Kwak and Kim, 2016). Apart from being viewed as an offering, service is also viewed as a process or as the exploitation of resources and competencies to gain benefits (Lusch, 2011). Recently, there have been references in literature to the productisation of services, but the true positioning of the concept in the field of service and its role in relation to service offering and processes have remained somewhat unclear. Harkonen et al. (2015, pp. 70-71) have made the most detailed attempt at defining the concept of service productisation (SP):Productisation is the process of analysing a need, defining and combining suitable elements into a product-like object, which is repeatable and comprehendible. Productisation activities cover those for commercial readiness to enable selling, delivering, using and invoicing. Productisation in the context of services addresses the objects of exchange that are typically abstract and intangible, and has a specific role in clarifying the service offering, creating replicability and enhancing understanding of the offering.\n2. Literature review: Over the past few decades, the literature has increasingly referred to SP, a concept that appears to be relatively new to service research, although it is seemingly widely used among practitioners and discussed in managerial magazines and seminars (Harkonen et al., 2015; Jaakkola, 2011; Valtakoski and Jarvi, 2016). As a first step, Table I highlights how SP is understood in the literature.\n3. Methodology: The research approach used in this study is qualitative and applies an inductive logic. The approach is influenced by Strauss and Corbin's (1997) work on the grounded theory development, according to which the researcher allows the theory to emerge from the data. The concept of an inductive analysis here refers to a research approach that relies on detailed readings of data in order to derive concepts, themes or a model by means of interpretation. The study has been completed by following the detailed guidance on a general inductive approach offered by Thomas (2006), which is guided by objectives or questions outlined by the researcher, condenses data into a brief format, establishes links between the research objectives and the summary findings, and develops a model or theory of findings that are evident in the data. In this study, the analysis is guided by the findings of a thorough literature review. This approach has been used by numerous authors working in a variety of fields, which indicates that it is not tied to any specific discipline and is hence appropriate for multidisciplinary work. Additionally, empirical data can be obtained from either primary or secondary sources. Table II details the overall approach adopted in this study. The benefit of using this particular research approach is that the findings are intimately tied to the evidence (Thomas, 2006).\n4. Findings: The findings of this study have been reported as elements of SP based on the patterns, themes and trends identified within and across the data. The framework created on the basis of the literature has been used to support the analysis. The findings are reported by grouping the evidence that support the identified elements and by presenting adequately rich details. The instances of SP have been indicated (A-M) so that the elements can be tracked back to the individual instances if necessary. The later synthesis section shows the relation between the empirical findings and the previous literature. To further address these findings, the appropriate literature has been cited and its relevance to SP and/or mismatches between the literature and the findings have been noted. The findings have then been further synthesised and a typology has been formulated to provide a clearer context for SP.\n5. Discussion: This study helps enhance the understanding of SP by bringing together past relevant journal discussions and by synthesising previously discrete elements into an ensemble of categories - a conceptual typology - for SP. This study identifies elements of SP and makes a clear contribution to practice and service literature. The created typology supports the understanding of what is considered as SP, both through the identified elements of SP and by presenting the elements in the context of commercial and technical service product portfolios to provide further clarity. The proposed typology is relevant for service product management to model services and support effective service management. The study clarifies the role of SP in relation to systematising and tangibilising the service offering and the related processes as well as in the creation of a suitable level of formalisation.\n",
"role": "user"
},
{
"content": "Purpose: The previous literature has indicated that the productisation of services may play a role in service management, although a certain level of obscurity still surrounds the concept. Therefore, the purpose of this paper is to clarify the meaning of service productisation (SP) as well as to contribute to a greater understanding of the concept.\nDesign/methodology/approach: An inductive analysis was applied to 13 instances of activities related to the productisation of services, with secondary data being analysed to identify practices relevant to SP and to examine their significance. The analysis is guided by an extensive literature review.\nFindings: SP has been found to play a role in systematising and tangibilising a service offering and its related processes as well as in formalising the processes and service offerings. The potential elements of SP have been identified and supporting evidence has been provided. The findings indicate that SP has a specific focus on the offering and its related processes, with the aim being to create a service product that can be sold, delivered and invoiced. SP may utilise various practices and techniques, and customer orientation also plays a significant role. A typology of SP has been created by reflecting on its commercial and technical aspects.\nPractical implications: This study has important implications for the service industry as it provides a structure and key considerations for productising services.\nOriginality/value: This study is one of the first to seek evidence for the concept of SP from multiple instances of SP as well as an extensive literature base. The typology created provides a context for discussing SP as well as reflecting on its commercial and technical aspects.\n",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.